Releases: svilupp/PromptingTools.jl
Releases · svilupp/PromptingTools.jl
v0.72.0
PromptingTools v0.72.0
Breaking Changes
Added
- Added new Gemini 2.0 Pro and Flash models to the model registry (aliases
gem20p
andgem20f
, respectively). Added Gemini 2.0 Flash Lite Preview model (aliasgem20fl
) and Gemini 2.0 Flash Thinking Experimental model (aliasgem20ft
). - Added BM25 normalization kwarg to
RAGTools.jl
to enable 0-1 query-specific normalization of BM25 scores for easier filtering and comparison. See?RT.bm25
and?RT.max_bm25_score
for more information.
Commits
Merged pull requests:
v0.71.0
PromptingTools v0.71.0
Breaking changes
Added
- Added the new OpenAI's O3 Mini reasoning model to the model registry (alias
o3m
). - Added the new DeepSeek R1 Distill Llama 70b model hosted on GroqCloud, which beats GPT 4o in many benchmarks while being upto 10x cheaper (alias
glmr
- stands for Groq Llama Medium(70b) Reasoning). - Added experimental support for "thinking tokens" that can be found with DeepSeek API and the reasoning model R1 (alias
dsr
). If the thought chain is provided, we save it in themsg.extras[:reasoning_content]
field for advanced users.
Commits
Merged pull requests:
v0.70.0
PromptingTools v0.70.0
Breaking Changes
Added
- Added new DeepSeek V3 model via Deepseek (alias
dschat
or simplyds
, because they are in a category of their own), Fireworks.ai and Together.ai (fds
andtds
for hosted Deepseek V3, respectively). Added Qwen 2.5 Coder 32B (aliasfqwen25c
ortqwen25c
for Fireworks.ai and Together.ai, respectively). - Added the reasoning Qwen QwQ 32B hosted on Together.ai.
- Added the new OpenAI's O1 model to the model registry (alias
o1
). - Added DeepSeek Reasoner model alias (
dsreason
or simplydsr
) from DeepSeek - Added MiniMax model support and alias (
minimax
) for MiniMax-Text-01 model
Commits
Merged pull requests:
v0.69.1
PromptingTools v0.69.1
Fixed
- Added assertion in
response_to_message
for missing:tool_calls
key in the response message. It's model failure but it wasn't obvious from the original error. - Fixed error for usage information in CamelCase from OpenAI servers (Gemini proxy now sends it in CamelCase).
Commits
Merged pull requests:
Closed issues:
- Update Gemini Experimental model to 1206 (#253)
v0.69.0
PromptingTools v0.69.0
Added
- Added a new Gemini 2.0 Flash Experimental model (
gemini-2.0-flash-exp
) and updated the aliasgem20f
with it.
Commits
Merged pull requests:
v0.68.0
PromptingTools v0.68.0
Added
- Added a new
cache=:all_but_last
cache strategy for Anthropic models to enable caching of the entire conversation except for the last user message (useful for longer conversations that you want to re-use, but not continue). See the docstrings for more information on which cache strategy to use. Thank you to @Sixzero for this!
Commits
Merged pull requests:
v0.67.0
PromptingTools v0.67.0
Added
- Added a new Gemini Experimental model from December 2024 (
gemini-exp-1206
) and updated thegemexp
alias to point to it.
Commits
Merged pull requests:
- Add Gemini experimental model 1206 and update alias (#252) (@devin-ai-integration[bot])
v0.66.0
PromptingTools v0.66.0
Added
- Added support for Groq's new Llama3.3 models. Updated
gllama370
,gl70
,glm
aliases tollama-3.3-70b-versatile
and addedgl70s
,glms
aliases tollama-3.3-70b-specdec
(faster with speculative decoding).
Fixed
- Fixed a bug in
extract_docstring
where it would not correctly block "empty" docstrings on Julia 1.11.
Commits
Merged pull requests:
v0.65.1
PromptingTools v0.65.1
Fixed
- Removed unnecessary printing to
stdout
during precompilation inprecompile.jl
. - Fixed a "bug-waiting-to-happen" in tool use.
to_json_type
now enforces users to provide concrete types, because abstract types can lead to errors during JSON3 deserialization. - Flowed through a bug fix in
StreamCallback
where the usage information was being included in the response even whenusage=nothing
. Lower bound ofStreamCallbacks
was bumped to0.5.1
.
Commits
Merged pull requests:
Closed issues:
v0.65.0
PromptingTools v0.65.0
Breaking
- Changed the official ENV variable for MistralAI API from
MISTRALAI_API_KEY
toMISTRAL_API_KEY
to be compatible with the Mistral docs.
Added
- Added a new Gemini Experimental model from November 2024 (
gemini-exp-1121
with aliasgemexp
). - Added a new
AnnotationMessage
type for keeping human-only information in the message changes. See?annotate!
on how to use it. - Added a new
ConversationMemory
type (exported) to enable long multi-turn conversations with a truncated memory of the conversation history. Truncation works in "batches" to not prevent caching. See?ConversationMemory
andget_last
for more information.
Updated
- Changed the ENV variable for MistralAI API from
MISTRALAI_API_KEY
toMISTRAL_API_KEY
to be compatible with the Mistral docs.
Commits
Merged pull requests: