Skip to content

Releases: svilupp/PromptingTools.jl

v0.24.0

07 May 08:47
Compare
Choose a tag to compare

PromptingTools v0.24.0

Diff since v0.23.0

Added

  • Added support for DeepSeek models via the dschat and dscode aliases. You can set the DEEPSEEK_API_KEY environment variable to your DeepSeek API key.

Commits

Merged pull requests:

v0.23.0

27 Apr 17:07
Compare
Choose a tag to compare

PromptingTools v0.23.0

Diff since v0.22.0

Added

  • Added new prompt templates for "Expert" tasks like LinuxBashExpertAsk, JavascriptExpertTask, etc.
  • Added new prompt templates for self-critiquing agents like ChiefEditorTranscriptCritic, JuliaExpertTranscriptCritic, etc.

Updated

  • Extended aicodefixer_feedback methods to work with AICode and AIGenerate.

Commits

Merged pull requests:

Closed issues:

  • Unable to get started (#132)

v0.22.0

20 Apr 06:43
Compare
Choose a tag to compare

PromptingTools v0.22.0

Diff since v0.21.0

### Added

  • Added support for Groq, the fastest LLM provider out there. It's free for now, so you can try it out - you just need to set your GROQ_API_KEY. We've added Llama3 8b (alias "gllama3"), 70b (alias "gllama370") and Mixtral 8x7b (alias "gmixtral"). For the shortcut junkies, we also added a shorthand Llama3 8b = "gl3" (first two letters and the last digit), Llama3 70b = "gl70" (first two letters and the last two digits).

Commits

Merged pull requests:

v0.21.0

19 Apr 20:24
Compare
Choose a tag to compare

PromptingTools v0.21.0

Diff since v0.20.1

Added

  • New models added to the model registry: Llama3 8b on Ollama (alias "llama3" for convenience) and on Together.ai (alias "tllama3", "t" stands for Together.ai), also adding the llama3 70b on Together.ai (alias "tllama370") and the powerful Mixtral-8x22b on Together.ai (alias "tmixtral22").

Fixed

  • Fixed a bug where pretty-printing RAGResult would forget a newline between the sources and context sections.

### Commits

Merged pull requests:

v0.20.1

18 Apr 09:04
Compare
Choose a tag to compare

PromptingTools v0.20.1

Diff since v0.20.0

### Fixed

  • Fixed truncate_dimension to ignore when 0 is provided (previously it would throw an error).

Commits

Merged pull requests:

v0.20.0

17 Apr 21:02
Compare
Choose a tag to compare

PromptingTools v0.20.0

Diff since v0.19.0

Added

  • Added a few new open-weights models hosted by Fireworks.ai to the registry (DBRX Instruct, Mixtral 8x22b Instruct, Qwen 72b). If you're curious about how well they work, try them!
  • Added basic support for observability downstream. Created custom callback infrastructure with initialize_tracer and finalize_tracer and dedicated types are TracerMessage and TracerMessageLike. See ?TracerMessage for more information and the corresponding aigenerate docstring.
  • Added MultiCandidateChunks which can hold candidates for retrieval across many indices (it's a flat structure to be similar to CandidateChunks and easy to reason about).
  • JSON serialization support extended for RAGResult, CandidateChunks, and MultiCandidateChunks to increase observability of RAG systems
  • Added a new search refiner TavilySearchRefiner - it will search the web via Tavily API to try to improve on the RAG answer (see ?refine!).
  • Introduced a few small utilities for manipulation of nested kwargs (necessary for RAG pipelines), check out getpropertynested, setpropertynested, merge_kwargs_nested.

Updated

  • [BREAKING] change to CandidateChunks where it's no longer allowed to be nested (ie, cc.positions being a list of several CandidateChunks). This is a breaking change for the RAGTools module only. We have introduced a new MultiCandidateChunks types that can refer to CandidateChunks across many indices.
  • Changed default model for RAGTools.CohereReranker to "cohere-rerank-english-v3.0".

Fixed

  • wrap_string utility now correctly splits only on spaces. Previously it would split on newlines, which would remove natural formatting of prompts/messages when displayed via pprint

Commits

Merged pull requests:

Closed issues:

  • [Question/FR] Using some HugginFace models (#127)
  • [DOCS] Markdown is not parsed in hero cards, use <pre> instead! (#128)
  • Make REPL mode (#131)

v0.19.0

10 Apr 06:45
Compare
Choose a tag to compare

PromptingTools v0.19.0

Diff since v0.18.0

Added

  • [BREAKING CHANGE] The default GPT-4 Turbo model alias ("gpt4t") now points to the official GPT-4 Turbo endpoint ("gpt-4-turbo").
  • Adds references to mistral-tiny (7bn parameter model from MistralAI) to the model registry for completeness.
  • Adds the new GPT-4 Turbo model ("gpt-4-turbo-2024-04-09"), but you can simply use alias "gpt4t" to access it.

Commits

Merged pull requests:

v0.18.0

05 Apr 13:38
Compare
Choose a tag to compare

PromptingTools v0.18.0

Diff since v0.17.1

Added

  • Adds support for binary embeddings in RAGTools (dispatch type for find_closest is finder=BinaryCosineSimilarity()), but you can also just convert the embeddings to binary yourself (always choose Matrix{Bool} for speed, not BitMatrix) and use without any changes (very little performance difference at the moment).
  • Added Ollama embedding models to the model registry ("nomic-embed-text", "mxbai-embed-large") and versioned MistralAI models.
  • Added template for data extraction with Chain-of-thought reasoning: :ExtractDataCoTXML.
  • Added data extraction support for Anthropic models (Claude 3) with aiextract. Try it with Claude-3 Haiku (model="claudeh") and Chain-of-though template (:ExtractDataCoTXML). See ?aiextract for more information and check Anthropic's recommended practices.

Fixed

Commits

Merged pull requests:

v0.17.1

27 Mar 13:29
Compare
Choose a tag to compare

PromptingTools v0.17.1

Diff since v0.17.0

Fixed

  • Fixed a bug in print_html where the custom kwargs were not being passed to the HTML constructor.

Commits

Merged pull requests:

v0.17.0

27 Mar 10:21
Compare
Choose a tag to compare

PromptingTools v0.17.0

Diff since v0.16.1

Added

  • Added support for aigenerate with Anthropic API. Preset model aliases are claudeo, claudes, and claudeh, for Claude 3 Opus, Sonnet, and Haiku, respectively.
  • Enabled the GoogleGenAI extension since GoogleGenAI.jl is now officially registered. You can use aigenerate by setting the model to gemini and providing the GOOGLE_API_KEY environment variable.
  • Added utilities to make preparation of finetuning datasets easier. You can now export your conversations in JSONL format with ShareGPT formatting (eg, for Axolotl). See ?PT.save_conversations for more information.
  • Added print_html utility for RAGTools module to print HTML-styled RAG answer annotations for web applications (eg, Genie.jl). See ?PromptingTools.Experimental.RAGTools.print_html for more information and examples.

Commits

Merged pull requests:

Closed issues:

  • [FR] Add support for Claude API (#96)