Skip to content

v6.3.0

Compare
Choose a tag to compare
@davidmezzetti davidmezzetti released this 02 Jan 14:54
· 316 commits to master since this release

This release adds new LLM inference methods, API Authorization and RAG improvements

📄 New LLM methods. llama.cpp and LiteLLM support added. LLM pipeline now supports Hugging Face models, GGUF files and LLM API inference all with one line of code.

🔒 API Authorization. Adds support for API keys and pluggable authentication methods when running through txtai API.

See below for full details on the new features, improvements and bug fixes.

New Features

  • Add llama.cpp support to LLM (#611)
  • Integrate with Litellm (#554)
  • Add API route dependencies (#623)
  • Add API Authorization (#263, #624)
  • Add notebook on how to build RAG pipelines (#605)
  • Add notebook showing how to use llama.cpp, LiteLLM and custom generation models (#615)

Improvements

  • Enhance textractor to better support RAG use cases (#603)
  • Update text extraction notebook (#604)
  • Extractor (RAG) pipeline improvements (#613)
  • Refactor LLM pipeline to support multiple framework methods (#614)
  • Change API startup event to lifespan event (#625)

Bug Fixes

  • Handle None input properly in Tokenizer (#607)
  • Issue with subdirectories and ZIP compression (#609)
  • Error in 52_Build_RAG_pipelines_with_txtai.ipynb (#620)
  • Add missing skl2onnx dependency (#622)