v6.3.0
This release adds new LLM inference methods, API Authorization and RAG improvements
📄 New LLM methods. llama.cpp and LiteLLM support added. LLM pipeline now supports Hugging Face models, GGUF files and LLM API inference all with one line of code.
🔒 API Authorization. Adds support for API keys and pluggable authentication methods when running through txtai API.
See below for full details on the new features, improvements and bug fixes.
New Features
- Add llama.cpp support to LLM (#611)
- Integrate with Litellm (#554)
- Add API route dependencies (#623)
- Add API Authorization (#263, #624)
- Add notebook on how to build RAG pipelines (#605)
- Add notebook showing how to use llama.cpp, LiteLLM and custom generation models (#615)
Improvements
- Enhance textractor to better support RAG use cases (#603)
- Update text extraction notebook (#604)
- Extractor (RAG) pipeline improvements (#613)
- Refactor LLM pipeline to support multiple framework methods (#614)
- Change API startup event to lifespan event (#625)