v0.6.5
What's new in 0.6.5 (2023-12-01)
These are the changes in inference v0.6.5.
New features
- FEAT: Support jina embedding models by @aresnow1 in #704
- FEAT: Support Yi-chat by @aresnow1 in #700
- FEAT: Support qwen 72b by @aresnow1 in #705
- FEAT: ChatGLM3 tool calls by @codingl2k1 in #701
Enhancements
- ENH: Specify actor pool port for distributed deployment by @ChengjieLi28 in #688
- ENH: Remove
xorbits
dependency by @ChengjieLi28 in #699 - ENH: User can just specify a string for prompt style when registering custom LLM models by @ChengjieLi28 in #682
- ENH: Add more models supported by vllm by @aresnow1 in #706
Bug fixes
- BUG: Fix xinference start failed if invalid custom model found by @codingl2k1 in #690
Documentation
- Doc: Fix some incorrect links in documentation by @aresnow1 in #684
- Doc: Update readme by @aresnow1 in #687
- DOC: documentation for docker and k8s by @lynnleelhl in #661
Others
New Contributors
- @lynnleelhl made their first contribution in #661
Full Changelog: v0.6.4...v0.6.5