Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DOC: update hot topics #456

Merged
merged 3 commits into from
Sep 15, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 5 additions & 5 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,15 +26,16 @@ potential of cutting-edge AI models.

## 🔥 Hot Topics
### Framework Enhancements
- Incorporate vLLM: [#445](https://github.com/xorbitsai/inference/pull/445)
- Embedding model support: [#418](https://github.com/xorbitsai/inference/pull/418)
- Custom model support: [#325](https://github.com/xorbitsai/inference/pull/325)
- LoRA support: [#271](https://github.com/xorbitsai/inference/issues/271)
- Multi-GPU support for PyTorch models: [#226](https://github.com/xorbitsai/inference/issues/226)
- Xinference dashboard: [#93](https://github.com/xorbitsai/inference/issues/93)
### New Models
- Built-in support for [CodeLLama](https://github.com/facebookresearch/codellama): [#414](https://github.com/xorbitsai/inference/pull/414) [#402](https://github.com/xorbitsai/inference/pull/402)
### Tools
- LlamaIndex plugin: [#7151](https://github.com/jerryjliu/llama_index/pull/7151)
### Integrations
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
- [Chatbox](https://chatboxai.app/): a desktop client for multiple cutting-edge LLM models, available on Windows, Mac and Linux.


## Key Features
Expand All @@ -57,8 +58,7 @@ for seamless management and monitoring.
allowing the seamless distribution of model inference across multiple devices or machines.

🔌 **Built-in Integration with Third-Party Libraries**: Xorbits Inference seamlessly integrates
with popular third-party libraries like [LangChain](https://python.langchain.com/docs/integrations/providers/xinference)
and [LlamaIndex](https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window).
with popular third-party libraries including [LangChain](https://python.langchain.com/docs/integrations/providers/xinference), [LlamaIndex](https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window), [Dify](https://docs.dify.ai/advanced/model-configuration/xinference), and [Chatbox](https://chatboxai.app/).

## Getting Started
Xinference can be installed via pip from PyPI. It is highly recommended to create a new virtual
Expand Down
14 changes: 7 additions & 7 deletions README_zh_CN.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,14 +23,17 @@ Xorbits Inference(Xinference)是一个性能强大且功能全面的分布

## 🔥 近期热点
### 框架增强
- 自定义模型: [#325](https://github.com/xorbitsai/inference/pull/325)
- 引入 vLLM: [#445](https://github.com/xorbitsai/inference/pull/445)
- Embedding 模型支持: [#418](https://github.com/xorbitsai/inference/pull/418)
- LoRA 支持: [#271](https://github.com/xorbitsai/inference/issues/271)
- PyTorch 模型多 GPU 支持: [#226](https://github.com/xorbitsai/inference/issues/226)
- Xinference 仪表盘: [#93](https://github.com/xorbitsai/inference/issues/93)
### 新模型
- 内置 [CodeLLama](https://github.com/facebookresearch/codellama): [#414](https://github.com/xorbitsai/inference/pull/414) [#402](https://github.com/xorbitsai/inference/pull/402)
### 工具
- LlamaIndex 插件: [#7151](https://github.com/jerryjliu/llama_index/pull/7151)
### 集成
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): 一个涵盖了大型语言模型开发、部署、维护和优化的 LLMOps 平台。
- [Chatbox](https://chatboxai.app/): 一个支持前沿大语言模型的桌面客户端,支持 Windows,Mac,以及 Linux。




Expand All @@ -39,16 +42,13 @@ Xorbits Inference(Xinference)是一个性能强大且功能全面的分布

⚡️ **前沿模型,应有尽有**:框架内置众多中英文的前沿大语言模型,包括 baichuan,chatglm2 等,一键即可体验!内置模型列表还在快速更新中!


🖥 **异构硬件,快如闪电**:通过 [ggml](https://github.com/ggerganov/ggml),同时使用你的 GPU 与 CPU 进行推理,降低延迟,提高吞吐!

⚙️ **接口调用,灵活多样**:提供多种使用模型的接口,包括 RPC,RESTful API,命令行,web UI 等等。方便模型的管理与监控。

🌐 **集群计算,分布协同**: 支持分布式部署,通过内置的资源调度器,让不同大小的模型按需调度到不同机器,充分使用集群资源。

🔌 **开放生态,无缝对接**: 与流行的三方库无缝对接,包括 [LangChain](https://python.langchain.com/docs/integrations/providers/xinference)
and [LlamaIndex](https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window)。
让开发者能够快速构建基于 AI 的应用。
🔌 **开放生态,无缝对接**: 与流行的三方库无缝对接,包括 [LangChain](https://python.langchain.com/docs/integrations/providers/xinference),[LlamaIndex](https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window),[Dify](https://docs.dify.ai/advanced/model-configuration/xinference),以及 [Chatbox](https://chatboxai.app/)。

## 快速入门
Xinference 可以通过 pip 从 PyPI 安装。我们非常推荐在安装前创建一个新的虚拟环境以避免依赖冲突。
Expand Down
19 changes: 11 additions & 8 deletions doc/source/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -33,28 +33,31 @@ allowing the seamless distribution of model inference across multiple devices or

🔌 **Built-in Integration with Third-Party Libraries**: Xorbits Inference seamlessly integrates
with popular third-party libraries like `LangChain <https://python.langchain.com/docs/integrations/providers/xinference>`_
and `LlamaIndex <https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window>`_.
, `LlamaIndex <https://gpt-index.readthedocs.io/en/stable/examples/llm/XinferenceLocalDeployment.html#i-run-pip-install-xinference-all-in-a-terminal-window>`_
, `Dify <https://docs.dify.ai/advanced/model-configuration/xinference>`_
, and `Chatbox <https://chatboxai.app/>`_.


🔥 Hot Topics
-------------

Framework Enhancements
~~~~~~~~~~~~~~~~~~~~~~
- Custom model support: `#325 <https://github.com/xorbitsai/inference/pull/325>`_
- Incorporate vLLM: `#445 <https://github.com/xorbitsai/inference/pull/445>`_
- Embedding model support: `#418 <https://github.com/xorbitsai/inference/pull/418>`_
- LoRA support: `#271 <https://github.com/xorbitsai/inference/issues/271>`_
- Multi-GPU support for PyTorch models: `#226 <https://github.com/xorbitsai/inference/issues/226>`_
- Xinference dashboard: `#93 <https://github.com/xorbitsai/inference/issues/93>`_

New Models
~~~~~~~~~~
- Built-in support for `Starcoder` in GGML: `#289 <https://github.com/xorbitsai/inference/pull/289>`_
- Built-in support for `MusicGen <https://github.com/facebookresearch/audiocraft/blob/main/docs/MUSICGEN.md>`_: `#313 <https://github.com/xorbitsai/inference/issues/313>`_
- Built-in support for `SD-XL <https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0>`_: `318 <https://github.com/xorbitsai/inference/issues/318>`_
- Built-in support for `CodeLLama <https://github.com/facebookresearch/codellama>`_: `#414 <https://github.com/xorbitsai/inference/pull/414>`_ `#402 <https://github.com/xorbitsai/inference/pull/402>`_

Tools
~~~~~
- LlamaIndex plugin: `7151 <https://github.com/jerryjliu/llama_index/pull/7151>`_

Integrations
~~~~~~~~~~~~
- `Dify <https://docs.dify.ai/advanced/model-configuration/xinference>`_: an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
- `Chatbox <https://chatboxai.app/>`_: a desktop client for multiple cutting-edge LLM models, available on Windows, Mac and Linux.


License
Expand Down
6 changes: 3 additions & 3 deletions doc/source/models/builtin/llama-2-chat.rst
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.


Model Spec 5 (pytorch, 13 Billion)
Expand All @@ -84,7 +84,7 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.

Model Spec 6 (pytorch, 70 Billion)
++++++++++++++++++++++++++++++++++
Expand All @@ -101,4 +101,4 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.
6 changes: 3 additions & 3 deletions doc/source/models/builtin/llama-2.rst
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.

Model Spec 5 (pytorch, 13 Billion)
++++++++++++++++++++++++++++++++++
Expand All @@ -82,7 +82,7 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.

Model Spec 6 (pytorch, 70 Billion)
++++++++++++++++++++++++++++++++++
Expand All @@ -99,4 +99,4 @@ chosen quantization method from the options listed above::

.. note::

4-bit quantization is not supported on macOS.
4-bit quantization is not supported on macOS.