-
Notifications
You must be signed in to change notification settings - Fork 477
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
FEAT: Support minicpm-4B #2263
FEAT: Support minicpm-4B #2263
Conversation
Support minicpm-4B
Support minicpm-4B
Support minicpm-4B
Support minicpm-4B
Support minicpm-4B
Support minicpm-4B
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you add this model to
inference/xinference/model/llm/vllm/core.py
Line 133 in d2a1196
] |
to support vllm?
vllm is not officially supported yet,See the vllm branch maintained by minicpm https://github.com/OpenBMB/vllm/tree/minicpm3 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
模型上线页面没有minicpm-4b模型 |
您是pip 安装的的xinference把,这个支持目前还没推送到pip源上,要等周五晚上发版 |
是用docker部署的,昨天pull的nightly-main镜像里找不到minicpm3-4b |
支持sglang推理吗?我看minicpm是有支持sglang的 |
sglang 添加模型只需要添加模型名到列表:
欢迎贡献 PR。 |
Testing has been done locally and the model requires the following two dependencies to be added - datamodel_code_generator and jsonschema