[Question]: From which version at least, the vllm supports to infrence and serve Qwen2.5-14B-Instruct model? #962
Answered
by
jklj077
zengqingfu1442
asked this question in
Q&A
-
Has this been raised before?
DescriptionWhich version of vllm support to inference and serve Qwen2.5-14B-Instruct model? |
Beta Was this translation helpful? Give feedback.
Answered by
jklj077
Sep 26, 2024
Replies: 2 comments
-
the latest version is recommended. tested with v0.6.1. qwen2 arch (Qwen1.5, Qwen2, Qwen2.5) is supported by vllm since v0.3.0. |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
zengqingfu1442
-
If you want to use the tools calling feature of Qwen2.5, you need to use version |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
the latest version is recommended. tested with v0.6.1. qwen2 arch (Qwen1.5, Qwen2, Qwen2.5) is supported by vllm since v0.3.0.