-
-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Vllm api server does not receive supported parameter truncate_prompt_tokens
#6890
Comments
Did you make a typo? I could not find |
max_prompt_tokens
truncate_prompt_tokens
@DarkLight1337, my bad. It's |
Does the server receive the request at all? Or is it a validation error from the OpenAI client? |
@DarkLight1337 I saw the request on server side, with 400 status code. |
Which vLLM version are you using? I'm not getting this problem on my end. Please run |
I am using vllm version 0.5.1, I checked the source codes for this version (in my env cache) and saw that |
Just in case, can you also check your Pydantic version? |
My Pydantic version is |
Similar issue when using "prompt_logprob" parameters, when serving "MiniCPM-Llama3-V-2_5" @DarkLight1337 Serve command:
Python code
Return
The environment
Pydantic version
|
Does this happen for any other model? |
Same issue when using |
|
Actually, there is an existing workaround for this: #6508 |
Thanks!! I will try to follow this on minicpm_v2_5 and internvl2-4B. |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
Your current environment
🐛 Describe the bug
I used the openai compatible server deployed with vllm:
When I send a request with the following snippet (openai client):
I got the error:
The following code, however, works:
I wonder why in https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#extra-parameters,
truncate_prompt_tokens
is supported but I am getting the error here ?The text was updated successfully, but these errors were encountered: