-
-
Notifications
You must be signed in to change notification settings - Fork 11k
[FIX] Throwing an exception when the model does not support pool tasks (#25840) #25855
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FIX] Throwing an exception when the model does not support pool tasks (#25840) #25855
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request correctly addresses a ValueError that occurs when _dummy_pooler_run is called for a model that does not support any pooling tasks. The fix introduces a check at the beginning of the function to raise a RuntimeError with a clear message, which is a good improvement. I have a couple of suggestions to further refine the code by removing a redundant function call and a now-unnecessary assertion.
15cfb10 to
25198aa
Compare
|
Sorry for overlooking as_reward_model in #20930 try add @default_pooling_type("ALL") before L390 vllm/vllm/model_executor/models/adapters.py Lines 387 to 392 in bd51f78
And comment: The Reward model uses all pooling by default refer to vllm/vllm/model_executor/models/qwen2_rm.py Lines 96 to 100 in bd51f78
|
|
@noooop why should add |
25198aa to
3bc918f
Compare
Lines 407 to 418 in 9360d34
L410 requires pooling_type.lower() != "last" anyway. This will automatically turn off chunked prefill. Please find a way to make setting @default_pooling_type("ALL") work |
3bc918f to
38d8700
Compare
Signed-off-by: zxw <1020938856@qq.com>
38d8700 to
ea0edbb
Compare
|
@yyzxw Thanks for your contribution language-models-test-extended-pooling failure is unrelated to this PR Will be fixed in #25817 @DarkLight1337 Is there anything else that needs to be modified in this PR? |
|
No |
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: 1994 <1994@users.noreply.github.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: Dhruvil Bhatt <bhattdbh@amazon.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: bbartels <benjamin@bartels.dev>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
vllm-project#25840) (vllm-project#25855) Signed-off-by: zxw <1020938856@qq.com> Co-authored-by: wang.yuqi <noooop@126.com> Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
Purpose
fix #25840
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.