-
-
Notifications
You must be signed in to change notification settings - Fork 11.3k
[CI] Enable encoder model compilation test #24442
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request enables a model compilation test for the BAAI/bge-base-en-v1.5 BERT-based model by uncommenting the corresponding TestSetting in test_basic_correctness.py. The attention backend for this test case has been updated from XFORMERS to FLASH_ATTN, aligning it with other tests in the suite. The changes are straightforward and improve test coverage for encoder models. The code looks good.
ProExpertProg
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@russellb is this supported in V1 already? If yes LGTM
|
@ProExpertProg It is supported in V1 now. You can take a look at this https://docs.vllm.ai/en/latest/models/supported_models.html#embedding |
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com> Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Purpose
Bert models has already been supported in V1 now.
Enable encoder model compilation test
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.