Skip to content

Conversation

@ZJY0516
Copy link
Contributor

@ZJY0516 ZJY0516 commented Sep 8, 2025

Purpose

Bert models has already been supported in V1 now.
Enable encoder model compilation test

Test Plan

pytest tests/compile/test_basic_correctness.py

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
@ZJY0516 ZJY0516 changed the title [Test] Enable encoder model compilation test [CI] Enable encoder model compilation test Sep 8, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request enables a model compilation test for the BAAI/bge-base-en-v1.5 BERT-based model by uncommenting the corresponding TestSetting in test_basic_correctness.py. The attention backend for this test case has been updated from XFORMERS to FLASH_ATTN, aligning it with other tests in the suite. The changes are straightforward and improve test coverage for encoder models. The code looks good.

Copy link
Collaborator

@ProExpertProg ProExpertProg left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@russellb is this supported in V1 already? If yes LGTM

@ZJY0516
Copy link
Contributor Author

ZJY0516 commented Sep 8, 2025

@ProExpertProg It is supported in V1 now. You can take a look at this https://docs.vllm.ai/en/latest/models/supported_models.html#embedding

@ProExpertProg ProExpertProg enabled auto-merge (squash) September 8, 2025 16:35
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 8, 2025
@vllm-bot vllm-bot merged commit 7be141b into vllm-project:main Sep 8, 2025
27 of 29 checks passed
@ZJY0516 ZJY0516 deleted the encoder-test branch September 9, 2025 03:19
eicherseiji pushed a commit to eicherseiji/vllm that referenced this pull request Sep 9, 2025
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
skyloevil pushed a commit to skyloevil/vllm that referenced this pull request Sep 13, 2025
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
FeiDaLI pushed a commit to FeiDaLI/vllm that referenced this pull request Sep 25, 2025
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 10, 2025
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
xuebwang-amd pushed a commit to xuebwang-amd/vllm that referenced this pull request Oct 24, 2025
Signed-off-by: zjy0516 <riverclouds.zhu@qq.com>
Signed-off-by: xuebwang-amd <xuebwang@amd.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants