Skip to content

Conversation

@linfeng-yuan
Copy link
Collaborator

@linfeng-yuan linfeng-yuan commented Jun 20, 2025

What this PR does / why we need it?

Fix the issue of insufficient cached cosine and sine length in MLA's TorchAir graph mode, which causes accuracy deviation during long-sequence inference.

Backported: #1331

Does this PR introduce any user-facing change?

No.

How was this patch tested?

We tested the accuracy of this patch with DeepSeek R1 e2e becnhmark serving, and get 83.33 sore for AIME2024 dataset with DP4TP4EP16 setting.
image

… long sequence scenarios

Signed-off-by: linfeng-yuan <1102311262@qq.com>
@wangxiyuan wangxiyuan merged commit a3a3d38 into vllm-project:v0.9.1-dev Jun 21, 2025
16 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants