Skip to content

Commit 2e610de

Browse files
authored
[CI/Build] Enable phi2 lora test (#20540)
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
1 parent 6e2c19c commit 2e610de

File tree

1 file changed

+0
-5
lines changed

1 file changed

+0
-5
lines changed

tests/lora/test_phi.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,6 @@
11
# SPDX-License-Identifier: Apache-2.0
22
# SPDX-FileCopyrightText: Copyright contributors to the vLLM project
33

4-
import pytest
5-
64
import vllm
75
from vllm.lora.request import LoRARequest
86

@@ -49,9 +47,6 @@ def do_sample(llm: vllm.LLM, lora_path: str, lora_id: int) -> list[str]:
4947
return generated_texts
5048

5149

52-
# Skipping for V1 for now as we are hitting,
53-
# "Head size 80 is not supported by FlashAttention." error.
54-
@pytest.mark.skip(reason="Head size 80 is not supported by FlashAttention")
5550
def test_phi2_lora(phi2_lora_files):
5651
# We enable enforce_eager=True here to reduce VRAM usage for lora-test CI,
5752
# Otherwise, the lora-test will fail due to CUDA OOM.

0 commit comments

Comments
 (0)