Skip to content

Commit b9aa9ac

Browse files
mend
Signed-off-by: hfadzxy <starmoon_zhang@163.com>
1 parent 2b598fd commit b9aa9ac

File tree

3 files changed

+5
-8
lines changed

3 files changed

+5
-8
lines changed

.github/workflows/vllm_ascend_test_long_term.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -96,8 +96,8 @@ jobs:
9696
- name: Run vllm-project/vllm-ascend long term test
9797
run: |
9898
if [[ "${{ matrix.os }}" == "linux-arm64-npu-1" ]]; then
99-
pytest -sv tests/e2e/long_term/test_accuracy.py
100-
# else
99+
pytest -sv tests/e2e/long_term/accuracy/accuracy_singlecard.py
100+
else
101101
# accuracy test multi card
102-
# VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/long_term/test_deepseek_v2_lite_tp2_accuracy.py
102+
pytest -sv tests/e2e/long_term/accuracy/accuracy_multicard.py
103103
fi

tests/e2e/long_term/accuracy/accuracy_multicard.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@
6161
EXPECTED_VALUE = {
6262
"Qwen/Qwen2.5-0.5B-Instruct": 0.316,
6363
"Qwen/Qwen3-30B-A3B": 0.888,
64-
"deepseek-ai/DeepSeek-V2-Lite": 0.376
64+
"deepseek-ai/DeepSeek-V2-Lite": 0.375
6565
}
6666
# Maximum context length configuration for each model.
6767
MAX_MODEL_LEN = {
@@ -98,7 +98,6 @@
9898
}
9999

100100
multiprocessing.set_start_method("spawn", force=True)
101-
os.environ["VLLM_USE_V1"] = "1"
102101

103102

104103
def run_test(queue, model, max_model_len, model_type, more_args):

tests/e2e/long_term/accuracy/accuracy_singlecard.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -98,10 +98,8 @@ def run_test(queue, model, max_model_len, model_type):
9898

9999

100100
@pytest.mark.parametrize("model", MODEL_NAME)
101-
@pytest.mark.parametrize("VLLM_USE_V1", ["0", "1"])
102-
def test_lm_eval_accuracy(monkeypatch: pytest.MonkeyPatch, model, VLLM_USE_V1):
101+
def test_lm_eval_accuracy(monkeypatch: pytest.MonkeyPatch, model):
103102
with monkeypatch.context() as m:
104-
m.setenv("VLLM_USE_V1", VLLM_USE_V1)
105103
result_queue: Queue[float] = multiprocessing.Queue()
106104
p = multiprocessing.Process(target=run_test,
107105
args=(result_queue, model,

0 commit comments

Comments
 (0)