Skip to content

Commit 79b8ec8

Browse files
committed
fix test
Signed-off-by: MengqingCao <cmq0113@163.com>
1 parent 6d9b45a commit 79b8ec8

File tree

1 file changed

+2
-8
lines changed

1 file changed

+2
-8
lines changed

tests/e2e/long_term/accuracy/accuracy_multicard.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@
2222
import subprocess
2323
import sys
2424
import time
25-
from multiprocessing import Queue
2625
from concurrent.futures import ThreadPoolExecutor, as_completed
26+
from multiprocessing import Queue
2727

2828
import lm_eval
2929
import pytest
@@ -146,10 +146,6 @@ def test_lm_eval_accuracy(monkeypatch: pytest.MonkeyPatch, model):
146146
f"Expected: {EXPECTED_VALUE[model]}±{RTOL} | Measured: {result}"
147147

148148

149-
150-
vllm serve "/opt/disk2/cache/modelscope/hub/models/Qwen/Qwen3-30B-A3B" --max-model-len 4096 --data-parallel-size 2
151-
152-
153149
@pytest.mark.parametrize("max_tokens", [10])
154150
@pytest.mark.parametrize("model",
155151
["Qwen/Qwen2.5-0.5B-Instruct", "Qwen/Qwen3-30B-A3B"])
@@ -165,7 +161,6 @@ def test_lm_eval_accuracy_dp(model, max_tokens):
165161
stdout=log_file,
166162
stderr=subprocess.DEVNULL)
167163

168-
169164
def _send_request(prompt, max_tokens=16, temperature=0.0):
170165
payload = {
171166
"prompt": prompt,
@@ -183,7 +178,6 @@ def _send_request(prompt, max_tokens=16, temperature=0.0):
183178
except Exception as e:
184179
return prompt, f"Error: {e}"
185180

186-
187181
try:
188182
for _ in range(300):
189183
try:
@@ -207,7 +201,7 @@ def _send_request(prompt, max_tokens=16, temperature=0.0):
207201
"paris is known for",
208202
"python is a",
209203
"chatgpt is",
210-
]*10
204+
] * 10
211205

212206
# 并发发送
213207
with ThreadPoolExecutor(max_workers=5) as executor:

0 commit comments

Comments
 (0)