Skip to content

Commit 293c004

Browse files
committed
[CI/UT][Refactor] Some refactors on UT
* move e2e spec decode and deepseek acc test to per pr * move test_fused_moe_allgather_ep.py to e2e/multicard * remove e2e test on deepseek-v2-lite due to already test acc Signed-off-by: MengqingCao <cmq0113@163.com>
1 parent 99e6855 commit 293c004

File tree

7 files changed

+7
-22
lines changed

7 files changed

+7
-22
lines changed

.github/workflows/vllm_ascend_test.yaml

Lines changed: 7 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -266,7 +266,13 @@ jobs:
266266
--ignore=tests/e2e/singlecard/test_offline_inference.py \
267267
--ignore=tests/e2e/singlecard/test_ilama_lora.py \
268268
--ignore=tests/e2e/singlecard/test_guided_decoding.py \
269-
--ignore=tests/e2e/singlecard/test_camem.py
269+
--ignore=tests/e2e/singlecard/test_camem.py \
270+
--ignore=tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_correctness.py \
271+
--ignore=tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
272+
# ------------------------------------ v1 spec decode test ------------------------------------ #
273+
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_correctness.py
274+
# TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
275+
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
270276
271277
- name: Run e2e test on V0 engine
272278
if: ${{ github.event_name == 'schedule' }}
@@ -360,7 +366,6 @@ jobs:
360366
# To avoid oom, we need to run the test in a single process.
361367
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_multistream_moe
362368
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
363-
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek
364369
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_topk
365370
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
366371
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_dbo
@@ -382,7 +387,6 @@ jobs:
382387
# Fixme: run VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py will raise error.
383388
# To avoid oom, we need to run the test in a single process.
384389
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
385-
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek
386390
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_topk
387391
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
388392
pytest -sv tests/e2e/multicard/test_data_parallel.py

.github/workflows/vllm_ascend_test_long_term.yaml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -100,10 +100,6 @@ jobs:
100100
# v0 spec decode test
101101
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/long_term/spec_decode_v0/e2e/test_mtp_correctness.py # it needs a clean process
102102
pytest -sv tests/e2e/long_term/spec_decode_v0 --ignore=tests/e2e/long_term/spec_decode_v0/e2e/test_mtp_correctness.py
103-
# v1 spec decode test
104-
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/long_term/spec_decode_v1/test_v1_mtp_correctness.py
105-
# TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
106-
VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/long_term/spec_decode_v1/test_v1_spec_decode.py
107103
# accuracy test single card
108104
pytest -sv tests/e2e/long_term/test_accuracy.py
109105
else

tests/e2e/multicard/test_offline_inference_distributed.py

Lines changed: 0 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -73,21 +73,6 @@ def test_models_distributed_DeepSeek_multistream_moe():
7373
vllm_model.generate_greedy(example_prompts, max_tokens)
7474

7575

76-
def test_models_distributed_DeepSeek():
77-
example_prompts = [
78-
"Hello, my name is",
79-
]
80-
dtype = "half"
81-
max_tokens = 5
82-
with VllmRunner(
83-
"deepseek-ai/DeepSeek-V2-Lite",
84-
dtype=dtype,
85-
tensor_parallel_size=4,
86-
distributed_executor_backend="mp",
87-
) as vllm_model:
88-
vllm_model.generate_greedy(example_prompts, max_tokens)
89-
90-
9176
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_TOPK_OPTIMIZE": "1"})
9277
def test_models_distributed_topk() -> None:
9378
example_prompts = [

0 commit comments

Comments
 (0)