@@ -268,7 +268,13 @@ jobs:
268268 --ignore=tests/e2e/singlecard/test_ilama_lora.py \
269269 --ignore=tests/e2e/singlecard/test_guided_decoding.py \
270270 --ignore=tests/e2e/singlecard/test_camem.py \
271- --ignore=tests/e2e/singlecard/test_embedding.py
271+ --ignore=tests/e2e/singlecard/test_embedding.py \
272+ --ignore=tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_correctness.py \
273+ --ignore=tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
274+ # ------------------------------------ v1 spec decode test ------------------------------------ #
275+ VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_correctness.py
276+ # TODO: revert me when test_v1_spec_decode.py::test_ngram_correctness is fixed
277+ VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
272278
273279 - name : Run e2e test on V0 engine
274280 if : ${{ github.event_name == 'schedule' }}
@@ -290,8 +296,6 @@ jobs:
290296 --ignore=tests/e2e/singlecard/test_guided_decoding.py \
291297 --ignore=tests/e2e/singlecard/test_camem.py \
292298 --ignore=tests/e2e/singlecard/test_prompt_embedding.py \
293- --ignore=tests/e2e/singlecard/core/test_ascend_scheduler.py \
294- --ignore=tests/e2e/singlecard/core/test_ascend_scheduler_e2e.py \
295299 --ignore=tests/e2e/singlecard/test_embedding.py
296300
297301 e2e-4-cards :
@@ -364,7 +368,6 @@ jobs:
364368 # To avoid oom, we need to run the test in a single process.
365369 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_multistream_moe
366370 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
367- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek
368371 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_topk
369372 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
370373 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_dbo
@@ -386,7 +389,6 @@ jobs:
386389 # Fixme: run VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py will raise error.
387390 # To avoid oom, we need to run the test in a single process.
388391 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
389- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek
390392 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_topk
391393 pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W8A8
392394 pytest -sv tests/e2e/multicard/test_data_parallel.py
0 commit comments