@@ -299,6 +299,7 @@ steps:
299299 commands :
300300 - pytest -v -s compile/test_pass_manager.py
301301 - pytest -v -s compile/test_fusion.py
302+ - pytest -v -s compile/test_sequence_parallelism.py
302303
303304- label : PyTorch Fullgraph Smoke Test # 9min
304305 source_file_dependencies :
@@ -318,13 +319,15 @@ steps:
318319 - pytest -v -s compile/test_full_graph.py
319320
320321- label : Kernels Core Operation Test
322+ mirror_hardwares : [amd]
321323 source_file_dependencies :
322324 - csrc/
323325 - tests/kernels/core
324326 commands :
325327 - pytest -v -s kernels/core
326328
327329- label : Kernels Attention Test %N
330+ mirror_hardwares : [amd]
328331 source_file_dependencies :
329332 - csrc/attention/
330333 - vllm/attention
@@ -335,6 +338,7 @@ steps:
335338 parallelism : 2
336339
337340- label : Kernels Quantization Test %N
341+ mirror_hardwares : [amd]
338342 source_file_dependencies :
339343 - csrc/quantization/
340344 - vllm/model_executor/layers/quantization
@@ -344,6 +348,7 @@ steps:
344348 parallelism : 2
345349
346350- label : Kernels MoE Test
351+ # mirror_hardwares: [amd]
347352 source_file_dependencies :
348353 - csrc/moe/
349354 - tests/kernels/moe
@@ -352,6 +357,7 @@ steps:
352357 - pytest -v -s kernels/moe
353358
354359- label : Kernels Mamba Test
360+ # mirror_hardwares: [amd]
355361 source_file_dependencies :
356362 - csrc/mamba/
357363 - tests/kernels/mamba
@@ -384,12 +390,13 @@ steps:
384390 commands :
385391 - pytest -v -s benchmarks/
386392
387- - label : Quantization Test # 33min
393+ - label : Quantization Test
388394 source_file_dependencies :
389395 - csrc/
390396 - vllm/model_executor/layers/quantization
391397 - tests/quantization
392- command : VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
398+ commands :
399+ - VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
393400
394401- label : LM Eval Small Models # 53min
395402 working_dir : " /vllm-workspace/.buildkite/lm-eval-harness"
@@ -435,82 +442,78 @@ steps:
435442 commands :
436443 - pytest -v -s models/test_transformers.py
437444 - pytest -v -s models/test_registry.py
445+ - pytest -v -s models/test_utils.py
446+ - pytest -v -s models/test_vision.py
438447 # V1 Test: https://github.com/vllm-project/vllm/issues/14531
439448 - VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2'
440449 - VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'llama4'
441450 - VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'plamo2'
442451
443- - label : Language Models Test (Standard) # 32min
452+ - label : Language Models Test (Standard)
444453 # mirror_hardwares: [amd]
445454 source_file_dependencies :
446455 - vllm/
447- - tests/models/decoder_only/language
448- - tests/models/embedding/language
449- - tests/models/encoder_decoder/language
456+ - tests/models/language
450457 commands :
451458 # Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
452- - pip install causal-conv1d
453- - pytest -v -s models/decoder_only/language -m 'core_model or quant_model'
454- - pytest -v -s models/embedding/language -m core_model
459+ - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8'
460+ - pytest -v -s models/language -m core_model
455461
456- - label : Language Models Test (Extended) # 1h10min
462+ - label : Language Models Test (Extended)
457463 optional : true
458464 source_file_dependencies :
459465 - vllm/
460- - tests/models/decoder_only/language
461- - tests/models/embedding/language
462- - tests/models/encoder_decoder/language
466+ - tests/models/language
463467 commands :
464468 # Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
465- - pip install causal-conv1d
466- - pytest -v -s models/decoder_only/language -m 'not core_model and not quant_model'
467- - pytest -v -s models/embedding/language -m 'not core_model'
469+ - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8'
470+ - pytest -v -s models/language -m 'not core_model'
468471
469- - label : Multi-Modal Models Test (Standard) # 40min
472+ - label : Multi-Modal Models Test (Standard)
470473 # mirror_hardwares: [amd]
471474 source_file_dependencies :
472475 - vllm/
473- - tests/models/decoder_only/audio_language
474- - tests/models/decoder_only/vision_language
475- - tests/models/embedding/vision_language
476- - tests/models/encoder_decoder/audio_language
477- - tests/models/encoder_decoder/vision_language
476+ - tests/models/multimodal
478477 commands :
479478 - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
480- - pytest -v -s models/multimodal
481- - pytest -v -s models/decoder_only/audio_language -m 'core_model or quant_model'
482- - pytest -v -s models/decoder_only/vision_language -m 'core_model or quant_model'
483- - pytest -v -s models/embedding/vision_language -m core_model
484- - pytest -v -s models/encoder_decoder/audio_language -m core_model
485- - pytest -v -s models/encoder_decoder/language -m core_model
486- - pytest -v -s models/encoder_decoder/vision_language -m core_model
487- - pytest -v -s models/decoder_only/vision_language/test_interleaved.py
488-
489- - label : Multi-Modal Models Test (Extended) 1 # 48m
479+ - pytest -v -s models/multimodal/processing
480+ - pytest -v -s --ignore models/multimodal/generation/test_whisper.py models/multimodal -m core_model
481+ - cd .. && pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work
482+
483+ - label : Multi-Modal Models Test (Extended) 1
490484 optional : true
491485 source_file_dependencies :
492486 - vllm/
493- - tests/models/decoder_only/audio_language
494- - tests/models/decoder_only/vision_language
495- - tests/models/embedding/vision_language
496- - tests/models/encoder_decoder/vision_language
487+ - tests/models/multimodal
497488 commands :
498489 - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
499- - pytest -v -s models/decoder_only/audio_language -m 'not core_model and not quant_model'
500- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=0) and not core_model and not quant_model'
501- - pytest -v -s --ignore models/decoder_only/vision_language/test_models.py models/decoder_only/vision_language -m 'not core_model and not quant_model'
502- - pytest -v -s models/embedding/vision_language -m 'not core_model'
503- - pytest -v -s models/encoder_decoder/language -m 'not core_model'
504- - pytest -v -s models/encoder_decoder/vision_language -m 'not core_model'
505-
506- - label : Multi-Modal Models Test (Extended) 2 # 38m
490+ - pytest -v -s --ignore models/multimodal/generation/test_common.py --ignore models/multimodal/processing models/multimodal -m 'not core_model'
491+
492+ - label : Multi-Modal Models Test (Extended) 2
507493 optional : true
508494 source_file_dependencies :
509495 - vllm/
510- - tests/models/decoder_only/vision_language
496+ - tests/models/multimodal
511497 commands :
512498 - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
513- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=1) and not core_model and not quant_model'
499+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=0) and not core_model'
500+
501+ - label : Multi-Modal Models Test (Extended) 3
502+ optional : true
503+ source_file_dependencies :
504+ - vllm/
505+ - tests/models/multimodal
506+ commands :
507+ - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
508+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=1) and not core_model'
509+
510+ - label : Quantized Models Test
511+ # mirror_hardwares: [amd]
512+ source_file_dependencies :
513+ - vllm/model_executor/layers/quantization
514+ - tests/models/quantization
515+ commands :
516+ - pytest -v -s models/quantization
514517
515518# This test is used only in PR development phase to test individual models and should never run on main
516519- label : Custom Models Test
@@ -580,9 +583,10 @@ steps:
580583 - TARGET_TEST_SUITE=L4 pytest basic_correctness/ -v -s -m 'distributed(num_gpus=2)'
581584 # Avoid importing model tests that cause CUDA reinitialization error
582585 - pytest models/test_transformers.py -v -s -m 'distributed(num_gpus=2)'
583- - pytest models/encoder_decoder/language/test_bart.py -v -s -m 'distributed(num_gpus=2)'
584- - pytest models/encoder_decoder/vision_language/test_broadcast.py -v -s -m 'distributed(num_gpus=2)'
585- - pytest models/decoder_only/vision_language/test_models.py -v -s -m 'distributed(num_gpus=2)'
586+ - pytest models/language -v -s -m 'distributed(num_gpus=2)'
587+ - pytest models/multimodal -v -s -m 'distributed(num_gpus=2)'
588+ # test sequence parallel
589+ - pytest -v -s distributed/test_sequence_parallel.py
586590 # this test fails consistently.
587591 # TODO: investigate and fix
588592 # - pytest -v -s spec_decode/e2e/test_integration_dist_tp2.py
0 commit comments