11# SPDX-License-Identifier: Apache-2.0
22
3+ import pytest
4+
35import vllm
46from vllm .lora .request import LoRARequest
7+ from vllm .platforms import current_platform
58
69from ..utils import create_new_process_for_each_test , multi_gpu_test
710
@@ -44,7 +47,6 @@ def do_sample(llm: vllm.LLM, lora_path: str, lora_id: int) -> list[str]:
4447 return generated_texts
4548
4649
47- @create_new_process_for_each_test ()
4850def test_ilama_lora (ilama_lora_files ):
4951 llm = vllm .LLM (MODEL_PATH ,
5052 max_model_len = 1024 ,
@@ -63,6 +65,8 @@ def test_ilama_lora(ilama_lora_files):
6365 assert output2 [i ] == EXPECTED_LORA_OUTPUT [i ]
6466
6567
68+ @pytest .mark .skipif (current_platform .is_cuda_alike (),
69+ reason = "Skipping to avoid redundant model tests" )
6670@multi_gpu_test (num_gpus = 4 )
6771@create_new_process_for_each_test ()
6872def test_ilama_lora_tp4 (ilama_lora_files ):
@@ -84,6 +88,8 @@ def test_ilama_lora_tp4(ilama_lora_files):
8488 assert output2 [i ] == EXPECTED_LORA_OUTPUT [i ]
8589
8690
91+ @pytest .mark .skipif (current_platform .is_cuda_alike (),
92+ reason = "Skipping to avoid redundant model tests" )
8793@multi_gpu_test (num_gpus = 4 )
8894@create_new_process_for_each_test ()
8995def test_ilama_lora_tp4_fully_sharded_loras (ilama_lora_files ):
0 commit comments