We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent c7f3b7d commit ad1fb75Copy full SHA for ad1fb75
tests/e2e/multicard/test_offline_inference_distributed.py
@@ -72,6 +72,21 @@ def test_models_distributed_topk() -> None:
72
vllm_model.generate(example_prompts, sampling_params)
73
74
75
+def test_models_distributed_DeepSeek():
76
+ example_prompts = [
77
+ "Hello, my name is",
78
+ ]
79
+ dtype = "half"
80
+ max_tokens = 5
81
+ with VllmRunner(
82
+ "deepseek-ai/DeepSeek-V2-Lite",
83
+ dtype=dtype,
84
+ tensor_parallel_size=4,
85
+ distributed_executor_backend="mp",
86
+ ) as vllm_model:
87
+ vllm_model.generate_greedy(example_prompts, max_tokens)
88
+
89
90
@patch.dict(os.environ, {"VLLM_ASCEND_ENABLE_DBO": "1"})
91
def test_models_distributed_DeepSeekV3_dbo():
92
example_prompts = ["The president of the United States is"] * 41
0 commit comments