Skip to content

Commit 647b5ae

Browse files
DarkLight1337albertoperdomo2
authored andcommitted
[CI/Build] Remove unnecessary flags from test registry (vllm-project#27353)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Alberto Perdomo <aperdomo@redhat.com>
1 parent c7d0cb3 commit 647b5ae

File tree

13 files changed

+89
-123
lines changed

13 files changed

+89
-123
lines changed

docs/models/supported_models.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -374,8 +374,8 @@ th {
374374
| `GraniteMoeSharedForCausalLM` | Granite MoE Shared | `ibm-research/moe-7b-1b-active-shared-experts` (test model) | ✅︎ | ✅︎ |
375375
| `GritLM` | GritLM | `parasail-ai/GritLM-7B-vllm`. | ✅︎ | ✅︎ |
376376
| `Grok1ModelForCausalLM` | Grok1 | `hpcai-tech/grok-1`. | ✅︎ | ✅︎ |
377-
| `HunYuanDenseV1ForCausalLM` | Hunyuan-7B-Instruct-0124 | `tencent/Hunyuan-7B-Instruct-0124` | ✅︎ | ✅︎ |
378-
| `HunYuanMoEV1ForCausalLM` | Hunyuan-80B-A13B | `tencent/Hunyuan-A13B-Instruct`, `tencent/Hunyuan-A13B-Pretrain`, `tencent/Hunyuan-A13B-Instruct-FP8`, etc. | ✅︎ | ✅︎ |
377+
| `HunYuanDenseV1ForCausalLM` | Hunyuan Dense | `tencent/Hunyuan-7B-Instruct` | ✅︎ | ✅︎ |
378+
| `HunYuanMoEV1ForCausalLM` | Hunyuan-A13B | `tencent/Hunyuan-A13B-Instruct`, `tencent/Hunyuan-A13B-Pretrain`, `tencent/Hunyuan-A13B-Instruct-FP8`, etc. | ✅︎ | ✅︎ |
379379
| `HCXVisionForCausalLM` | HyperCLOVAX-SEED-Vision-Instruct-3B | `naver-hyperclovax/HyperCLOVAX-SEED-Vision-Instruct-3B` | | |
380380
| `InternLMForCausalLM` | InternLM | `internlm/internlm-7b`, `internlm/internlm-chat-7b`, etc. | ✅︎ | ✅︎ |
381381
| `InternLM2ForCausalLM` | InternLM2 | `internlm/internlm2-7b`, `internlm/internlm2-chat-7b`, etc. | ✅︎ | ✅︎ |

tests/distributed/test_pipeline_parallel.py

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -244,7 +244,7 @@ def _compare_tp(
244244
tokenizer_mode = model_info.tokenizer_mode
245245
hf_overrides = model_info.hf_overrides
246246
hf_config = get_config(model_id, trust_remote_code)
247-
skip_tokenizer_init = model_info.skip_tokenizer_init
247+
require_embed_inputs = model_info.require_embed_inputs
248248
max_num_seqs = model_info.max_num_seqs
249249

250250
dtype = "float16"
@@ -299,8 +299,14 @@ def _compare_tp(
299299
common_args.extend(["--load-format", load_format])
300300
if hf_overrides:
301301
common_args.extend(["--hf-overrides", json.dumps(hf_overrides)])
302-
if skip_tokenizer_init:
303-
common_args.append("--skip-tokenizer-init")
302+
if require_embed_inputs:
303+
common_args.extend(
304+
[
305+
"--skip-tokenizer-init",
306+
"--enable-prompt-embeds",
307+
"--enable-mm-embeds",
308+
]
309+
)
304310
if max_num_seqs:
305311
common_args.extend(["--max-num-seqs", f"{max_num_seqs}"])
306312

tests/distributed/test_sequence_parallel.py

Lines changed: 9 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -181,7 +181,7 @@ def _compare_sp(
181181
trust_remote_code = model_info.trust_remote_code
182182
tokenizer_mode = model_info.tokenizer_mode
183183
hf_overrides = model_info.hf_overrides
184-
skip_tokenizer_init = model_info.skip_tokenizer_init
184+
require_embed_inputs = model_info.require_embed_inputs
185185

186186
if load_format == "dummy":
187187
# Avoid OOM
@@ -233,8 +233,14 @@ def _compare_sp(
233233
common_args.extend(["--load-format", load_format])
234234
if hf_overrides:
235235
common_args.extend(["--hf-overrides", json.dumps(hf_overrides)])
236-
if skip_tokenizer_init:
237-
common_args.append("--skip-tokenizer-init")
236+
if require_embed_inputs:
237+
common_args.extend(
238+
[
239+
"--skip-tokenizer-init",
240+
"--enable-prompt-embeds",
241+
"--enable-mm-embeds",
242+
]
243+
)
238244

239245
compilation_config = {
240246
"mode": CompilationMode.VLLM_COMPILE,

tests/entrypoints/openai/test_chat_template.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,9 @@ def test_get_gen_prompt(
114114
trust_remote_code=model_info.trust_remote_code,
115115
revision=model_info.revision,
116116
hf_overrides=model_info.hf_overrides,
117-
skip_tokenizer_init=model_info.skip_tokenizer_init,
117+
skip_tokenizer_init=model_info.require_embed_inputs,
118+
enable_prompt_embeds=model_info.require_embed_inputs,
119+
enable_mm_embeds=model_info.require_embed_inputs,
118120
enforce_eager=model_info.enforce_eager,
119121
dtype=model_info.dtype,
120122
)

tests/entrypoints/test_chat_utils.py

Lines changed: 12 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1742,7 +1742,9 @@ def test_resolve_hf_chat_template(sample_json_schema, model, use_tools):
17421742
revision=model_info.revision,
17431743
trust_remote_code=model_info.trust_remote_code,
17441744
hf_overrides=model_info.hf_overrides,
1745-
skip_tokenizer_init=model_info.skip_tokenizer_init,
1745+
skip_tokenizer_init=model_info.require_embed_inputs,
1746+
enable_prompt_embeds=model_info.require_embed_inputs,
1747+
enable_mm_embeds=model_info.require_embed_inputs,
17461748
enforce_eager=model_info.enforce_eager,
17471749
dtype=model_info.dtype,
17481750
)
@@ -1842,7 +1844,9 @@ def test_resolve_hf_chat_template_kwargs(sample_json_schema, model, expected_kwa
18421844
revision=model_info.revision,
18431845
trust_remote_code=model_info.trust_remote_code,
18441846
hf_overrides=model_info.hf_overrides,
1845-
skip_tokenizer_init=model_info.skip_tokenizer_init,
1847+
skip_tokenizer_init=model_info.require_embed_inputs,
1848+
enable_prompt_embeds=model_info.require_embed_inputs,
1849+
enable_mm_embeds=model_info.require_embed_inputs,
18461850
enforce_eager=model_info.enforce_eager,
18471851
dtype=model_info.dtype,
18481852
)
@@ -1903,7 +1907,9 @@ def test_resolve_content_format_hf_defined(model, expected_format):
19031907
revision=model_info.revision,
19041908
trust_remote_code=model_info.trust_remote_code,
19051909
hf_overrides=model_info.hf_overrides,
1906-
skip_tokenizer_init=model_info.skip_tokenizer_init,
1910+
skip_tokenizer_init=model_info.require_embed_inputs,
1911+
enable_prompt_embeds=model_info.require_embed_inputs,
1912+
enable_mm_embeds=model_info.require_embed_inputs,
19071913
enforce_eager=model_info.enforce_eager,
19081914
dtype=model_info.dtype,
19091915
)
@@ -1961,7 +1967,9 @@ def test_resolve_content_format_fallbacks(model, expected_format):
19611967
revision=model_info.revision,
19621968
trust_remote_code=model_info.trust_remote_code,
19631969
hf_overrides=model_info.hf_overrides,
1964-
skip_tokenizer_init=model_info.skip_tokenizer_init,
1970+
skip_tokenizer_init=model_info.require_embed_inputs,
1971+
enable_prompt_embeds=model_info.require_embed_inputs,
1972+
enable_mm_embeds=model_info.require_embed_inputs,
19651973
enforce_eager=model_info.enforce_eager,
19661974
dtype=model_info.dtype,
19671975
)

tests/models/multimodal/generation/vlm_utils/core.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -71,8 +71,9 @@ def run_test(
7171
vllm_runner_kwargs_["tokenizer_mode"] = model_info.tokenizer_mode
7272
if model_info.hf_overrides:
7373
vllm_runner_kwargs_["hf_overrides"] = model_info.hf_overrides
74-
if model_info.skip_tokenizer_init:
75-
vllm_runner_kwargs_["skip_tokenizer_init"] = model_info.skip_tokenizer_init
74+
if model_info.require_embed_inputs:
75+
for k in ("skip_tokenizer_init", "enable_prompt_embeds", "enable_mm_embeds"):
76+
vllm_runner_kwargs_[k] = model_info.require_embed_inputs
7677

7778
if vllm_runner_kwargs:
7879
vllm_runner_kwargs_.update(vllm_runner_kwargs)

tests/models/multimodal/processing/test_common.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,9 @@ def _test_processing_correctness(
108108
hf_overrides=model_info.hf_overrides,
109109
# Ensure that the cache can fit all of the data
110110
mm_processor_cache_gb=2048,
111-
skip_tokenizer_init=model_info.skip_tokenizer_init,
111+
skip_tokenizer_init=model_info.require_embed_inputs,
112+
enable_prompt_embeds=model_info.require_embed_inputs,
113+
enable_mm_embeds=model_info.require_embed_inputs,
112114
enforce_eager=model_info.enforce_eager,
113115
dtype=model_info.dtype,
114116
)

tests/models/multimodal/processing/test_tensor_schema.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -218,7 +218,9 @@ def test_model_tensor_schema(model_arch: str, model_id: str):
218218
revision=model_info.revision,
219219
trust_remote_code=model_info.trust_remote_code,
220220
hf_overrides=hf_overrides_fn,
221-
skip_tokenizer_init=model_info.skip_tokenizer_init,
221+
skip_tokenizer_init=model_info.require_embed_inputs,
222+
enable_prompt_embeds=model_info.require_embed_inputs,
223+
enable_mm_embeds=model_info.require_embed_inputs,
222224
enforce_eager=model_info.enforce_eager,
223225
dtype=model_info.dtype,
224226
)

tests/models/multimodal/test_mapping.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,9 @@ def test_hf_model_weights_mapper(model_arch: str):
5959
revision=model_info.revision,
6060
trust_remote_code=model_info.trust_remote_code,
6161
hf_overrides=model_info.hf_overrides,
62-
skip_tokenizer_init=model_info.skip_tokenizer_init,
62+
skip_tokenizer_init=model_info.require_embed_inputs,
63+
enable_prompt_embeds=model_info.require_embed_inputs,
64+
enable_mm_embeds=model_info.require_embed_inputs,
6365
enforce_eager=model_info.enforce_eager,
6466
dtype=model_info.dtype,
6567
)

0 commit comments

Comments
 (0)