Skip to content

Commit 6ad393f

Browse files
yeqcharlottehouseroad
authored andcommitted
clean up model names and whitespaces (#120)
Signed-off-by: Ye (Charlotte) Qi <yeq@meta.com> Signed-off-by: Lu Fang <lufang@fb.com>
1 parent 188bb52 commit 6ad393f

File tree

8 files changed

+8
-18
lines changed

8 files changed

+8
-18
lines changed

examples/offline_inference/vision_language.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -585,8 +585,7 @@ def run_mllama(questions: list[str], modality: str) -> ModelRequestData:
585585
def run_llama4(questions: list[str], modality: str):
586586
assert modality == "image"
587587

588-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
589-
model_name = "ll-re/Llama-4-Scout-17B-16E-Instruct"
588+
model_name = "meta-llama/Llama-4-Scout-17B-16E-Instruct"
590589

591590
engine_args = EngineArgs(
592591
model=model_name,

examples/offline_inference/vision_language_multi_image.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -254,8 +254,7 @@ def load_internvl(question: str, image_urls: list[str]) -> ModelRequestData:
254254

255255

256256
def load_llama4(question: str, image_urls: list[str]) -> ModelRequestData:
257-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
258-
model_name = "ll-re/Llama-4-Scout-17B-16E-Instruct"
257+
model_name = "meta-llama/Llama-4-Scout-17B-16E-Instruct"
259258

260259
engine_args = EngineArgs(
261260
model=model_name,

tests/models/decoder_only/vision_language/test_models.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -537,8 +537,7 @@
537537
)],
538538
),
539539
"llama4": VLMTestInfo(
540-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
541-
models=["ll-re/Llama-4-Scout-17B-16E-Instruct"],
540+
models=["meta-llama/Llama-4-Scout-17B-16E-Instruct"],
542541
prompt_formatter=lambda img_prompt: f"<|begin_of_text|><|header_start|>user<|header_end|>\n\n{img_prompt}<|eot|><|header_start|>assistant<|header_end|>\n\n", # noqa: E501
543542
img_idx_to_prompt=lambda _: "<|image|>",
544543
test_type=(VLMTestType.IMAGE, VLMTestType.MULTI_IMAGE),

tests/models/multimodal/processing/test_common.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -280,8 +280,7 @@ def _test_processing_correctness_mistral(
280280
"Skywork/Skywork-R1V-38B",
281281
"fixie-ai/ultravox-v0_5-llama-3_2-1b",
282282
"openai/whisper-large-v3",
283-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
284-
"ll-re/Llama-4-Scout-17B-16E-Instruct",
283+
"meta-llama/Llama-4-Scout-17B-16E-Instruct",
285284
])
286285
@pytest.mark.parametrize("hit_rate", [0.3, 0.5, 1.0])
287286
@pytest.mark.parametrize("num_batches", [32])

tests/models/multimodal/processing/test_llama4.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@
1010
from ...utils import build_model_context
1111

1212

13-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
14-
@pytest.mark.parametrize("model_id", ["ll-re/Llama-4-Scout-17B-16E-Instruct"])
13+
@pytest.mark.parametrize("model_id",
14+
["meta-llama/Llama-4-Scout-17B-16E-Instruct"])
1515
@pytest.mark.parametrize("mm_processor_kwargs", [{}])
1616
@pytest.mark.parametrize("num_imgs", [1, 5])
1717
@pytest.mark.parametrize("disable_mm_preprocessor_cache", [True, False])

tests/models/registry.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -337,8 +337,7 @@ def check_available_online(
337337
tokenizer="facebook/bart-base",
338338
trust_remote_code=True), # noqa: E501
339339
"MllamaForConditionalGeneration": _HfExamplesInfo("meta-llama/Llama-3.2-11B-Vision-Instruct"), # noqa: E501
340-
# FIXME: meta-llama/Llama-4-Scout-17B-16E-Instruct
341-
"Llama4ForConditionalGeneration": _HfExamplesInfo("ll-re/Llama-4-Scout-17B-16E-Instruct"), # noqa: E501
340+
"Llama4ForConditionalGeneration": _HfExamplesInfo("meta-llama/Llama-4-Scout-17B-16E-Instruct"), # noqa: E501
342341
"WhisperForConditionalGeneration": _HfExamplesInfo("openai/whisper-large-v3"), # noqa: E501
343342
}
344343

vllm/model_executor/layers/fused_moe/cutlass_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -103,7 +103,7 @@ def cutlass_moe_fp8(
103103
per_act_token = a1_scale.numel() != 1 if a1_scale is not None else (
104104
a2_scale.numel() != 1 if a2_scale is not None else False)
105105
if apply_router_weight_on_input:
106-
# FIXME: this only works for topK=1, will need to update for topK>1
106+
# TODO: this only works for topK=1, will need to update for topK>1
107107
a = a * topk_weights.to(out_dtype)
108108

109109
a_q, a1_scale = ops.scaled_fp8_quant(

vllm/model_executor/models/mllama4.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -310,9 +310,7 @@ def forward(
310310
):
311311
# Self Attention
312312
residual = hidden_state
313-
314313
hidden_state = self.input_layernorm(hidden_state)
315-
316314
hidden_state = self.self_attn(hidden_state)
317315
hidden_state = residual + hidden_state
318316

@@ -463,15 +461,12 @@ def forward(
463461
positional_embedding = self.positional_embedding_vlm.to(
464462
dtype=hidden_state.dtype, device=hidden_state.device)
465463
hidden_state = hidden_state + positional_embedding
466-
467464
hidden_state = self.layernorm_pre(hidden_state)
468-
469465
hidden_state = hidden_state.view(num_tiles, -1, hidden_dim)
470466

471467
# Apply encoder
472468
output = self.model(hidden_state)
473469
hidden_state = output.last_hidden_state
474-
475470
hidden_state = self.layernorm_post(hidden_state)
476471

477472
# Remove CLS token output

0 commit comments

Comments
 (0)