Skip to content

Commit 320feae

Browse files
authored
[Model] Lfm2Moe (#26344)
Signed-off-by: Paul Pak <paulpak58@gmail.com>
1 parent 1e4ecca commit 320feae

File tree

8 files changed

+967
-7
lines changed

8 files changed

+967
-7
lines changed

docs/models/supported_models.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -390,6 +390,7 @@ th {
390390
| `JAISLMHeadModel` | Jais | `inceptionai/jais-13b`, `inceptionai/jais-13b-chat`, `inceptionai/jais-30b-v3`, `inceptionai/jais-30b-chat-v3`, etc. | | ✅︎ | ✅︎ |
391391
| `JambaForCausalLM` | Jamba | `ai21labs/AI21-Jamba-1.5-Large`, `ai21labs/AI21-Jamba-1.5-Mini`, `ai21labs/Jamba-v0.1`, etc. | ✅︎ | ✅︎ | ✅︎ |
392392
| `Lfm2ForCausalLM` | LFM2 | `LiquidAI/LFM2-1.2B`, `LiquidAI/LFM2-700M`, `LiquidAI/LFM2-350M`, etc. | ✅︎ | ✅︎ | ✅︎ |
393+
| `Lfm2MoeForCausalLM` | LFM2MoE | `LiquidAI/LFM2-8B-A1B-preview`, etc. | ✅︎ | ✅︎ | ✅︎ |
393394
| `LlamaForCausalLM` | Llama 3.1, Llama 3, Llama 2, LLaMA, Yi | `meta-llama/Meta-Llama-3.1-405B-Instruct`, `meta-llama/Meta-Llama-3.1-70B`, `meta-llama/Meta-Llama-3-70B-Instruct`, `meta-llama/Llama-2-70b-hf`, `01-ai/Yi-34B`, etc. | ✅︎ | ✅︎ | ✅︎ |
394395
| `MambaForCausalLM` | Mamba | `state-spaces/mamba-130m-hf`, `state-spaces/mamba-790m-hf`, `state-spaces/mamba-2.8b-hf`, etc. | | ✅︎ | ✅︎ |
395396
| `Mamba2ForCausalLM` | Mamba2 | `mistralai/Mamba-Codestral-7B-v0.1`, etc. | | ✅︎ | ✅︎ |

tests/models/registry.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -321,6 +321,9 @@ def check_available_online(
321321
"Lfm2ForCausalLM": _HfExamplesInfo(
322322
"LiquidAI/LFM2-1.2B", min_transformers_version="4.54"
323323
),
324+
"Lfm2MoeForCausalLM": _HfExamplesInfo(
325+
"LiquidAI/LFM2-8B-A1B", min_transformers_version="4.58"
326+
),
324327
"LlamaForCausalLM": _HfExamplesInfo(
325328
"meta-llama/Llama-3.2-1B-Instruct",
326329
extras={

vllm/model_executor/models/lfm2.py

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -71,14 +71,14 @@ def __init__(
7171
output_sizes=[ff_dim] * 2,
7272
bias=False,
7373
quant_config=quant_config,
74-
prefix=f"{prefix}.gate_up_proj",
74+
prefix=f"{prefix}.w1",
7575
)
7676
self.w2 = RowParallelLinear(
7777
input_size=ff_dim,
7878
output_size=dim,
7979
bias=False,
8080
quant_config=quant_config,
81-
prefix=f"{prefix}.down_proj",
81+
prefix=f"{prefix}.w2",
8282
)
8383
self.act_fn = SiluAndMul()
8484

@@ -484,17 +484,12 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None:
484484
quant_config = vllm_config.quant_config
485485
cache_config = vllm_config.cache_config
486486
lora_config = vllm_config.lora_config
487-
scheduler_config = vllm_config.scheduler_config
488487
assert not cache_config.enable_prefix_caching, (
489488
"Lfm2 currently does not support prefix caching"
490489
)
491490

492491
super().__init__()
493492
self.config = config
494-
self.vllm_config = vllm_config
495-
self.scheduler_config = scheduler_config
496-
self.model_config = vllm_config.model_config
497-
498493
self.model = Lfm2Model(
499494
vllm_config=vllm_config, prefix=maybe_prefix(prefix, "model")
500495
)

0 commit comments

Comments
 (0)