We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f58f8b5 commit e073438Copy full SHA for e073438
vllm/model_executor/layers/quantization/moe_wna16.py
@@ -293,9 +293,10 @@ def apply(
293
custom_routing_function: Optional[Callable] = None,
294
scoring_func: str = "softmax",
295
e_score_correction_bias: Optional[torch.Tensor] = None,
296
+ activation: str = "silu",
297
) -> torch.Tensor:
298
from vllm.model_executor.layers.fused_moe import fused_experts
-
299
+ assert activation == "silu", "Only SiLU activation is supported."
300
topk_weights, topk_ids = FusedMoE.select_experts(
301
hidden_states=x,
302
router_logits=router_logits,
0 commit comments