Skip to content

Commit 6f33557

Browse files
committed
update
1 parent a31d222 commit 6f33557

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

vllm_ascend/ops/moe/moe_mlp.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -227,7 +227,8 @@ def unified_apply_mlp(hidden_states: torch.Tensor,
227227
w2_scale_bias: torch.Tensor = None,
228228
topk_scales: Optional[torch.Tensor] = None,
229229
with_quant: bool = False,
230-
fusion: bool = False) -> torch.Tensor:
230+
fusion: bool = False,
231+
need_trans: bool = True) -> torch.Tensor:
231232
if with_quant:
232233
return quant_apply_mlp(hidden_states=hidden_states,
233234
w1=w1,

vllm_ascend/quantization/w8a8_dynamic.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
from vllm_ascend.ascend_forward_context import FusedMoEState
2929
from vllm_ascend.distributed.parallel_state import get_mc2_group
3030
from vllm_ascend.ops.fused_moe import unified_fused_experts_eager
31-
from vllm_ascend.ops.layers.experts_selector import select_experts
31+
from vllm_ascend.ops.moe.experts_selector import select_experts
3232
from vllm_ascend.utils import ACL_FORMAT_FRACTAL_NZ
3333

3434

0 commit comments

Comments
 (0)