Skip to content

Commit c13b334

Browse files
refactor common_fused_moe.py
Co-Authored-By: weijinqian0 <12153182+weijinqian0@users.noreply.github.com> Signed-off-by: weichen <calvin_zhu0210@outlook.com> Signed-off-by: Pr0Wh1teGivee <calvin_zhu0210@outlook.com>
1 parent 2693196 commit c13b334

21 files changed

+1050
-932
lines changed

tests/e2e/singlecard/ops/test_fused_moe.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -28,9 +28,8 @@
2828
import torch_npu
2929
from vllm.model_executor.layers.activation import SiluAndMul
3030

31-
from vllm_ascend.ops.layers.experts_selector import select_experts
32-
from vllm_ascend.ops.moe_dispatcher.token_dispatcher import \
33-
TokenDispatcherWithAllGather
31+
from vllm_ascend.ops.moe.experts_selector import select_experts
32+
from vllm_ascend.ops.moe.token_dispatcher import TokenDispatcherWithAllGather
3433

3534
NUM_EXPERTS = [8, 64]
3635
EP_SIZE = [1]
@@ -209,7 +208,7 @@ def test_select_experts(
209208
dtype=torch.int32)
210209
custom_routing_function.return_value = (mock_weights, mock_ids)
211210

212-
with patch("vllm_ascend.ops.layers.experts_selector._native_grouped_topk"
211+
with patch("vllm_ascend.ops.moe.experts_selector._native_grouped_topk"
213212
) as mock_native_grouped_topk:
214213
mock_native_grouped_topk.side_effect = lambda x, num_groups, k: torch.randn_like(
215214
x)

tests/e2e/singlecard/ops/test_moe_comm.py

Lines changed: 0 additions & 175 deletions
This file was deleted.

0 commit comments

Comments
 (0)