Skip to content

Commit 2ca339a

Browse files
committed
[XPU][bugfix] fix rope for llama4 and deepseek
Signed-off-by: Yan Ma <yan.ma@intel.com>
1 parent 4fb8771 commit 2ca339a

File tree

2 files changed

+5
-21
lines changed

2 files changed

+5
-21
lines changed

vllm/model_executor/layers/rotary_embedding/deepseek_scaling_rope.py

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -146,11 +146,5 @@ def forward_native(
146146
key = key_rot
147147
return query, key
148148

149-
def forward_cuda(
150-
self,
151-
positions: torch.Tensor,
152-
query: torch.Tensor,
153-
key: torch.Tensor | None = None,
154-
offsets: torch.Tensor | None = None,
155-
) -> tuple[torch.Tensor, torch.Tensor | None]:
156-
return self.forward_native(positions, query, key, offsets)
149+
forward_cuda = forward_native
150+
forward_xpu = forward_native

vllm/model_executor/layers/rotary_embedding/llama4_vision_rope.py

Lines changed: 3 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -72,16 +72,6 @@ def forward_native( # type: ignore[override]
7272
key_out = torch.view_as_real(key_ * freqs_ci).flatten(3)
7373
return query_out.type_as(query), key_out.type_as(key)
7474

75-
def forward_cuda( # type: ignore[override]
76-
self,
77-
query: torch.Tensor,
78-
key: torch.Tensor | None = None,
79-
) -> tuple[torch.Tensor, torch.Tensor | None]:
80-
return self.forward_native(query, key)
81-
82-
def forward_hip( # type: ignore[override]
83-
self,
84-
query: torch.Tensor,
85-
key: torch.Tensor | None = None,
86-
) -> tuple[torch.Tensor, torch.Tensor | None]:
87-
return self.forward_native(query, key)
75+
forward_cuda = forward_native
76+
forward_hip = forward_native
77+
forward_xpu = forward_native

0 commit comments

Comments
 (0)