We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 940afec commit b8e4387Copy full SHA for b8e4387
vllm/lora/layers.py
@@ -242,12 +242,12 @@ def forward(self, x: torch.Tensor) -> torch.Tensor:
242
embeddings_indices = torch.narrow(
243
self.punica_wrapper._embeddings_indices, 1, 0, x.size(0))
244
245
- indices = embeddings_indices[1].view_as(x)
+ indices = embeddings_indices[1]
246
full_lora_a_embeddings = F.embedding(
247
x + indices,
248
self.lora_a_stacked_2d,
249
)
250
- indices = embeddings_indices[0].view_as(x)
+ indices = embeddings_indices[0]
251
full_output = self.base_layer.forward(x +
252
(indices * added_tokens_mask))
253
0 commit comments