Skip to content

Commit 1f731ae

Browse files
committed
format
Signed-off-by: Qubitium <qubitium@modelcloud.ai>
1 parent 10a2275 commit 1f731ae

File tree

1 file changed

+1
-1
lines changed
  • vllm/model_executor/layers/quantization/kernels/mixed_precision

1 file changed

+1
-1
lines changed

vllm/model_executor/layers/quantization/kernels/mixed_precision/marlin.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -117,7 +117,7 @@ def apply_weights(self,
117117
bias: Optional[torch.Tensor] = None) -> torch.Tensor:
118118
# marlin requires contiguous memory layout
119119
# kv/prefill caching may cause x to be non-contiguous
120-
x = x.contiguous() # no-op if already contiguous
120+
x = x.contiguous() # no-op if already contiguous
121121

122122
c = self.config
123123
w_q, w_s, w_zp, w_gidx = self._get_weight_params(layer)

0 commit comments

Comments
 (0)