We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 10a2275 commit 1f731aeCopy full SHA for 1f731ae
vllm/model_executor/layers/quantization/kernels/mixed_precision/marlin.py
@@ -117,7 +117,7 @@ def apply_weights(self,
117
bias: Optional[torch.Tensor] = None) -> torch.Tensor:
118
# marlin requires contiguous memory layout
119
# kv/prefill caching may cause x to be non-contiguous
120
- x = x.contiguous() # no-op if already contiguous
+ x = x.contiguous() # no-op if already contiguous
121
122
c = self.config
123
w_q, w_s, w_zp, w_gidx = self._get_weight_params(layer)
0 commit comments