We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 19a53b2 commit f04d604Copy full SHA for f04d604
vllm/attention/layer.py
@@ -209,7 +209,7 @@ def forward(
209
if self.use_output:
210
output_shape = (output_shape
211
if output_shape is not None else query.shape)
212
- output = torch.empty(output_shape,
+ output = torch.zeros(output_shape,
213
dtype=query.dtype,
214
device=query.device)
215
hidden_size = output_shape[-1]
0 commit comments