We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 506f798 commit 44ef111Copy full SHA for 44ef111
vllm/attention/layer.py
@@ -532,7 +532,6 @@ def unified_attention_with_output(
532
# Not all layers can use RoPE fusing, so check that they were given all
533
# needed inputs along with the environment variable to enable this.
534
if (
535
-
536
and hasattr(self.impl, "rotary_emb")
537
and self.impl.rotary_emb is not None
538
and positions is not None
0 commit comments