Skip to content

Commit e1ba235

Browse files
[BugFix] Fix failing test quantization/test_compressed_tensors.py::test_compressed_tensors_fp8_block_enabled (#26436)
Signed-off-by: morrison-turnansky <mturnans@redhat.com>
1 parent b82f430 commit e1ba235

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/config/vllm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -597,7 +597,7 @@ def has_blocked_weights():
597597
# https://github.com/vllm-project/vllm/issues/25094
598598
if has_blocked_weights():
599599
custom_ops = self.compilation_config.custom_ops
600-
if "none" not in custom_ops and "-quant_fp8" not in custom_ops:
600+
if "-quant_fp8" not in custom_ops:
601601
custom_ops.append("+quant_fp8")
602602

603603
def update_sizes_for_sequence_parallelism(self, possible_sizes: list) -> list:

0 commit comments

Comments
 (0)