File tree Expand file tree Collapse file tree 1 file changed +8
-6
lines changed
vllm/model_executor/layers/fused_moe Expand file tree Collapse file tree 1 file changed +8
-6
lines changed Original file line number Diff line number Diff line change @@ -741,12 +741,14 @@ def __init__(
741741
742742 # we padding globally so EP buffer allocation works
743743 if quant_config and quant_config .get_name () == "mxfp4" :
744- if not is_torch_equal_or_newer ("2.8.0" ):
745- raise RuntimeError ("Mxfp4 on hopper requires torch >= 2.8.0" )
746- if current_platform .is_device_capability (
747- 90 ) and not has_triton_kernels ():
748- raise NotImplementedError (
749- "Triton kernels must be installed for mxfp4 on hopper" )
744+ if not current_platform .is_device_capability (100 ):
745+ if not is_torch_equal_or_newer ("2.8.0" ):
746+ raise RuntimeError (
747+ "Mxfp4 on non-blackwell requires torch >= 2.8.0" )
748+ if not has_triton_kernels ():
749+ raise NotImplementedError (
750+ "triton_kernels must be installed for "
751+ "mxfp4 on non-blackwell" )
750752 if (current_platform .is_rocm ()
751753 or envs .VLLM_USE_FLASHINFER_MOE_MXFP4_MXFP8
752754 or envs .VLLM_USE_FLASHINFER_MOE_MXFP4_BF16 ):
You can’t perform that action at this time.
0 commit comments