@@ -168,7 +168,7 @@ def test_splitting_ops_dynamic():
168168 if is_torch_equal_or_newer ("2.9.0.dev" ):
169169 config = VllmConfig (
170170 compilation_config = CompilationConfig (
171- level = CompilationMode .VLLM_COMPILE ,
171+ mode = CompilationMode .VLLM_COMPILE ,
172172 use_inductor_graph_partition = True ,
173173 splitting_ops = ["vllm::unified_attention" ],
174174 )
@@ -180,7 +180,7 @@ def test_splitting_ops_dynamic():
180180 # When attn_fusion pass enabled, splitting_ops now default to attention ops.
181181 config = VllmConfig (
182182 compilation_config = CompilationConfig (
183- level = CompilationMode .VLLM_COMPILE ,
183+ mode = CompilationMode .VLLM_COMPILE ,
184184 pass_config = {"enable_attn_fusion" : True , "enable_noop" : True },
185185 custom_ops = ["+quant_fp8" ],
186186 cudagraph_mode = CUDAGraphMode .PIECEWISE ,
@@ -195,7 +195,7 @@ def test_splitting_ops_dynamic():
195195 if is_torch_equal_or_newer ("2.9.0.dev" ):
196196 config = VllmConfig (
197197 compilation_config = CompilationConfig (
198- level = CompilationMode .VLLM_COMPILE ,
198+ mode = CompilationMode .VLLM_COMPILE ,
199199 use_inductor_graph_partition = True ,
200200 pass_config = {"enable_attn_fusion" : True , "enable_noop" : True },
201201 custom_ops = ["+quant_fp8" ],
0 commit comments