Skip to content

Commit 8f18feb

Browse files
authored
Remove last level references not removed in #26355 (#27260)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
1 parent ed540d6 commit 8f18feb

File tree

6 files changed

+13
-13
lines changed

6 files changed

+13
-13
lines changed

tests/compile/piecewise/test_toy_llama.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -355,13 +355,13 @@ def test_toy_llama(
355355
)
356356

357357
compile_config_no_compile = CompilationConfig(
358-
level=CompilationMode.NONE,
358+
mode=CompilationMode.NONE,
359359
cudagraph_mode=CUDAGraphMode.NONE,
360360
backend="eager",
361361
)
362362

363363
compile_config_no_split = CompilationConfig(
364-
level=CompilationMode.VLLM_COMPILE,
364+
mode=CompilationMode.VLLM_COMPILE,
365365
use_inductor_graph_partition=use_inductor_graph_partition,
366366
cudagraph_mode=CUDAGraphMode.PIECEWISE,
367367
backend=backend,

tests/compile/test_aot_compile.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ def forward(self, x: torch.Tensor):
3838
def make_vllm_config() -> VllmConfig:
3939
return VllmConfig(
4040
compilation_config=CompilationConfig(
41-
level=CompilationMode.VLLM_COMPILE,
41+
mode=CompilationMode.VLLM_COMPILE,
4242
)
4343
)
4444

tests/compile/test_config.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -168,7 +168,7 @@ def test_splitting_ops_dynamic():
168168
if is_torch_equal_or_newer("2.9.0.dev"):
169169
config = VllmConfig(
170170
compilation_config=CompilationConfig(
171-
level=CompilationMode.VLLM_COMPILE,
171+
mode=CompilationMode.VLLM_COMPILE,
172172
use_inductor_graph_partition=True,
173173
splitting_ops=["vllm::unified_attention"],
174174
)
@@ -180,7 +180,7 @@ def test_splitting_ops_dynamic():
180180
# When attn_fusion pass enabled, splitting_ops now default to attention ops.
181181
config = VllmConfig(
182182
compilation_config=CompilationConfig(
183-
level=CompilationMode.VLLM_COMPILE,
183+
mode=CompilationMode.VLLM_COMPILE,
184184
pass_config={"enable_attn_fusion": True, "enable_noop": True},
185185
custom_ops=["+quant_fp8"],
186186
cudagraph_mode=CUDAGraphMode.PIECEWISE,
@@ -195,7 +195,7 @@ def test_splitting_ops_dynamic():
195195
if is_torch_equal_or_newer("2.9.0.dev"):
196196
config = VllmConfig(
197197
compilation_config=CompilationConfig(
198-
level=CompilationMode.VLLM_COMPILE,
198+
mode=CompilationMode.VLLM_COMPILE,
199199
use_inductor_graph_partition=True,
200200
pass_config={"enable_attn_fusion": True, "enable_noop": True},
201201
custom_ops=["+quant_fp8"],

tests/compile/test_full_graph.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -198,7 +198,7 @@ def run_model(compile_config: int | CompilationConfig, model: str, **model_kwarg
198198
compilation_config = (
199199
compile_config
200200
if isinstance(compile_config, CompilationConfig)
201-
else CompilationConfig(level=compile_config)
201+
else CompilationConfig(mode=compile_config)
202202
)
203203

204204
prompts = [

tests/compile/test_fusions_e2e.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -151,7 +151,7 @@ def test_attn_quant(
151151
cudagraph_mode=mode,
152152
splitting_ops=splitting_ops,
153153
# Common
154-
level=CompilationMode.VLLM_COMPILE,
154+
mode=CompilationMode.VLLM_COMPILE,
155155
pass_config=PassConfig(enable_attn_fusion=True, enable_noop=True),
156156
# Inductor caches custom passes by default as well via uuid
157157
inductor_compile_config={"force_disable_caches": True},
@@ -236,7 +236,7 @@ def test_tp2_attn_quant_allreduce_rmsnorm(
236236
custom_ops=custom_ops_list,
237237
splitting_ops=splitting_ops,
238238
# Common
239-
level=CompilationMode.VLLM_COMPILE,
239+
mode=CompilationMode.VLLM_COMPILE,
240240
pass_config=PassConfig(
241241
enable_attn_fusion=True,
242242
enable_noop=True,
@@ -273,7 +273,7 @@ def run_model(compile_config: int | CompilationConfig, model: str, **model_kwarg
273273
compilation_config = (
274274
compile_config
275275
if isinstance(compile_config, CompilationConfig)
276-
else CompilationConfig(level=compile_config)
276+
else CompilationConfig(mode=compile_config)
277277
)
278278

279279
prompts = [

tests/model_executor/test_enabled_custom_ops.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ class Relu3(ReLUSquaredActivation):
3636

3737

3838
@pytest.mark.parametrize(
39-
"env, torch_level, backend, ops_enabled, default_on",
39+
"env, compilation_mode, backend, ops_enabled, default_on",
4040
[
4141
# Default values based on compile level
4242
# - All by default (no Inductor compilation)
@@ -77,15 +77,15 @@ class Relu3(ReLUSquaredActivation):
7777
)
7878
def test_enabled_ops(
7979
env: str | None,
80-
torch_level: int,
80+
compilation_mode: int,
8181
backend: str,
8282
ops_enabled: list[int],
8383
default_on: bool,
8484
):
8585
custom_ops = env.split(",") if env else []
8686
vllm_config = VllmConfig(
8787
compilation_config=CompilationConfig(
88-
backend=backend, level=torch_level, custom_ops=custom_ops
88+
backend=backend, mode=compilation_mode, custom_ops=custom_ops
8989
)
9090
)
9191
with set_current_vllm_config(vllm_config):

0 commit comments

Comments
 (0)