@@ -322,7 +322,7 @@ def test_toy_llama():
322322 num_graphs_seen = 0 ,
323323 num_piecewise_graphs_seen = 0 ,
324324 num_piecewise_capturable_graphs_seen = 0 ,
325- num_inductor_compilations = 0 ,
325+ num_backend_compilations = 0 ,
326326 num_cudagraph_caputured = 0 ,
327327 ):
328328 outputs .append (run_model (llama_config , use_compile = False ))
@@ -332,7 +332,7 @@ def test_toy_llama():
332332 num_graphs_seen = 1 , # one graph for the model
333333 num_piecewise_graphs_seen = 1 ,
334334 num_piecewise_capturable_graphs_seen = 1 ,
335- num_inductor_compilations = 1 , # num_piecewise_capturable_graphs_seen
335+ num_backend_compilations = 1 , # num_piecewise_capturable_graphs_seen
336336 num_cudagraph_caputured =
337337 2 , # num_cudagraph_sizes * num_piecewise_capturable_graphs_seen
338338 ):
@@ -345,7 +345,7 @@ def test_toy_llama():
345345 1 , # 2 * num_layers + 1
346346 num_piecewise_capturable_graphs_seen = 1 +
347347 llama_config .num_layers , # 1 + num_layers
348- num_inductor_compilations = 1 +
348+ num_backend_compilations = 1 +
349349 llama_config .num_layers , # num_piecewise_capturable_graphs_seen
350350 num_cudagraph_caputured = 2 *
351351 (1 + llama_config .num_layers
0 commit comments