@@ -327,7 +327,7 @@ def _test_toy_llama(*, use_inductor):
327327            num_piecewise_graphs_seen = 0 ,
328328            num_piecewise_capturable_graphs_seen = 0 ,
329329            num_backend_compilations = 0 ,
330-             num_cudagraph_caputured = 0 ,
330+             num_cudagraph_captured = 0 ,
331331    ):
332332        outputs .append (
333333            run_model (llama_config , use_inductor = False , use_compile = False ))
@@ -343,7 +343,7 @@ def _test_toy_llama(*, use_inductor):
343343            num_piecewise_graphs_seen = 1 ,
344344            num_piecewise_capturable_graphs_seen = 1 ,
345345            num_backend_compilations = 1 ,  # num_piecewise_capturable_graphs_seen 
346-             num_cudagraph_caputured = 
346+             num_cudagraph_captured = 
347347            2 ,  # num_cudagraph_sizes * num_piecewise_capturable_graphs_seen 
348348            ** kwargs ,
349349    ):
@@ -361,7 +361,7 @@ def _test_toy_llama(*, use_inductor):
361361            llama_config .num_layers ,  # 1 + num_layers 
362362            num_backend_compilations = 1  + 
363363            llama_config .num_layers ,  # num_piecewise_capturable_graphs_seen 
364-             num_cudagraph_caputured = 2  * 
364+             num_cudagraph_captured = 2  * 
365365        (1  +  llama_config .num_layers 
366366         ),  # num_cudagraph_sizes * num_piecewise_capturable_graphs_seen 
367367    ):
0 commit comments