@@ -161,8 +161,7 @@ def test_attn_quant(
161161 run_model (compilation_config , model_name , ** model_kwargs )
162162
163163 matches = re .findall (
164- r"\[compilation/fusion_attn.py:\d+] "
165- r"Fused quant onto (\d+) attention nodes" ,
164+ r"fusion_attn.py:\d+] Fused quant onto (\d+) attention nodes" ,
166165 log_holder .text ,
167166 )
168167 assert len (matches ) == 1 , log_holder .text
@@ -252,8 +251,7 @@ def test_tp2_attn_quant_allreduce_rmsnorm(
252251 compilation_config , model_name , tensor_parallel_size = 2 , ** model_kwargs
253252 )
254253 matches = re .findall (
255- r"\[compilation/fusion_attn.py:\d+] "
256- r"Fused quant onto (\d+) attention nodes" ,
254+ r"fusion_attn.py:\d+] Fused quant onto (\d+) attention nodes" ,
257255 log_holder .text ,
258256 )
259257 assert len (matches ) == 2 , log_holder .text
@@ -262,8 +260,7 @@ def test_tp2_attn_quant_allreduce_rmsnorm(
262260 assert int (matches [1 ]) == attention_fusions
263261
264262 matches = re .findall (
265- r"\[compilation/collective_fusion.py:\d+] "
266- r"Replaced (\d+) patterns" ,
263+ r"collective_fusion.py:\d+] Replaced (\d+) patterns" ,
267264 log_holder .text ,
268265 )
269266 assert len (matches ) == 2 , log_holder .text
0 commit comments