@@ -212,18 +212,18 @@ def _config_reduction(config, task):
212212 print (f"-- MODEL CONVERTED IN { time .perf_counter () - begin } " )
213213 model = onnx .load (filename , load_external_data = False )
214214 if attention == "PACKED" :
215- self .assertIn ("PackedMultiHeadAttention" , str (model ))
215+ self .assertIn (' "PackedMultiHeadAttention"' , str (model ))
216216 elif attention == "BIGMASK" :
217- self .assertNotIn ("PackedMultiHeadAttention" , str (model ))
217+ self .assertNotIn (' "PackedMultiHeadAttention"' , str (model ))
218218 self .assertNotIn ("MultiHeadAttention" , str (model ))
219219 self .assertNotIn ("Loop" , {n .op_type for n in model .graph .node })
220220 elif attention == "LOOPMHA" :
221- self .assertNotIn ("PackedMultiHeadAttention" , str (model ))
222- self .assertIn ("MultiHeadAttention" , str (model ))
221+ self .assertNotIn (' "PackedMultiHeadAttention"' , str (model ))
222+ self .assertIn (' "MultiHeadAttention"' , str (model ))
223223 self .assertIn ("Loop" , {n .op_type for n in model .graph .node })
224224 elif attention == "LOOPA24" :
225- self .assertNotIn ("PackedMultiHeadAttention" , str (model ))
226- self .assertNotIn ("MultiHeadAttention" , str (model ))
225+ self .assertNotIn (' "PackedMultiHeadAttention"' , str (model ))
226+ self .assertNotIn (' "MultiHeadAttention"' , str (model ))
227227 self .assertIn ("Loop" , {n .op_type for n in model .graph .node })
228228 else :
229229 raise AssertionError (f"attention={ attention !r} not expected" )
@@ -257,7 +257,7 @@ def _config_reduction(config, task):
257257 else ["CPUExecutionProvider" ]
258258 ),
259259 use_ort = True ,
260- atol = 0.02 ,
260+ atol = 0.05 ,
261261 rtol = 10 ,
262262 # ep=pt2_file,
263263 expected = expected ,
0 commit comments