@@ -264,6 +264,7 @@ def generate_ttir(
264264
265265 assert isinstance (kernel , JITFunction )
266266
267+ # pyrefly: ignore # missing-attribute
267268 context = triton ._C .libtriton .ir .context ()
268269 target = triton .runtime .driver .active .get_current_target ()
269270 backend = triton .compiler .compiler .make_backend (target )
@@ -305,6 +306,7 @@ def generate_ttir(
305306 base_tensor = torch .empty (
306307 [elements_per_dim ] * len (block_shape ), dtype = a .dtype
307308 )
309+ # pyrefly: ignore # bad-argument-type
308310 ordered_args [name ] = TensorDescriptor .from_tensor (base_tensor , block_shape )
309311 elif isinstance (a , (FakeTensor , torch ._inductor .ir .TensorBox )):
310312 with torch ._C ._DisableTorchDispatch ():
@@ -368,6 +370,7 @@ def _get_specialization(args): # type: ignore[no-untyped-def]
368370
369371 target = triton .runtime .driver .active .get_current_target ()
370372 backend_ = triton .compiler .compiler .make_backend (target )
373+ # pyrefly: ignore # missing-attribute
371374 return backend_ .get_attrs_descriptor (args , kernel .params )
372375 else :
373376 assert (
@@ -384,6 +387,7 @@ def _get_specialization(args): # type: ignore[no-untyped-def]
384387 except TypeError : # Unknown arg `specialize_extra`
385388 # Older versions of Triton take specialize_extra as an arg to specialize_impl
386389 specialize_impl = functools .partial (
390+ # pyrefly: ignore # missing-argument
387391 triton .runtime .jit .create_specialize_impl (),
388392 specialize_extra = backend .get_arg_specialization ,
389393 )
@@ -468,6 +472,7 @@ def get_signature_value(idx: int, arg: Any) -> str:
468472 if i not in constexprs
469473 }
470474
475+ # pyrefly: ignore # missing-attribute
471476 triton ._C .libtriton .ir .load_dialects (context )
472477 backend .load_dialects (context )
473478
@@ -477,22 +482,29 @@ def get_signature_value(idx: int, arg: Any) -> str:
477482 # backward compatibility here.
478483 make_ir_sig_params = len (inspect .signature (src .make_ir ).parameters )
479484 get_codegen_implementation_sig_params = len (
485+ # pyrefly: ignore # missing-attribute
480486 inspect .signature (backend .get_codegen_implementation ).parameters
481487 )
482488 if make_ir_sig_params == 2 :
489+ # pyrefly: ignore # missing-argument
483490 ttir_module = src .make_ir (options , context )
484491 elif make_ir_sig_params == 3 :
492+ # pyrefly: ignore # missing-attribute
485493 codegen_fns = backend .get_codegen_implementation ()
494+ # pyrefly: ignore # missing-argument
486495 ttir_module = src .make_ir (options , codegen_fns , context )
487496 elif make_ir_sig_params == 4 :
488497 codegen_args = [options ] if get_codegen_implementation_sig_params == 1 else []
498+ # pyrefly: ignore # missing-attribute
489499 codegen_fns = backend .get_codegen_implementation (* codegen_args )
490500 module_map = backend .get_module_map ()
491501 ttir_module = src .make_ir (options , codegen_fns , module_map , context )
492502 else :
493503 codegen_args = [options ] if get_codegen_implementation_sig_params == 1 else []
504+ # pyrefly: ignore # missing-attribute
494505 codegen_fns = backend .get_codegen_implementation (* codegen_args )
495506 module_map = backend .get_module_map ()
507+ # pyrefly: ignore # bad-argument-count
496508 ttir_module = src .make_ir (target , options , codegen_fns , module_map , context )
497509 if not ttir_module .verify ():
498510 raise RuntimeError ("Verification for TTIR module has failed" )
@@ -1102,6 +1114,7 @@ def triton_kernel_wrapper_mutation_dense(
11021114 from triton .tools .tensor_descriptor import TensorDescriptor
11031115
11041116 block_shape = stable_meta [0 ]
1117+ # pyrefly: ignore # bad-argument-type
11051118 kwargs [k ] = TensorDescriptor .from_tensor (tensor , block_shape )
11061119
11071120 # move as many positional arguments from dicts to args as we
@@ -1658,6 +1671,7 @@ def call_triton_kernel(
16581671 "Passing multiple @triton.autotune decorators is not supported. "
16591672 "Please use a single @triton.autotune decorator instead."
16601673 )
1674+ # pyrefly: ignore # missing-attribute
16611675 iter_kernel = iter_kernel .fn
16621676
16631677 # Process the @triton.heuristics decorator:
@@ -1868,6 +1882,7 @@ def call_triton_kernel(
18681882
18691883 # Both for grid's meta as well as for the kernel, we need combined
18701884 # args and kwargs combined and normalized
1885+ # pyrefly: ignore # missing-attribute
18711886 combined_args_raw = {** dict (zip (variable .kernel .arg_names , args )), ** kwargs }
18721887
18731888 # precompute the grid for the kernel
@@ -2061,6 +2076,7 @@ def __init__(
20612076 kernel_idx : Optional [int ],
20622077 grid : Optional ["TritonGridType" ],
20632078 ) -> None :
2079+ # pyrefly: ignore # bad-assignment
20642080 self .kernel = None
20652081 self .grid = None
20662082 tracing_triton_hopifier_singleton .init_variable (self , kernel , kernel_idx , grid )
0 commit comments