File tree Expand file tree Collapse file tree 2 files changed +5
-1
lines changed
backends/xnnpack/test/ops
examples/models/llama/source_transformation Expand file tree Collapse file tree 2 files changed +5
-1
lines changed Original file line number Diff line number Diff line change @@ -395,7 +395,9 @@ def _test_groupwise_dq_linear(
395395 quantize_ (
396396 mod ,
397397 Int8DynamicActivationIntxWeightConfig (
398- weight_dtype = torch .int4 , weight_granularity = PerGroup (group_size )
398+ # pyre-ignore[16]
399+ weight_dtype = torch .int4 ,
400+ weight_granularity = PerGroup (group_size ),
399401 ),
400402 )
401403 unwrap_tensor_subclass (mod )
Original file line number Diff line number Diff line change @@ -135,6 +135,7 @@ def quantize( # noqa C901
135135 PerAxis (0 ) if group_size == 0 else PerGroup (group_size )
136136 ),
137137 weight_mapping_type = MappingType .SYMMETRIC ,
138+ # pyre-ignore[6]
138139 intx_packing_format = "opaque_torchao_auto" ,
139140 ),
140141 )
@@ -157,6 +158,7 @@ def quantize( # noqa C901
157158 quantize_ (
158159 model ,
159160 Int8DynamicActivationIntxWeightConfig (
161+ # pyre-ignore[16]
160162 weight_dtype = torch .int4 ,
161163 weight_granularity = PerGroup (group_size ),
162164 ),
You can’t perform that action at this time.
0 commit comments