@@ -902,7 +902,6 @@ def test_int8_dynamic_quant_subclass(self, device, dtype):
902
902
)
903
903
904
904
@parameterized .expand (COMMON_DEVICE_DTYPE )
905
- @unittest .skip ("flaky test, will fix in another PR" )
906
905
def test_int8_weight_only_quant_subclass (self , device , dtype ):
907
906
self ._test_lin_weight_subclass_impl (
908
907
Int8WeightOnlyQuantizedLinearWeight .from_float , device , 40 , test_dtype = dtype
@@ -976,7 +975,6 @@ def test_int8_dynamic_quant_subclass_api(self, device, dtype):
976
975
)
977
976
978
977
@parameterized .expand (COMMON_DEVICE_DTYPE )
979
- @unittest .skip ("flaky test, will fix in another PR" )
980
978
def test_int8_weight_only_quant_subclass_api (self , device , dtype ):
981
979
self ._test_lin_weight_subclass_api_impl (
982
980
change_linear_weights_to_int8_woqtensors , device , 40 , test_dtype = dtype
@@ -1157,7 +1155,6 @@ def test_save_load_dqtensors(self, device, dtype):
1157
1155
1158
1156
@parameterized .expand (COMMON_DEVICE_DTYPE )
1159
1157
@torch .no_grad ()
1160
- @unittest .skip ("flaky test, will fix in another PR" )
1161
1158
def test_save_load_int8woqtensors (self , device , dtype ):
1162
1159
self ._test_handle_save_load_meta_impl (change_linear_weights_to_int8_woqtensors , device , test_dtype = dtype )
1163
1160
0 commit comments