@@ -150,7 +150,6 @@ def run_downsample_if_req(
150
150
height_sharding = None ,
151
151
packer_l1_accum_enabled = True ,
152
152
enable_act_double_buffer = False ,
153
- enable_split_reader = False ,
154
153
):
155
154
if self .downsample :
156
155
logger .debug (f"Running downsample" )
@@ -180,7 +179,6 @@ def run_downsample_if_req(
180
179
if input_width < 56
181
180
else False ,
182
181
enable_weights_double_buffer = True if input_width < 56 else False ,
183
- enable_split_reader = enable_split_reader ,
184
182
full_inner_dim = True ,
185
183
),
186
184
}
@@ -217,7 +215,6 @@ def __call__(
217
215
eltwise_binary_out_in_place = True ,
218
216
packer_l1_acc = True ,
219
217
enable_act_double_buffer = False ,
220
- enable_split_reader = False ,
221
218
ops_parallel_config = None ,
222
219
layer_module = None ,
223
220
):
@@ -287,7 +284,6 @@ def __call__(
287
284
height_sharding ,
288
285
packer_l1_accum_enabled = packer_l1_acc ,
289
286
enable_act_double_buffer = False ,
290
- enable_split_reader = enable_split_reader ,
291
287
)
292
288
if layer_module and layer_module == "layer4_module1" :
293
289
if ops_parallel_config and "layer4_module1_downsample" not in ops_parallel_config :
@@ -331,7 +327,6 @@ def __call__(
331
327
reshard_if_not_optimal = reshard_if_not_optimal ,
332
328
enable_act_double_buffer = enable_act_double_buffer ,
333
329
enable_weights_double_buffer = True ,
334
- enable_split_reader = enable_split_reader ,
335
330
full_inner_dim = True ,
336
331
),
337
332
}
@@ -439,7 +434,6 @@ def __call__(
439
434
height_sharding ,
440
435
packer_l1_accum_enabled = packer_l1_acc ,
441
436
enable_act_double_buffer = enable_act_double_buffer ,
442
- enable_split_reader = enable_split_reader ,
443
437
)
444
438
445
439
assert ds_out is not None , "ds_out is None"
@@ -578,7 +572,6 @@ def __init__(
578
572
deallocate_activation = dealloc_input ,
579
573
act_block_h_override = act_block_h_override ,
580
574
enable_act_double_buffer = is_wormhole_b0 () or is_blackhole (),
581
- enable_split_reader = True ,
582
575
shard_layout = ttnn .TensorMemoryLayout .HEIGHT_SHARDED ,
583
576
reshard_if_not_optimal = False ,
584
577
# otherwise act block h is not big enough for the reuse
@@ -812,7 +805,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
812
805
reshard_if_not_optimal = reshard ,
813
806
height_sharding = height_shard ,
814
807
enable_act_double_buffer = True ,
815
- enable_split_reader = True ,
816
808
)
817
809
818
810
if is_first_run :
@@ -833,7 +825,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
833
825
x_height ,
834
826
x_width ,
835
827
enable_act_double_buffer = False ,
836
- enable_split_reader = True ,
837
828
layer_module = "layer1_module2" ,
838
829
)
839
830
@@ -845,7 +836,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
845
836
x_height ,
846
837
x_width ,
847
838
enable_act_double_buffer = False ,
848
- enable_split_reader = True ,
849
839
layer_module = "layer1_module3" ,
850
840
)
851
841
@@ -864,7 +854,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
864
854
reshard_if_not_optimal = reshard ,
865
855
height_sharding = height_shard ,
866
856
enable_act_double_buffer = True ,
867
- enable_split_reader = True ,
868
857
layer_module = "layer2_module1" ,
869
858
)
870
859
@@ -886,7 +875,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
886
875
x_height ,
887
876
x_width ,
888
877
enable_act_double_buffer = True ,
889
- enable_split_reader = True ,
890
878
layer_module = "layer2_module2" ,
891
879
)
892
880
@@ -898,7 +886,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
898
886
x_height ,
899
887
x_width ,
900
888
enable_act_double_buffer = True ,
901
- enable_split_reader = True ,
902
889
layer_module = "layer2_module3" ,
903
890
)
904
891
@@ -910,7 +897,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
910
897
x_height ,
911
898
x_width ,
912
899
enable_act_double_buffer = True ,
913
- enable_split_reader = True ,
914
900
layer_module = "layer2_module4" ,
915
901
)
916
902
@@ -931,7 +917,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
931
917
reshard_if_not_optimal = reshard ,
932
918
height_sharding = height_shard ,
933
919
enable_act_double_buffer = True ,
934
- enable_split_reader = False ,
935
920
)
936
921
937
922
if is_first_run :
@@ -952,7 +937,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
952
937
x_height ,
953
938
x_width ,
954
939
enable_act_double_buffer = True ,
955
- enable_split_reader = False ,
956
940
)
957
941
958
942
logger .debug (f"==== Running layer 3 module 3" )
@@ -963,7 +947,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
963
947
x_height ,
964
948
x_width ,
965
949
enable_act_double_buffer = True ,
966
- enable_split_reader = False ,
967
950
layer_module = "layer3_module3" ,
968
951
)
969
952
@@ -975,7 +958,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
975
958
x_height ,
976
959
x_width ,
977
960
enable_act_double_buffer = True ,
978
- enable_split_reader = False ,
979
961
layer_module = "layer3_module4" ,
980
962
)
981
963
@@ -987,7 +969,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
987
969
x_height ,
988
970
x_width ,
989
971
enable_act_double_buffer = True ,
990
- enable_split_reader = False ,
991
972
layer_module = "layer3_module5" ,
992
973
)
993
974
@@ -1000,7 +981,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
1000
981
x_width ,
1001
982
eltwise_binary_out_in_place = True ,
1002
983
enable_act_double_buffer = True ,
1003
- enable_split_reader = False ,
1004
984
)
1005
985
1006
986
reshard = is_blackhole () and self .batch_size == 20
@@ -1031,7 +1011,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
1031
1011
reshard_if_not_optimal = reshard ,
1032
1012
height_sharding = height_shard ,
1033
1013
enable_act_double_buffer = True ,
1034
- enable_split_reader = False ,
1035
1014
ops_parallel_config = ops_parallel_config ,
1036
1015
layer_module = "layer4_module1" ,
1037
1016
)
@@ -1044,7 +1023,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
1044
1023
x_height ,
1045
1024
x_width ,
1046
1025
enable_act_double_buffer = True ,
1047
- enable_split_reader = False ,
1048
1026
layer_module = "layer4_module2" ,
1049
1027
)
1050
1028
@@ -1056,7 +1034,6 @@ def run(self, input_tensor, device, ops_parallel_config) -> ttnn.Tensor:
1056
1034
x_height ,
1057
1035
x_width ,
1058
1036
enable_act_double_buffer = True ,
1059
- enable_split_reader = False ,
1060
1037
layer_module = "layer4_module3" ,
1061
1038
)
1062
1039
0 commit comments