@@ -673,18 +673,6 @@ def _create_predictor(self, predictor_args: PredictorArgument):
673
673
config .enable_use_gpu (100 , device_id )
674
674
config .enable_new_executor ()
675
675
676
- if self .tensor_parallel_degree > 1 :
677
- trainer_endpoints = fleet .worker_endpoints ()
678
- current_endpoint = trainer_endpoints [self .tensor_parallel_rank ]
679
-
680
- dist_config = config .dist_config ()
681
- dist_config .set_ranks (self .tensor_parallel_degree , self .tensor_parallel_rank )
682
- dist_config .set_endpoints (trainer_endpoints , current_endpoint )
683
- dist_config .enable_dist_model (True )
684
-
685
- dist_config .set_comm_init_config (os .path .join (predictor_args .model_name_or_path , "rank_mapping.csv" ))
686
- config .set_dist_config (dist_config )
687
-
688
676
predictor = paddle .inference .create_predictor (config )
689
677
return predictor
690
678
@@ -1178,18 +1166,6 @@ def _create_predictor(self, predictor_args: PredictorArgument):
1178
1166
pass_builder = config .pass_builder ()
1179
1167
passes .addPasses (pass_builder , self .model_config .model_type , self .model_config .quant_type )
1180
1168
1181
- if self .tensor_parallel_degree > 1 :
1182
- trainer_endpoints = fleet .worker_endpoints ()
1183
- current_endpoint = trainer_endpoints [self .tensor_parallel_rank ]
1184
-
1185
- dist_config = config .dist_config ()
1186
- dist_config .set_ranks (self .tensor_parallel_degree , self .tensor_parallel_rank )
1187
- dist_config .set_endpoints (trainer_endpoints , current_endpoint )
1188
- dist_config .enable_dist_model (True )
1189
-
1190
- dist_config .set_comm_init_config (os .path .join (predictor_args .model_name_or_path , "rank_mapping.csv" ))
1191
- config .set_dist_config (dist_config )
1192
-
1193
1169
self .predictor = paddle .inference .create_predictor (config )
1194
1170
1195
1171
def predict (self , input_texts : list [str ], return_tokens = False ):
0 commit comments