11/*
2- * Copyright (c) 2018-2020 The University of Tennessee and The University
2+ * Copyright (c) 2018-2023 The University of Tennessee and The University
33 * of Tennessee Research Foundation. All rights
44 * reserved.
55 * Copyright (c) 2020 Bull S.A.S. All rights reserved.
@@ -63,7 +63,7 @@ mca_coll_han_set_bcast_args(mca_coll_han_bcast_args_t * args, mca_coll_task_t *
6363 * iter 4 | | | | lb | task: t1, contains lb
6464 */
6565int
66- mca_coll_han_bcast_intra (void * buff ,
66+ mca_coll_han_bcast_intra (void * buf ,
6767 int count ,
6868 struct ompi_datatype_t * dtype ,
6969 int root ,
@@ -84,8 +84,8 @@ mca_coll_han_bcast_intra(void *buff,
8484 * future calls will then be automatically redirected.
8585 */
8686 HAN_LOAD_FALLBACK_COLLECTIVES (han_module , comm );
87- return comm -> c_coll -> coll_bcast ( buff , count , dtype , root ,
88- comm , comm -> c_coll -> coll_bcast_module );
87+ return han_module -> previous_bcast ( buf , count , dtype , root ,
88+ comm , han_module -> previous_bcast_module );
8989 }
9090 /* Topo must be initialized to know rank distribution which then is used to
9191 * determine if han can be used */
@@ -97,8 +97,8 @@ mca_coll_han_bcast_intra(void *buff,
9797 * future calls will then be automatically redirected.
9898 */
9999 HAN_LOAD_FALLBACK_COLLECTIVE (han_module , comm , bcast );
100- return comm -> c_coll -> coll_bcast ( buff , count , dtype , root ,
101- comm , comm -> c_coll -> coll_bcast_module );
100+ return han_module -> previous_bcast ( buf , count , dtype , root ,
101+ comm , han_module -> previous_bcast_module );
102102 }
103103
104104 ompi_datatype_get_extent (dtype , & lb , & extent );
@@ -129,7 +129,7 @@ mca_coll_han_bcast_intra(void *buff,
129129 mca_coll_task_t * t0 = OBJ_NEW (mca_coll_task_t );
130130 /* Setup up t0 task arguments */
131131 mca_coll_han_bcast_args_t * t = malloc (sizeof (mca_coll_han_bcast_args_t ));
132- mca_coll_han_set_bcast_args (t , t0 , (char * ) buff , seg_count , dtype ,
132+ mca_coll_han_set_bcast_args (t , t0 , (char * )buf , seg_count , dtype ,
133133 root_up_rank , root_low_rank , up_comm , low_comm ,
134134 num_segments , 0 , w_rank , count - (num_segments - 1 ) * seg_count ,
135135 low_rank != root_low_rank );
@@ -222,7 +222,7 @@ int mca_coll_han_bcast_t1_task(void *task_args)
222222 * communications without tasks.
223223 */
224224int
225- mca_coll_han_bcast_intra_simple (void * buff ,
225+ mca_coll_han_bcast_intra_simple (void * buf ,
226226 int count ,
227227 struct ompi_datatype_t * dtype ,
228228 int root ,
@@ -246,8 +246,8 @@ mca_coll_han_bcast_intra_simple(void *buff,
246246 * future calls will then be automatically redirected.
247247 */
248248 HAN_LOAD_FALLBACK_COLLECTIVES (han_module , comm );
249- return comm -> c_coll -> coll_bcast ( buff , count , dtype , root ,
250- comm , comm -> c_coll -> coll_bcast_module );
249+ return han_module -> previous_bcast ( buf , count , dtype , root ,
250+ comm , han_module -> previous_bcast_module );
251251 }
252252 /* Topo must be initialized to know rank distribution which then is used to
253253 * determine if han can be used */
@@ -259,8 +259,8 @@ mca_coll_han_bcast_intra_simple(void *buff,
259259 * future calls will then be automatically redirected.
260260 */
261261 HAN_LOAD_FALLBACK_COLLECTIVE (han_module , comm , bcast );
262- return comm -> c_coll -> coll_bcast ( buff , count , dtype , root ,
263- comm , comm -> c_coll -> coll_bcast_module );
262+ return han_module -> previous_bcast ( buf , count , dtype , root ,
263+ comm , han_module -> previous_bcast_module );
264264 }
265265
266266 low_comm = han_module -> sub_comm [INTRA_NODE ];
@@ -277,18 +277,18 @@ mca_coll_han_bcast_intra_simple(void *buff,
277277 w_rank , root_low_rank , root_up_rank ));
278278
279279 if (low_rank == root_low_rank ) {
280- up_comm -> c_coll -> coll_bcast (buff , count , dtype , root_up_rank ,
280+ up_comm -> c_coll -> coll_bcast (buf , count , dtype , root_up_rank ,
281281 up_comm , up_comm -> c_coll -> coll_bcast_module );
282282
283283 /* To remove when han has better sub-module selection.
284284 For now switching to ibcast enables to make runs with libnbc. */
285285 //ompi_request_t req;
286- //up_comm->c_coll->coll_ibcast(buff , count, dtype, root_up_rank,
286+ //up_comm->c_coll->coll_ibcast(buf , count, dtype, root_up_rank,
287287 // up_comm, &req, up_comm->c_coll->coll_ibcast_module);
288288 //ompi_request_wait(&req, MPI_STATUS_IGNORE);
289289
290290 }
291- low_comm -> c_coll -> coll_bcast (buff , count , dtype , root_low_rank ,
291+ low_comm -> c_coll -> coll_bcast (buf , count , dtype , root_low_rank ,
292292 low_comm , low_comm -> c_coll -> coll_bcast_module );
293293
294294 return OMPI_SUCCESS ;
0 commit comments