66 * Copyright (c) 2024 NVIDIA Corporation. All rights reserved.
77 * Copyright (c) 2024 Research Organization for Information Science
88 * and Technology (RIST). All rights reserved.
9+ * Copyright (c) 2025 Triad National Security, LLC. All rights
10+ * reserved.
911 *
1012 * $COPYRIGHT$
1113 *
@@ -114,11 +116,11 @@ _Generic((*(out)), \
114116#if defined(GENERATE_NEON_CODE )
115117#define OP_AARCH64_FUNC (name , type_name , type_size , type_cnt , type , op ) \
116118 static void OP_CONCAT(ompi_op_aarch64_2buff_##name##_##type##type_size##_t, \
117- APPEND)(const void *_in, void *_out, int *count, \
119+ APPEND)(const void *_in, void *_out, size_t *count, \
118120 struct ompi_datatype_t **dtype, \
119- struct ompi_op_base_module_1_0_0_t *module) \
121+ struct ompi_op_base_module_2_0_0_t *module) \
120122 { \
121- int left_over = *count; \
123+ size_t left_over = *count; \
122124 type##type_size##_t *in = (type##type_size##_t *) _in, \
123125 *out = (type##type_size##_t *) _out; \
124126 OP_CONCAT(OMPI_OP_TYPE_PREPEND, type##type_size##x##type_cnt##_t) vsrc, vdst; \
@@ -138,12 +140,12 @@ _Generic((*(out)), \
138140#elif defined(GENERATE_SVE_CODE )
139141#define OP_AARCH64_FUNC (name , type_name , type_size , type_cnt , type , op ) \
140142 static void OP_CONCAT(ompi_op_aarch64_2buff_##name##_##type##type_size##_t, APPEND) \
141- (const void *_in, void *_out, int *count, \
143+ (const void *_in, void *_out, size_t *count, \
142144 struct ompi_datatype_t **dtype, \
143- struct ompi_op_base_module_1_0_0_t *module) \
145+ struct ompi_op_base_module_2_0_0_t *module) \
144146 { \
145147 const int types_per_step = svcnt(*((type##type_size##_t *) _in)); \
146- const int cnt = *count; \
148+ const size_t cnt = *count; \
147149 type##type_size##_t *in = (type##type_size##_t *) _in, \
148150 *out = (type##type_size##_t *) _out; \
149151 OP_CONCAT(OMPI_OP_TYPE_PREPEND, type##type_size##_t) vsrc, vdst; \
@@ -279,11 +281,11 @@ _Generic((*(out)), \
279281#if defined(GENERATE_NEON_CODE )
280282#define OP_AARCH64_FUNC_3BUFF (name , type_name , type_size , type_cnt , type , op ) \
281283static void OP_CONCAT(ompi_op_aarch64_3buff_##name##_##type##type_size##_t, APPEND) \
282- (const void *_in1, const void *_in2, void *_out, int *count, \
284+ (const void *_in1, const void *_in2, void *_out, size_t *count, \
283285 struct ompi_datatype_t **dtype, \
284- struct ompi_op_base_module_1_0_0_t *module) \
286+ struct ompi_op_base_module_2_0_0_t *module) \
285287{ \
286- int left_over = *count; \
288+ size_t left_over = *count; \
287289 type##type_size##_t *in1 = (type##type_size##_t *) _in1, \
288290 *in2 = (type##type_size##_t *) _in2, \
289291 *out = (type##type_size##_t *) _out; \
@@ -304,17 +306,17 @@ static void OP_CONCAT(ompi_op_aarch64_3buff_##name##_##type##type_size##_t, APPE
304306#elif defined(GENERATE_SVE_CODE )
305307#define OP_AARCH64_FUNC_3BUFF (name , type_name , type_size , type_cnt , type , op ) \
306308static void OP_CONCAT(ompi_op_aarch64_3buff_##name##_##type##type_size##_t, APPEND) \
307- (const void *_in1, const void *_in2, void *_out, int *count, \
309+ (const void *_in1, const void *_in2, void *_out, size_t *count, \
308310 struct ompi_datatype_t **dtype, \
309- struct ompi_op_base_module_1_0_0_t *module) \
311+ struct ompi_op_base_module_2_0_0_t *module) \
310312{ \
311313 const int types_per_step = svcnt(*((type##type_size##_t *) _in1)); \
312314 type##type_size##_t *in1 = (type##type_size##_t *) _in1, \
313315 *in2 = (type##type_size##_t *) _in2, \
314316 *out = (type##type_size##_t *) _out; \
315- const int cnt = *count; \
317+ const size_t cnt = *count; \
316318 OP_CONCAT(OMPI_OP_TYPE_PREPEND, type##type_size##_t) vsrc, vdst; \
317- for (int idx=0; idx < cnt; idx += types_per_step) { \
319+ for (size_t idx=0; idx < cnt; idx += types_per_step) { \
318320 svbool_t pred = svwhilelt_b##type_size(idx, cnt); \
319321 vsrc = svld1(pred, &in1[idx]); \
320322 vdst = svld1(pred, &in2[idx]); \
0 commit comments