@@ -330,7 +330,7 @@ void * ggml_aligned_malloc(size_t size) {
330330 #endif
331331 if (result != 0 ) {
332332 // Handle allocation failure
333- const char * error_desc = "unknown allocation error" ;
333+ const char * error_desc = "unknown allocation error" ;
334334 switch (result ) {
335335 case EINVAL :
336336 error_desc = "invalid alignment value" ;
@@ -1174,13 +1174,13 @@ void ggml_print_objects(const struct ggml_context * ctx) {
11741174int64_t ggml_nelements (const struct ggml_tensor * tensor ) {
11751175 static_assert (GGML_MAX_DIMS == 4 , "GGML_MAX_DIMS is not 4 - update this function" );
11761176
1177- return tensor -> ne [0 ]* tensor -> ne [1 ]* tensor -> ne [2 ]* tensor -> ne [3 ];
1177+ return tensor -> ne [0 ] * tensor -> ne [1 ] * tensor -> ne [2 ] * tensor -> ne [3 ];
11781178}
11791179
11801180int64_t ggml_nrows (const struct ggml_tensor * tensor ) {
11811181 static_assert (GGML_MAX_DIMS == 4 , "GGML_MAX_DIMS is not 4 - update this function" );
11821182
1183- return tensor -> ne [1 ]* tensor -> ne [2 ]* tensor -> ne [3 ];
1183+ return tensor -> ne [1 ] * tensor -> ne [2 ] * tensor -> ne [3 ];
11841184}
11851185
11861186size_t ggml_nbytes (const struct ggml_tensor * tensor ) {
@@ -1194,14 +1194,14 @@ size_t ggml_nbytes(const struct ggml_tensor * tensor) {
11941194 const size_t blck_size = ggml_blck_size (tensor -> type );
11951195 if (blck_size == 1 ) {
11961196 nbytes = ggml_type_size (tensor -> type );
1197- for (int i = 0 ; i < GGML_MAX_DIMS ; ++ i ) {
1198- nbytes += (tensor -> ne [i ] - 1 )* tensor -> nb [i ];
1197+ for (size_t i = 0 ; i < GGML_MAX_DIMS ; ++ i ) {
1198+ nbytes += (tensor -> ne [i ] - 1 ) * tensor -> nb [i ];
11991199 }
12001200 }
12011201 else {
1202- nbytes = tensor -> ne [0 ]* tensor -> nb [0 ]/ blck_size ;
1203- for (int i = 1 ; i < GGML_MAX_DIMS ; ++ i ) {
1204- nbytes += (tensor -> ne [i ] - 1 )* tensor -> nb [i ];
1202+ nbytes = tensor -> ne [0 ] * tensor -> nb [0 ] / blck_size ;
1203+ for (size_t i = 1 ; i < GGML_MAX_DIMS ; ++ i ) {
1204+ nbytes += (tensor -> ne [i ] - 1 ) * tensor -> nb [i ];
12051205 }
12061206 }
12071207
@@ -1222,11 +1222,11 @@ size_t ggml_type_size(enum ggml_type type) {
12221222
12231223size_t ggml_row_size (enum ggml_type type , int64_t ne ) {
12241224 assert (ne % ggml_blck_size (type ) == 0 );
1225- return ggml_type_size (type )* ne / ggml_blck_size (type );
1225+ return ggml_type_size (type ) * ne / ggml_blck_size (type );
12261226}
12271227
12281228double ggml_type_sizef (enum ggml_type type ) {
1229- return ((double )(type_traits [type ].type_size ))/ type_traits [type ].blck_size ;
1229+ return ((double )(type_traits [type ].type_size )) / type_traits [type ].blck_size ;
12301230}
12311231
12321232const char * ggml_type_name (enum ggml_type type ) {
@@ -1304,19 +1304,19 @@ enum ggml_type ggml_ftype_to_ggml_type(enum ggml_ftype ftype) {
13041304 enum ggml_type wtype = GGML_TYPE_COUNT ;
13051305
13061306 switch (ftype ) {
1307- case GGML_FTYPE_ALL_F32 : wtype = GGML_TYPE_F32 ; break ;
1308- case GGML_FTYPE_MOSTLY_F16 : wtype = GGML_TYPE_F16 ; break ;
1309- case GGML_FTYPE_MOSTLY_BF16 : wtype = GGML_TYPE_BF16 ; break ;
1310- case GGML_FTYPE_MOSTLY_Q4_0 : wtype = GGML_TYPE_Q4_0 ; break ;
1311- case GGML_FTYPE_MOSTLY_Q4_1 : wtype = GGML_TYPE_Q4_1 ; break ;
1312- case GGML_FTYPE_MOSTLY_Q5_0 : wtype = GGML_TYPE_Q5_0 ; break ;
1313- case GGML_FTYPE_MOSTLY_Q5_1 : wtype = GGML_TYPE_Q5_1 ; break ;
1314- case GGML_FTYPE_MOSTLY_Q8_0 : wtype = GGML_TYPE_Q8_0 ; break ;
1315- case GGML_FTYPE_MOSTLY_Q2_K : wtype = GGML_TYPE_Q2_K ; break ;
1316- case GGML_FTYPE_MOSTLY_Q3_K : wtype = GGML_TYPE_Q3_K ; break ;
1317- case GGML_FTYPE_MOSTLY_Q4_K : wtype = GGML_TYPE_Q4_K ; break ;
1318- case GGML_FTYPE_MOSTLY_Q5_K : wtype = GGML_TYPE_Q5_K ; break ;
1319- case GGML_FTYPE_MOSTLY_Q6_K : wtype = GGML_TYPE_Q6_K ; break ;
1307+ case GGML_FTYPE_ALL_F32 : wtype = GGML_TYPE_F32 ; break ;
1308+ case GGML_FTYPE_MOSTLY_F16 : wtype = GGML_TYPE_F16 ; break ;
1309+ case GGML_FTYPE_MOSTLY_BF16 : wtype = GGML_TYPE_BF16 ; break ;
1310+ case GGML_FTYPE_MOSTLY_Q4_0 : wtype = GGML_TYPE_Q4_0 ; break ;
1311+ case GGML_FTYPE_MOSTLY_Q4_1 : wtype = GGML_TYPE_Q4_1 ; break ;
1312+ case GGML_FTYPE_MOSTLY_Q5_0 : wtype = GGML_TYPE_Q5_0 ; break ;
1313+ case GGML_FTYPE_MOSTLY_Q5_1 : wtype = GGML_TYPE_Q5_1 ; break ;
1314+ case GGML_FTYPE_MOSTLY_Q8_0 : wtype = GGML_TYPE_Q8_0 ; break ;
1315+ case GGML_FTYPE_MOSTLY_Q2_K : wtype = GGML_TYPE_Q2_K ; break ;
1316+ case GGML_FTYPE_MOSTLY_Q3_K : wtype = GGML_TYPE_Q3_K ; break ;
1317+ case GGML_FTYPE_MOSTLY_Q4_K : wtype = GGML_TYPE_Q4_K ; break ;
1318+ case GGML_FTYPE_MOSTLY_Q5_K : wtype = GGML_TYPE_Q5_K ; break ;
1319+ case GGML_FTYPE_MOSTLY_Q6_K : wtype = GGML_TYPE_Q6_K ; break ;
13201320 case GGML_FTYPE_MOSTLY_IQ2_XXS : wtype = GGML_TYPE_IQ2_XXS ; break ;
13211321 case GGML_FTYPE_MOSTLY_IQ2_XS : wtype = GGML_TYPE_IQ2_XS ; break ;
13221322 case GGML_FTYPE_MOSTLY_IQ3_XXS : wtype = GGML_TYPE_IQ3_XXS ; break ;
@@ -1326,8 +1326,8 @@ enum ggml_type ggml_ftype_to_ggml_type(enum ggml_ftype ftype) {
13261326 case GGML_FTYPE_MOSTLY_IQ4_XS : wtype = GGML_TYPE_IQ4_XS ; break ;
13271327 case GGML_FTYPE_MOSTLY_IQ3_S : wtype = GGML_TYPE_IQ3_S ; break ;
13281328 case GGML_FTYPE_MOSTLY_IQ2_S : wtype = GGML_TYPE_IQ2_S ; break ;
1329- case GGML_FTYPE_UNKNOWN : wtype = GGML_TYPE_COUNT ; break ;
1330- case GGML_FTYPE_MOSTLY_Q4_1_SOME_F16 : wtype = GGML_TYPE_COUNT ; break ;
1329+ case GGML_FTYPE_UNKNOWN : wtype = GGML_TYPE_COUNT ; break ;
1330+ case GGML_FTYPE_MOSTLY_Q4_1_SOME_F16 : wtype = GGML_TYPE_COUNT ; break ;
13311331 }
13321332
13331333 GGML_ASSERT (wtype != GGML_TYPE_COUNT );
@@ -1348,7 +1348,7 @@ static bool ggml_is_contiguous_n(const struct ggml_tensor * tensor, int n) {
13481348 if (tensor -> ne [0 ] != ggml_blck_size (tensor -> type ) && tensor -> nb [0 ] != next_nb ) {
13491349 return false;
13501350 }
1351- next_nb *= tensor -> ne [0 ]/ ggml_blck_size (tensor -> type );
1351+ next_nb *= tensor -> ne [0 ] / ggml_blck_size (tensor -> type );
13521352 for (int i = 1 ; i < GGML_MAX_DIMS ; i ++ ) {
13531353 if (tensor -> ne [i ] != 1 ) {
13541354 if (i > n ) {
@@ -1358,7 +1358,7 @@ static bool ggml_is_contiguous_n(const struct ggml_tensor * tensor, int n) {
13581358 next_nb *= tensor -> ne [i ];
13591359 } else {
13601360 // this dimension does not need to be contiguous
1361- next_nb = tensor -> ne [i ]* tensor -> nb [i ];
1361+ next_nb = tensor -> ne [i ] * tensor -> nb [i ];
13621362 }
13631363 }
13641364 }
@@ -1382,7 +1382,7 @@ bool ggml_is_contiguous_2(const struct ggml_tensor * tensor) {
13821382}
13831383
13841384bool ggml_is_contiguously_allocated (const struct ggml_tensor * tensor ) {
1385- return ggml_nbytes (tensor ) == ggml_nelements (tensor ) * ggml_type_size (tensor -> type )/ ggml_blck_size (tensor -> type );
1385+ return ggml_nbytes (tensor ) == ggml_nelements (tensor ) * ggml_type_size (tensor -> type ) / ggml_blck_size (tensor -> type );
13861386}
13871387
13881388bool ggml_is_permuted (const struct ggml_tensor * tensor ) {
@@ -1409,8 +1409,8 @@ static inline bool ggml_is_padded_1d(const struct ggml_tensor * tensor) {
14091409
14101410 return
14111411 tensor -> nb [0 ] == ggml_type_size (tensor -> type ) &&
1412- tensor -> nb [2 ] == tensor -> nb [1 ]* tensor -> ne [1 ] &&
1413- tensor -> nb [3 ] == tensor -> nb [2 ]* tensor -> ne [2 ];
1412+ tensor -> nb [2 ] == tensor -> nb [1 ] * tensor -> ne [1 ] &&
1413+ tensor -> nb [3 ] == tensor -> nb [2 ] * tensor -> ne [2 ];
14141414}
14151415
14161416bool ggml_is_empty (const struct ggml_tensor * tensor ) {
@@ -1675,9 +1675,9 @@ static struct ggml_tensor * ggml_new_tensor_impl(
16751675 }
16761676
16771677 result -> nb [0 ] = ggml_type_size (type );
1678- result -> nb [1 ] = result -> nb [0 ]* (result -> ne [0 ]/ ggml_blck_size (type ));
1679- for (int i = 2 ; i < GGML_MAX_DIMS ; i ++ ) {
1680- result -> nb [i ] = result -> nb [i - 1 ]* result -> ne [i - 1 ];
1678+ result -> nb [1 ] = result -> nb [0 ] * (result -> ne [0 ] / ggml_blck_size (type ));
1679+ for (size_t i = 2 ; i < GGML_MAX_DIMS ; i ++ ) {
1680+ result -> nb [i ] = result -> nb [i - 1 ] * result -> ne [i - 1 ];
16811681 }
16821682
16831683 ctx -> n_objects ++ ;
@@ -1745,22 +1745,22 @@ void ggml_unravel_index(const struct ggml_tensor * tensor, int64_t i, int64_t *
17451745 const int64_t ne1 = tensor -> ne [1 ];
17461746 const int64_t ne0 = tensor -> ne [0 ];
17471747
1748- const int64_t i3_ = (i / (ne2 * ne1 * ne0 ));
1749- const int64_t i2_ = (i - i3_ * ne2 * ne1 * ne0 )/ (ne1 * ne0 );
1750- const int64_t i1_ = (i - i3_ * ne2 * ne1 * ne0 - i2_ * ne1 * ne0 )/ ne0 ;
1751- const int64_t i0_ = (i - i3_ * ne2 * ne1 * ne0 - i2_ * ne1 * ne0 - i1_ * ne0 );
1748+ const int64_t i3_ = (i / (ne2 * ne1 * ne0 ));
1749+ const int64_t i2_ = (i - i3_ * ne2 * ne1 * ne0 ) / (ne1 * ne0 );
1750+ const int64_t i1_ = (i - i3_ * ne2 * ne1 * ne0 - i2_ * ne1 * ne0 ) / ne0 ;
1751+ const int64_t i0_ = (i - i3_ * ne2 * ne1 * ne0 - i2_ * ne1 * ne0 - i1_ * ne0 );
17521752
17531753 if (i0 ) {
1754- * i0 = i0_ ;
1754+ * i0 = i0_ ;
17551755 }
17561756 if (i1 ) {
1757- * i1 = i1_ ;
1757+ * i1 = i1_ ;
17581758 }
17591759 if (i2 ) {
1760- * i2 = i2_ ;
1760+ * i2 = i2_ ;
17611761 }
17621762 if (i3 ) {
1763- * i3 = i3_ ;
1763+ * i3 = i3_ ;
17641764 }
17651765}
17661766
@@ -1810,7 +1810,7 @@ struct ggml_tensor * ggml_view_tensor(
18101810 struct ggml_tensor * result = ggml_new_tensor_impl (ctx , src -> type , GGML_MAX_DIMS , src -> ne , src , 0 );
18111811 ggml_format_name (result , "%s (view)" , src -> name );
18121812
1813- for (int i = 0 ; i < GGML_MAX_DIMS ; i ++ ) {
1813+ for (size_t i = 0 ; i < GGML_MAX_DIMS ; i ++ ) {
18141814 result -> nb [i ] = src -> nb [i ];
18151815 }
18161816
@@ -2290,7 +2290,7 @@ struct ggml_tensor * ggml_sum_rows(
22902290 struct ggml_context * ctx ,
22912291 struct ggml_tensor * a ) {
22922292 int64_t ne [GGML_MAX_DIMS ] = { 1 };
2293- for (int i = 1 ; i < GGML_MAX_DIMS ; ++ i ) {
2293+ for (size_t i = 1 ; i < GGML_MAX_DIMS ; ++ i ) {
22942294 ne [i ] = a -> ne [i ];
22952295 }
22962296
@@ -2678,7 +2678,7 @@ static struct ggml_tensor * ggml_glu_impl(
26782678 GGML_ASSERT (a -> type == b -> type );
26792679 }
26802680
2681- int64_t ne [GGML_MAX_DIMS ] = { a -> ne [0 ] / 2 }; for (int i = 1 ; i < GGML_MAX_DIMS ; i ++ ) ne [i ] = a -> ne [i ];
2681+ int64_t ne [GGML_MAX_DIMS ] = { a -> ne [0 ] / 2 }; for (size_t i = 1 ; i < GGML_MAX_DIMS ; i ++ ) ne [i ] = a -> ne [i ];
26822682 struct ggml_tensor * result = ggml_new_tensor_impl (ctx , a -> type , GGML_MAX_DIMS , b ? a -> ne : ne , NULL , 0 );
26832683
26842684 ggml_set_op_params_i32 (result , 0 , (int32_t ) op );
@@ -2963,9 +2963,9 @@ struct ggml_tensor * ggml_l2_norm_inplace(
29632963static inline bool ggml_can_mul_mat (const struct ggml_tensor * t0 , const struct ggml_tensor * t1 ) {
29642964 static_assert (GGML_MAX_DIMS == 4 , "GGML_MAX_DIMS is not 4 - update this function" );
29652965
2966- return (t0 -> ne [0 ] == t1 -> ne [0 ]) &&
2967- (t1 -> ne [2 ]% t0 -> ne [2 ] == 0 ) && // verify t0 is broadcastable
2968- (t1 -> ne [3 ]% t0 -> ne [3 ] == 0 );
2966+ return (t0 -> ne [0 ] == t1 -> ne [0 ]) &&
2967+ (t1 -> ne [2 ] % t0 -> ne [2 ] == 0 ) && // verify t0 is broadcastable
2968+ (t1 -> ne [3 ] % t0 -> ne [3 ] == 0 );
29692969}
29702970
29712971struct ggml_tensor * ggml_mul_mat (
@@ -3040,9 +3040,9 @@ struct ggml_tensor * ggml_mul_mat_id(
30403040static inline bool ggml_can_out_prod (const struct ggml_tensor * t0 , const struct ggml_tensor * t1 ) {
30413041 static_assert (GGML_MAX_DIMS == 4 , "GGML_MAX_DIMS is not 4 - update this function" );
30423042
3043- return (t0 -> ne [1 ] == t1 -> ne [1 ]) &&
3044- (t1 -> ne [2 ]% t0 -> ne [2 ] == 0 ) && // verify t0 is broadcastable
3045- (t1 -> ne [3 ]% t0 -> ne [3 ] == 0 );
3043+ return (t0 -> ne [1 ] == t1 -> ne [1 ]) &&
3044+ (t1 -> ne [2 ] % t0 -> ne [2 ] == 0 ) && // verify t0 is broadcastable
3045+ (t1 -> ne [3 ] % t0 -> ne [3 ] == 0 );
30463046}
30473047
30483048struct ggml_tensor * ggml_out_prod (
@@ -3720,8 +3720,8 @@ static struct ggml_tensor * ggml_soft_max_impl(
37203720 GGML_ASSERT (ggml_is_contiguous (mask ));
37213721 GGML_ASSERT (mask -> ne [0 ] == a -> ne [0 ]);
37223722 GGML_ASSERT (mask -> ne [1 ] >= a -> ne [1 ]);
3723- GGML_ASSERT (a -> ne [2 ]% mask -> ne [2 ] == 0 );
3724- GGML_ASSERT (a -> ne [3 ]% mask -> ne [3 ] == 0 );
3723+ GGML_ASSERT (a -> ne [2 ] % mask -> ne [2 ] == 0 );
3724+ GGML_ASSERT (a -> ne [3 ] % mask -> ne [3 ] == 0 );
37253725 }
37263726
37273727 if (max_bias > 0.0f ) {
@@ -4418,7 +4418,7 @@ struct ggml_tensor * ggml_conv_transpose_2d_p0(
44184418 a -> ne [2 ], b -> ne [3 ],
44194419 };
44204420
4421- struct ggml_tensor * result = ggml_new_tensor (ctx , GGML_TYPE_F32 , 4 , ne );
4421+ struct ggml_tensor * result = ggml_new_tensor (ctx , GGML_TYPE_F32 , 4 , ne );
44224422
44234423 ggml_set_op_params_i32 (result , 0 , stride );
44244424
@@ -5043,8 +5043,8 @@ static struct ggml_tensor * ggml_add_rel_pos_impl(
50435043 GGML_ASSERT (ph -> type == GGML_TYPE_F32 );
50445044 GGML_ASSERT (pw -> type == GGML_TYPE_F32 );
50455045 GGML_ASSERT (pw -> ne [3 ] == a -> ne [2 ]);
5046- GGML_ASSERT (pw -> ne [0 ]* pw -> ne [0 ] == a -> ne [0 ]);
5047- GGML_ASSERT (pw -> ne [1 ]* pw -> ne [2 ] == a -> ne [1 ]);
5046+ GGML_ASSERT (pw -> ne [0 ] * pw -> ne [0 ] == a -> ne [0 ]);
5047+ GGML_ASSERT (pw -> ne [1 ] * pw -> ne [2 ] == a -> ne [1 ]);
50485048
50495049 struct ggml_tensor * result = inplace ? ggml_view_tensor (ctx , a ) : ggml_dup_tensor (ctx , a );
50505050 ggml_set_op_params_i32 (result , 0 , inplace ? 1 : 0 );
@@ -5544,7 +5544,7 @@ size_t ggml_hash_size(size_t min_sz) {
55445544 size_t l = 0 ;
55455545 size_t r = n_primes ;
55465546 while (l < r ) {
5547- size_t m = (l + r )/ 2 ;
5547+ size_t m = (l + r ) / 2 ;
55485548 if (primes [m ] < min_sz ) {
55495549 l = m + 1 ;
55505550 } else {
@@ -6138,7 +6138,7 @@ static size_t ggml_visit_parents(struct ggml_cgraph * cgraph, struct ggml_tensor
61386138 return node_hash_pos ;
61396139 }
61406140
6141- for (int i = 0 ; i < GGML_MAX_SRC ; ++ i ) {
6141+ for (size_t i = 0 ; i < GGML_MAX_SRC ; ++ i ) {
61426142 const int k =
61436143 (cgraph -> order == GGML_CGRAPH_EVAL_ORDER_LEFT_TO_RIGHT ) ? i :
61446144 (cgraph -> order == GGML_CGRAPH_EVAL_ORDER_RIGHT_TO_LEFT ) ? (GGML_MAX_SRC - 1 - i ) :
@@ -6260,7 +6260,7 @@ void ggml_build_backward_expand(
62606260 default :
62616261 break ;
62626262 }
6263- for (int j = 0 ; j < GGML_MAX_SRC ; ++ j ) {
6263+ for (size_t j = 0 ; j < GGML_MAX_SRC ; ++ j ) {
62646264 if (!node -> src [j ] || ignore_src [j ] || !grads_needed [ggml_hash_find (& cgraph -> visited_hash_set , node -> src [j ])]) {
62656265 continue ;
62666266 }
0 commit comments