@@ -1997,12 +1997,6 @@ size_t ggml_nbytes_pad(const struct ggml_tensor * tensor) {
19971997 return GGML_PAD(ggml_nbytes(tensor), GGML_MEM_ALIGN);
19981998}
19991999
2000- size_t ggml_nbytes_split(const struct ggml_tensor * tensor, int nrows_split) {
2001- static_assert(GGML_MAX_DIMS == 4, "GGML_MAX_DIMS is not 4 - update this function");
2002-
2003- return (nrows_split*tensor->ne[0]*ggml_type_size(tensor->type))/ggml_blck_size(tensor->type);
2004- }
2005-
20062000int ggml_blck_size(enum ggml_type type) {
20072001 return type_traits[type].blck_size;
20082002}
@@ -2491,7 +2485,7 @@ static struct ggml_tensor * ggml_new_tensor_impl(
24912485 view_src = view_src->view_src;
24922486 }
24932487
2494- size_t data_size = ggml_type_size (type)*( ne[0]/ggml_blck_size(type) );
2488+ size_t data_size = ggml_row_size (type, ne[0]);
24952489 for (int i = 1; i < n_dims; i++) {
24962490 data_size *= ne[i];
24972491 }
@@ -9698,7 +9692,7 @@ static void ggml_compute_forward_mul_mat(
96989692 if (params->type == GGML_TASK_INIT) {
96999693 if (src1->type != vec_dot_type) {
97009694 char * wdata = params->wdata;
9701- const size_t row_size = ne10*ggml_type_size (vec_dot_type)/ggml_blck_size(vec_dot_type );
9695+ const size_t row_size = ggml_row_size (vec_dot_type, ne10 );
97029696
97039697 assert(params->wsize >= ne11*ne12*ne13*row_size);
97049698 assert(src1->type == GGML_TYPE_F32);
@@ -9721,7 +9715,7 @@ static void ggml_compute_forward_mul_mat(
97219715 }
97229716
97239717 const void * wdata = (src1->type == vec_dot_type) ? src1->data : params->wdata;
9724- const size_t row_size = ne10*ggml_type_size (vec_dot_type)/ggml_blck_size(vec_dot_type );
9718+ const size_t row_size = ggml_row_size (vec_dot_type, ne10 );
97259719
97269720 const int64_t nr0 = ne01; // src0 rows
97279721 const int64_t nr1 = cne1*ne12*ne13; // src1 rows
@@ -16326,7 +16320,7 @@ struct ggml_cplan ggml_graph_plan(struct ggml_cgraph * cgraph, int n_threads) {
1632616320 } else
1632716321#endif
1632816322 if (node->src[1]->type != vec_dot_type) {
16329- cur = ggml_type_size (vec_dot_type)* ggml_nelements(node->src[1])/ggml_blck_size(vec_dot_type );
16323+ cur = ggml_row_size (vec_dot_type, ggml_nelements(node->src[1]));
1633016324 }
1633116325 } break;
1633216326 case GGML_OP_MUL_MAT_ID:
@@ -16343,7 +16337,7 @@ struct ggml_cplan ggml_graph_plan(struct ggml_cgraph * cgraph, int n_threads) {
1634316337 } else
1634416338#endif
1634516339 if (b->type != vec_dot_type) {
16346- cur = ggml_type_size (vec_dot_type)* ggml_nelements(b)/ggml_blck_size(vec_dot_type );
16340+ cur = ggml_row_size (vec_dot_type, ggml_nelements(b));
1634716341 }
1634816342 } break;
1634916343 case GGML_OP_OUT_PROD:
@@ -18703,7 +18697,7 @@ struct gguf_context * gguf_init_from_file(const char * fname, struct gguf_init_p
1870318697 return NULL;
1870418698 }
1870518699
18706- const size_t size_cur = (ne*ggml_type_size( info->type))/ggml_blck_size(info->type );
18700+ const size_t size_cur = ggml_row_size( info->type, ne );
1870718701
1870818702 ctx->size += GGML_PAD(size_cur, ctx->alignment);
1870918703 }
0 commit comments