@@ -286,7 +286,7 @@ static void launch_bin_bcast_pack(const ggml_tensor * src0, const ggml_tensor *
286286 ne12, ne13,
287287 /* s0, */ s1, s2, s3,
288288 /* s00,*/ s01, s02, s03,
289- /* s10,*/ s11, s12, s13, (const src1_t *) dst->src [I + 1 ]-> data ...);
289+ /* s10,*/ s11, s12, s13, (const src1_t *) tensor_data ( dst->src [I + 1 ]) ...);
290290 } else {
291291 k_bin_bcast_unravel<bin_op, src0_t , src1_t , dst_t >
292292 <<<block_num, block_size, 0 , stream>>> (src0_dd, src1_dd, dst_dd, ne0_fastdiv, ne1_fastdiv,
@@ -302,7 +302,7 @@ static void launch_bin_bcast_pack(const ggml_tensor * src0, const ggml_tensor *
302302 src0_dd, src1_dd, dst_dd, ne0, ne1, ne2, ne3_fastdiv, ne10, ne11, ne12, ne13,
303303 /* s0, */ s1, s2, s3,
304304 /* s00,*/ s01, s02, s03,
305- /* s10,*/ s11, s12, s13, (const src1_t *) dst->src [I + 1 ]-> data ...);
305+ /* s10,*/ s11, s12, s13, (const src1_t *) tensor_data ( dst->src [I + 1 ]) ...);
306306 } else {
307307 k_bin_bcast<bin_op, src0_t , src1_t , dst_t ><<<block_nums, block_dims, 0 , stream>>> (
308308 src0_dd, src1_dd, dst_dd, ne0, ne1, ne2, ne3_fastdiv, ne10, ne11, ne12, ne13,
@@ -389,23 +389,23 @@ static void ggml_cuda_op_bin_bcast(
389389}
390390
391391void ggml_cuda_op_repeat (ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
392- ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_repeat, 0 >>(dst, dst->src [0 ], dst, nullptr , dst->src [0 ]-> data , dst-> data , ctx.stream ());
392+ ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_repeat, 0 >>(dst, dst->src [0 ], dst, nullptr , tensor_data ( dst->src [0 ]), tensor_data ( dst) , ctx.stream ());
393393}
394394
395395void ggml_cuda_op_add (ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
396- ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_add>>(dst->src [0 ], dst->src [1 ], dst, dst->src [0 ]-> data , dst->src [1 ]-> data , dst-> data , ctx.stream ());
396+ ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_add>>(dst->src [0 ], dst->src [1 ], dst, tensor_data ( dst->src [0 ]), tensor_data ( dst->src [1 ]), tensor_data ( dst) , ctx.stream ());
397397}
398398
399399void ggml_cuda_op_sub (ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
400- ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_sub>>(dst->src [0 ], dst->src [1 ], dst, dst->src [0 ]-> data , dst->src [1 ]-> data , dst-> data , ctx.stream ());
400+ ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_sub>>(dst->src [0 ], dst->src [1 ], dst, tensor_data ( dst->src [0 ]), tensor_data ( dst->src [1 ]), tensor_data ( dst) , ctx.stream ());
401401}
402402
403403void ggml_cuda_op_mul (ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
404- ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_mul>>(dst->src [0 ], dst->src [1 ], dst, dst->src [0 ]-> data , dst->src [1 ]-> data , dst-> data , ctx.stream ());
404+ ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_mul>>(dst->src [0 ], dst->src [1 ], dst, tensor_data ( dst->src [0 ]), tensor_data ( dst->src [1 ]), tensor_data ( dst) , ctx.stream ());
405405}
406406
407407void ggml_cuda_op_div (ggml_backend_cuda_context & ctx, ggml_tensor * dst) {
408- ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_div>>(dst->src [0 ], dst->src [1 ], dst, dst->src [0 ]-> data , dst->src [1 ]-> data , dst-> data , ctx.stream ());
408+ ggml_cuda_op_bin_bcast<bin_bcast_cuda<op_div>>(dst->src [0 ], dst->src [1 ], dst, tensor_data ( dst->src [0 ]), tensor_data ( dst->src [1 ]), tensor_data ( dst) , ctx.stream ());
409409}
410410
411411template <float (*op)(const float , const float ), int n_fuse>
@@ -417,19 +417,19 @@ static void ggml_cuda_op_fused_binbcast_impl(ggml_backend_cuda_context & ctx, gg
417417
418418 if (src0->type == GGML_TYPE_F32 && dst->type == GGML_TYPE_F32) {
419419 launch_bin_bcast_pack<op, float , float , float >(src0, src1, dst,
420- (const float *) src0-> data , (const float *) src1-> data , (float *) dst-> data ,
420+ (const float *) tensor_data ( src0) , (const float *) tensor_data ( src1) , (float *) tensor_data ( dst) ,
421421 stream, std::make_index_sequence<n_fuse>{});
422422 } else if (src0->type == GGML_TYPE_F16 && src1->type == GGML_TYPE_F16 && dst->type == GGML_TYPE_F16) {
423423 launch_bin_bcast_pack<op, half, half, half>(src0, src1, dst,
424- (const half *) src0-> data , (const half *) src1-> data , (half *) dst-> data ,
424+ (const half *) tensor_data ( src0) , (const half *) tensor_data ( src1) , (half *) tensor_data ( dst) ,
425425 stream, std::make_index_sequence<n_fuse>{});
426426 } else if (src0->type == GGML_TYPE_F16 && src1->type == GGML_TYPE_F32 && dst->type == GGML_TYPE_F16) {
427427 launch_bin_bcast_pack<op, half, float , half>(src0, src1, dst,
428- (const half *) src0-> data , (const float *) src1-> data , (half *) dst-> data ,
428+ (const half *) tensor_data ( src0) , (const float *) tensor_data ( src1) , (half *) tensor_data ( dst) ,
429429 stream, std::make_index_sequence<n_fuse>{});
430430 } else if (src0->type == GGML_TYPE_F16 && dst->type == GGML_TYPE_F32) {
431431 launch_bin_bcast_pack<op, half, float , float >(src0, src1, dst,
432- (const half *) src0-> data , (const float *) src1-> data , (float *) dst-> data ,
432+ (const half *) tensor_data ( src0) , (const float *) tensor_data ( src1) , (float *) tensor_data ( dst) ,
433433 stream, std::make_index_sequence<n_fuse>{});
434434 } else {
435435 fprintf (stderr,
@@ -491,8 +491,8 @@ void ggml_cuda_op_repeat_back(ggml_backend_cuda_context & ctx, ggml_tensor * dst
491491
492492 switch (dst->type ) {
493493 case GGML_TYPE_F32: {
494- const float * src0_d = (const float *) src0-> data ;
495- float * dst_d = (float *) dst-> data ;
494+ const float * src0_d = (const float *) tensor_data ( src0) ;
495+ float * dst_d = (float *) tensor_data ( dst) ;
496496 repeat_back_cuda (src0_d, dst_d, ne00, ne01, ne02, ne03, s00, s01, s02, s03, ne0, ne1, ne2, ne3, stream);
497497 } break ;
498498 default : {
0 commit comments