@@ -2184,7 +2184,7 @@ static void aclnn_cache_init(ggml_backend_cann_context& ctx, ggml_tensor* dst,
21842184 ACL_CHECK (aclrtFree (ctx.init_ptr ));
21852185 }
21862186 ACL_CHECK (aclrtMalloc (&ctx.init_ptr , theta_scale_length * sizeof (float_t ), ACL_MEM_MALLOC_HUGE_FIRST));
2187-
2187+
21882188 aclTensor* acl_theta_scale_tensor =
21892189 ggml_cann_create_tensor (ctx.init_ptr , ACL_FLOAT, sizeof (float_t ),
21902190 theta_scale_ne, theta_scale_nb, GGML_MAX_DIMS);
@@ -2198,7 +2198,7 @@ static void aclnn_cache_init(ggml_backend_cann_context& ctx, ggml_tensor* dst,
21982198 aclScalar* acl_theta_scale = aclCreateScalar (&theta_scale, aclDataType::ACL_FLOAT);
21992199 GGML_CANN_CALL_ACLNN_OP (ctx, PowScalarTensor, acl_theta_scale, acl_theta_scale_tensor,
22002200 acl_theta_scale_tensor);
2201-
2201+
22022202 // freq_scale
22032203 if (freq_scale != 1 ) {
22042204 aclnn_muls (ctx, acl_theta_scale_tensor, freq_scale, nullptr , true );
@@ -2231,7 +2231,7 @@ static void aclnn_cache_init(ggml_backend_cann_context& ctx, ggml_tensor* dst,
22312231 }
22322232
22332233 bool is_fisrt_layer = (std::strncmp (dst->name , " Qcur-0" , GGML_MAX_NAME) == 0 );
2234-
2234+
22352235 if (is_fisrt_layer || !is_attention) {
22362236
22372237 aclTensor* acl_theta_scale_tensor =
@@ -2242,13 +2242,13 @@ static void aclnn_cache_init(ggml_backend_cann_context& ctx, ggml_tensor* dst,
22422242 aclTensor* acl_position_tensor = ggml_cann_create_tensor (
22432243 src1->data , ggml_cann_type_mapping (src1->type ),
22442244 ggml_type_size (src1->type ), position_ne, position_nb, GGML_MAX_DIMS);
2245-
2245+
22462246 // power * position
22472247 int64_t theta_length = theta_scale_length * position_length;
22482248 ggml_cann_pool_alloc theta_allocator (ctx.pool (),
22492249 theta_length * sizeof (float_t ));
22502250 void * theta_buffer = theta_allocator.get ();
2251-
2251+
22522252 aclTensor* acl_theta_tensor =
22532253 ggml_cann_create_tensor (theta_buffer, ACL_FLOAT, sizeof (float_t ),
22542254 theta_ne, theta_nb, GGML_MAX_DIMS);
@@ -2277,7 +2277,7 @@ static void aclnn_cache_init(ggml_backend_cann_context& ctx, ggml_tensor* dst,
22772277 aclTensor* acl_cos_tensor = ggml_cann_create_tensor (
22782278 ctx.cos_ptr , ACL_FLOAT, sizeof (float_t ), theta_ne, theta_nb,
22792279 GGML_MAX_DIMS, ACL_FORMAT_ND);
2280-
2280+
22812281 // attn_factor
22822282 if (attn_factor != 1 ) {
22832283 aclnn_muls (ctx, acl_sin_tensor, attn_factor, nullptr , true );
0 commit comments