@@ -384,8 +384,8 @@ void ggml_gemv_q4_0_4x4_q8_0(int n, float * restrict s, size_t bs, const void *
384384 UNUSED (blocklen );
385385
386386#if defined(__ARM_FEATURE_SVE )
387- if (svcntw () == 8 ) {
388- GGML_ASSERT (!(ggml_cpu_has_sve () && (svcntw () == 8 )) &&
387+ if (ggml_sve_cnt_b == QK8_0 ) {
388+ GGML_ASSERT (!(ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
389389 "__ARM_FEATURE_SVE defined, use the Q4_0_8_8 quantization format for optimal performance" );
390390 }
391391#endif
@@ -496,8 +496,8 @@ void ggml_gemv_q4_0_4x8_q8_0(int n, float * restrict s, size_t bs, const void *
496496 UNUSED (blocklen );
497497
498498#if defined(__ARM_FEATURE_SVE )
499- if (svcntw () == 8 ) {
500- GGML_ASSERT (!(ggml_cpu_has_sve () && (svcntw () == 8 )) &&
499+ if (ggml_sve_cnt_b == QK8_0 ) {
500+ GGML_ASSERT (!(ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
501501 "__ARM_FEATURE_SVE defined, use the Q4_0_8_8 quantization format for optimal performance" );
502502 }
503503#endif
@@ -614,7 +614,7 @@ void ggml_gemv_q4_0_8x8_q8_0(int n, float * restrict s, size_t bs, const void *
614614 UNUSED (blocklen );
615615
616616#if defined(__ARM_FEATURE_SVE ) && ! ((defined(_MSC_VER )) && ! defined(__clang__ ))
617- if (svcntw () == 8 ) {
617+ if (ggml_sve_cnt_b == QK8_0 ) {
618618 const void * b_ptr = vx ;
619619 const void * a_ptr = vy ;
620620 float * res_ptr = s ;
@@ -680,12 +680,12 @@ void ggml_gemv_q4_0_8x8_q8_0(int n, float * restrict s, size_t bs, const void *
680680 return ;
681681 }
682682 else if (ggml_cpu_has_neon () && ggml_cpu_has_matmul_int8 ()) {
683- GGML_ASSERT ((ggml_cpu_has_sve () && (svcntw () == 8 )) &&
683+ GGML_ASSERT ((ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
684684 "__ARM_FEATURE_SVE for vector size of 256-bits not defined, use the Q4_0_4_8 quantization format for optimal "
685685 "performance" );
686686 }
687687 else if (ggml_cpu_has_neon ()) {
688- GGML_ASSERT (((ggml_cpu_has_sve () && (svcntw () == 8 )) || ggml_cpu_has_matmul_int8 ()) &&
688+ GGML_ASSERT (((ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) || ggml_cpu_has_matmul_int8 ()) &&
689689 "__ARM_FEATURE_SVE for vector size of 256-bits and __ARM_FEATURE_MATMUL_INT8 not defined, use the Q4_0_4_4 "
690690 "quantization format for optimal performance" );
691691 }
@@ -745,8 +745,8 @@ void ggml_gemm_q4_0_4x4_q8_0(int n, float * restrict s, size_t bs, const void *
745745 UNUSED (blocklen );
746746
747747#if defined(__ARM_FEATURE_SVE ) && defined(__ARM_FEATURE_MATMUL_INT8 )
748- if (svcntw () == 8 ) {
749- GGML_ASSERT (!(ggml_cpu_has_sve () && (svcntw () == 8 )) &&
748+ if (ggml_sve_cnt_b == QK8_0 ) {
749+ GGML_ASSERT (!(ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
750750 "__ARM_FEATURE_SVE defined, use the Q4_0_8_8 quantization format for optimal performance" );
751751 }
752752#endif
@@ -1266,8 +1266,8 @@ void ggml_gemm_q4_0_4x8_q8_0(int n, float * restrict s, size_t bs, const void *
12661266 UNUSED (blocklen );
12671267
12681268#if defined(__ARM_FEATURE_SVE ) && defined(__ARM_FEATURE_MATMUL_INT8 )
1269- if (svcntw () == 8 ) {
1270- GGML_ASSERT (!(ggml_cpu_has_sve () && (svcntw () == 8 )) &&
1269+ if (ggml_sve_cnt_b == QK8_0 ) {
1270+ GGML_ASSERT (!(ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
12711271 "__ARM_FEATURE_SVE defined, use the Q4_0_8_8 quantization format for optimal performance" );
12721272 }
12731273#endif
@@ -1728,7 +1728,7 @@ void ggml_gemm_q4_0_8x8_q8_0(int n, float * restrict s, size_t bs, const void *
17281728 UNUSED (blocklen );
17291729
17301730#if defined(__ARM_FEATURE_SVE ) && defined(__ARM_FEATURE_MATMUL_INT8 ) && ! ((defined(_MSC_VER )) && ! defined(__clang__ ))
1731- if (svcntw () == 8 ) {
1731+ if (ggml_sve_cnt_b == QK8_0 ) {
17321732 const void * b_ptr = vx ;
17331733 const void * a_ptr = vy ;
17341734 float * res_ptr = s ;
@@ -2139,12 +2139,12 @@ void ggml_gemm_q4_0_8x8_q8_0(int n, float * restrict s, size_t bs, const void *
21392139 return ;
21402140 }
21412141 else if (ggml_cpu_has_neon () && ggml_cpu_has_matmul_int8 ()) {
2142- GGML_ASSERT ((ggml_cpu_has_sve () && (svcntw () == 8 )) &&
2142+ GGML_ASSERT ((ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) &&
21432143 "__ARM_FEATURE_SVE for vector size of 256-bits not defined, use the Q4_0_4_8 quantization format for optimal "
21442144 "performance" );
21452145 }
21462146 else if (ggml_cpu_has_neon ()) {
2147- GGML_ASSERT (((ggml_cpu_has_sve () && (svcntw () == 8 )) || ggml_cpu_has_matmul_int8 ()) &&
2147+ GGML_ASSERT (((ggml_cpu_has_sve () && (ggml_sve_cnt_b == QK8_0 )) || ggml_cpu_has_matmul_int8 ()) &&
21482148 "__ARM_FEATURE_SVE for vector size of 256-bits and __ARM_FEATURE_MATMUL_INT8 not defined, use the Q4_0_4_4 "
21492149 "quantization format for optimal performance" );
21502150 }
0 commit comments