Skip to content

Commit a134884

Browse files
author
zhouwg
committed
ggml-qnn: enable release build with necessary logs to make reviewers happy
1 parent 302e014 commit a134884

File tree

4 files changed

+8
-9
lines changed

4 files changed

+8
-9
lines changed

ggml/src/ggml-qnn/ggml-qnn-impl.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,7 @@ class qnn_instance;
8989
struct ggml_backend_qnn_context;
9090
void ggmlqnn_log_internal(ggml_log_level level, const char * file, const char * func, int line, const char * format, ...);
9191

92-
#ifdef NDEBUG
92+
#if 0//def NDEBUG
9393
#define GGMLQNN_DEBUG 0
9494
#define ENABLE_QNNBACKEND_PERF 0 // enable/disable op's perf info
9595
#define GGMLQNN_PRINT_QNN_INTERNAL_LOG 0 // enable/disable QNN's internal log

ggml/src/ggml-qnn/ggml-qnn-ops.cpp

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -91,9 +91,7 @@ void ggml_qnn_general_node(ggml_backend_qnn_context * ctx, ggml_tensor * op) {
9191
p_tensor1 = ggmlqnn_create_compute_tensor(src1);
9292
p_tensor2 = ggmlqnn_create_compute_tensor(dst);
9393
}
94-
#if GGMLQNN_PRINT_OP_ADD_LOG
95-
print_tensors_info(__func__, ctx, src0, src1, dst);
96-
#endif
94+
ggmlqnn_print_tensors_info(__func__, ctx, src0, src1, dst);
9795

9896
//ensure QNN tensor has correct tensor type
9997
QNN_VER_PTR(*p_tensor0)->type = QNN_TENSOR_TYPE_APP_WRITE;

ggml/src/ggml-qnn/ggml-qnn.cpp

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2232,7 +2232,7 @@ static bool ggml_qnn_can_handle_op(const ggml_backend_qnn_context * ctx, const s
22322232
}
22332233

22342234
if (tensor->op == GGML_OP_MUL_MAT) {
2235-
dump_op_info(tensor);
2235+
//dump_op_info(tensor);
22362236
if (src0_rank != src1_rank) // make QNN SDK happy
22372237
return false;
22382238
if (src0_rank < 2) // QNN's limitation, make QNN SDK happy
@@ -2244,15 +2244,16 @@ static bool ggml_qnn_can_handle_op(const ggml_backend_qnn_context * ctx, const s
22442244

22452245
if (ctx->device == QNN_BACKEND_NPU)
22462246
if (2 == src0_rank)
2247-
return (src0->type == GGML_TYPE_F32 || src0->type == GGML_TYPE_F16
2247+
return (src0->type == GGML_TYPE_F32
22482248
|| src0->type == GGML_TYPE_Q4_0 || src0->type == GGML_TYPE_Q8_0
22492249
|| src0->type == GGML_TYPE_Q6_K || src0->type == GGML_TYPE_Q8_K
22502250
) && (src1->type == GGML_TYPE_F32) && (tensor->type == GGML_TYPE_F32);
22512251
else
22522252
return (src0->type == GGML_TYPE_F32) && (src1->type == GGML_TYPE_F32) && (tensor->type == GGML_TYPE_F32);
22532253
else
2254-
return (src0->type == GGML_TYPE_F32 || src0->type == GGML_TYPE_F16 || ggml_is_quantized(src0->type))
2255-
&& (src1->type == GGML_TYPE_F32) && (tensor->type == GGML_TYPE_F32);
2254+
return (src0->type == GGML_TYPE_F32 || src0->type == GGML_TYPE_Q4_0
2255+
|| src0->type == GGML_TYPE_Q8_0 || src0->type == GGML_TYPE_Q6_K || src0->type == GGML_TYPE_Q8_K)
2256+
&& (src1->type == GGML_TYPE_F32) && (tensor->type == GGML_TYPE_F32);
22562257
}
22572258

22582259
if (tensor->op == GGML_OP_MUL) {

scripts/build-run-android.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ function check_and_download_ndk()
7474

7575
function build_arm64
7676
{
77-
cmake -H. -B./out/android -DCMAKE_BUILD_TYPE=Debug -DGGML_USE_QNN=ON -DCMAKE_TOOLCHAIN_FILE=${ANDROID_NDK}/build/cmake/android.toolchain.cmake -DANDROID_ABI=arm64-v8a -DANDROID_PLATFORM=latest -DCMAKE_C_FLAGS=-march=armv8.7-a -DGGML_QNN=ON -DGGML_QNN_SDK_PATH=${QNN_SDK_PATH}
77+
cmake -H. -B./out/android -DCMAKE_BUILD_TYPE=Release -DGGML_USE_QNN=ON -DCMAKE_TOOLCHAIN_FILE=${ANDROID_NDK}/build/cmake/android.toolchain.cmake -DANDROID_ABI=arm64-v8a -DANDROID_PLATFORM=latest -DCMAKE_C_FLAGS=-march=armv8.7-a -DGGML_QNN=ON -DGGML_QNN_SDK_PATH=${QNN_SDK_PATH}
7878
cd out/android
7979
make -j16
8080
show_pwd

0 commit comments

Comments
 (0)