@@ -482,7 +482,21 @@ static struct ggml_backend_hexagon_context g_hexagon_mgr[GGML_HEXAGON_MAX_DEVICE
482482 .backend = nullptr ,
483483 .raw_interface = {},
484484 .raw_system_interface = {},
485- .socinfo = {}},
485+ .socinfo = {},
486+ .qnn_singlenode_graph_map = {},
487+ .work_data = nullptr ,
488+ .tasks = {},
489+ .work_size = 0 ,
490+ .desired_size = 0 ,
491+ .n_threads = 8 ,
492+ .rpc_mempool_capacity = 0 ,
493+ .rpc_mempool_len = 0 ,
494+ .rpc_mempool_usage = 0 ,
495+ .rpc_mempool = nullptr ,
496+ .rpc_mempool_handle = 0 ,
497+ .ggmlop_handle = 0 ,
498+ .domain_id = HEXAGON_CDSP,
499+ },
486500
487501 { .device = 1 ,
488502 .name = " qnn-gpu" ,
@@ -496,7 +510,21 @@ static struct ggml_backend_hexagon_context g_hexagon_mgr[GGML_HEXAGON_MAX_DEVICE
496510 .backend = nullptr ,
497511 .raw_interface = {},
498512 .raw_system_interface = {},
499- .socinfo = {}},
513+ .socinfo = {},
514+ .qnn_singlenode_graph_map = {},
515+ .work_data = nullptr ,
516+ .tasks = {},
517+ .work_size = 0 ,
518+ .desired_size = 0 ,
519+ .n_threads = 8 ,
520+ .rpc_mempool_capacity = 0 ,
521+ .rpc_mempool_len = 0 ,
522+ .rpc_mempool_usage = 0 ,
523+ .rpc_mempool = nullptr ,
524+ .rpc_mempool_handle = 0 ,
525+ .ggmlop_handle = 0 ,
526+ .domain_id = HEXAGON_CDSP,
527+ },
500528
501529 { .device = 2 ,
502530 .name = " qnn-npu" ,
@@ -510,7 +538,21 @@ static struct ggml_backend_hexagon_context g_hexagon_mgr[GGML_HEXAGON_MAX_DEVICE
510538 .backend = nullptr ,
511539 .raw_interface = {},
512540 .raw_system_interface = {},
513- .socinfo = {}},
541+ .socinfo = {},
542+ .qnn_singlenode_graph_map = {},
543+ .work_data = nullptr ,
544+ .tasks = {},
545+ .work_size = 0 ,
546+ .desired_size = 0 ,
547+ .n_threads = 8 ,
548+ .rpc_mempool_capacity = 0 ,
549+ .rpc_mempool_len = 0 ,
550+ .rpc_mempool_usage = 0 ,
551+ .rpc_mempool = nullptr ,
552+ .rpc_mempool_handle = 0 ,
553+ .ggmlop_handle = 0 ,
554+ .domain_id = HEXAGON_CDSP,
555+ },
514556};
515557
516558static domain hexagon_supported_domains[] = {
@@ -3857,7 +3899,9 @@ static Qnn_Tensor_t * ggmlqnn_create_general_tensor(qnn_instance * instance, Qnn
38573899 .dataFormat = QNN_TENSOR_DATA_FORMAT_FLAT_BUFFER,
38583900 .dataType = qnn_data_type,
38593901 .quantizeParams = {.encodingDefinition = QNN_DEFINITION_UNDEFINED,
3860- .quantizationEncoding = QNN_QUANTIZATION_ENCODING_UNDEFINED},
3902+ .quantizationEncoding = QNN_QUANTIZATION_ENCODING_UNDEFINED,
3903+ .scaleOffsetEncoding = {.scale = 0 .0000000000000000f , .offset = 0 }
3904+ },
38613905 .rank = rank,
38623906 .dimensions = tensor_dims,
38633907 .memType = QNN_TENSORMEMTYPE_RAW,
@@ -4559,12 +4603,6 @@ static void ggmlqnn_compute_rms_norm(ggml_backend_hexagon_context * ctx, ggml_te
45594603 GGML_UNUSED (dst);
45604604}
45614605
4562- static void ggmlqnn_compute_diag_mask (ggml_backend_hexagon_context * ctx, ggml_tensor * dst, float value) {
4563- GGML_UNUSED (ctx);
4564- GGML_UNUSED (dst);
4565- GGML_UNUSED (value);
4566- }
4567-
45684606static void ggmlqnn_compute_im2col (ggml_backend_hexagon_context * ctx, ggml_tensor * dst) {
45694607 GGML_UNUSED (ctx);
45704608 GGML_UNUSED (dst);
@@ -5203,9 +5241,8 @@ static void ggmlhexagon_deinit_cdsp(ggml_backend_hexagon_context * ctx) {
52035241 hexagon_error = ggmlop_dsp_close (ctx->ggmlop_handle );
52045242 if (AEE_SUCCESS != hexagon_error) {
52055243 GGMLHEXAGON_LOG_WARN (" error 0x%x: failed to close ggmlop dsp handle" , hexagon_error);
5206- } else {
5207- ctx->ggmlop_handle = 0 ;
52085244 }
5245+ ctx->ggmlop_handle = 0 ;
52095246 }
52105247
52115248 ggmlhexagon_deinit_rpcmempool (ctx);
@@ -5722,9 +5759,6 @@ static bool ggmlhexagon_compute_forward(ggml_backend_t backend, struct ggml_tens
57225759 case GGML_OP_PERMUTE:
57235760 case GGML_OP_TRANSPOSE:
57245761 break ;
5725- case GGML_OP_DIAG_MASK_INF:
5726- ggmlqnn_compute_diag_mask (ctx, dst, -INFINITY);
5727- break ;
57285762 case GGML_OP_SOFT_MAX:
57295763 ggmlqnn_compute_softmax (ctx, dst);
57305764 break ;
0 commit comments