From 9614a3f24622692999902ebaa850071afba0891d Mon Sep 17 00:00:00 2001 From: Roman Janik Date: Mon, 9 Jun 2025 17:47:16 +0200 Subject: [PATCH 1/2] NXP backend: Turn off bias tensor reuse in Convolution converter --- .../node_converters/ops_converters/convolution_converter.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py b/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py index dff003445ae..0f3a4b9bb5a 100644 --- a/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py +++ b/backends/nxp/backend/ir/converter/node_converters/ops_converters/convolution_converter.py @@ -263,7 +263,7 @@ def _convert_unpadded_2D( ) b = self.builder.create_zeros_tensor( - [output_channels], "zero_bias", bias_type, True + [output_channels], "zero_bias", bias_type, False ) # Compute scale and zero point for bias tensor From 30316cd5e8f6c3ef7ac3d8964b64a445046ecd92 Mon Sep 17 00:00:00 2001 From: Roman Janik Date: Mon, 9 Jun 2025 18:45:58 +0200 Subject: [PATCH 2/2] NXP backend: Fix _has_shared_q_params_if_quantized in Node converter --- backends/nxp/backend/ir/converter/node_converter.py | 9 ++------- 1 file changed, 2 insertions(+), 7 deletions(-) diff --git a/backends/nxp/backend/ir/converter/node_converter.py b/backends/nxp/backend/ir/converter/node_converter.py index d646e507769..ed624aaa411 100755 --- a/backends/nxp/backend/ir/converter/node_converter.py +++ b/backends/nxp/backend/ir/converter/node_converter.py @@ -132,13 +132,8 @@ def _has_shared_q_params_if_quantized(node: Node) -> bool: # Some exotic operator (only consumer or only produces) return True - pre_node = node.prev - post_node = node.next - - if pre_node.name == node.all_input_nodes[0] and post_node.name == node.users[0]: - raise RuntimeError( - "Prev & next nodes are not the same as inputs and outputs." - ) + pre_node = node.all_input_nodes[0] + post_node = list(node.users)[0] if _is_dequant_node(pre_node) and _is_quant_node(post_node): # Node is quantized