From 9a432924ab09773b16b8e4dc08093b565c6a067d Mon Sep 17 00:00:00 2001 From: d7howe <623640221@qq.com> Date: Fri, 17 Oct 2025 17:03:14 +0800 Subject: [PATCH] =?UTF-8?q?fix:=E4=BF=AE=E5=A4=8D=E8=BF=90=E8=A1=8C?= =?UTF-8?q?=E6=8A=A5=E9=94=99transpose=5Fweight=5Fkeys=E4=B8=8D=E5=AD=98?= =?UTF-8?q?=E5=9C=A8?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- paddlex/inference/models/common/vlm/transformers/model_utils.py | 2 ++ 1 file changed, 2 insertions(+) diff --git a/paddlex/inference/models/common/vlm/transformers/model_utils.py b/paddlex/inference/models/common/vlm/transformers/model_utils.py index aaa7600543..5920c8ae05 100644 --- a/paddlex/inference/models/common/vlm/transformers/model_utils.py +++ b/paddlex/inference/models/common/vlm/transformers/model_utils.py @@ -1607,6 +1607,7 @@ def _fuse_or_split_keys( except NotImplementedError: if convert_from_hf: raise ValueError("`convert_from_hf=True` is not supported") + transpose_weight_keys = None state_dict = load_state_dict( shard_file, tp_actions if pre_tensor_parallel_split else None, @@ -1934,6 +1935,7 @@ def from_pretrained( except NotImplementedError: if convert_from_hf: raise ValueError("`convert_from_hf=True` is not supported") + transpose_weight_keys = None state_dict = load_state_dict( resolved_archive_file, convert_from_hf=convert_from_hf,