Skip to content

Support converting Qwen2.5-based MiniVLA checkpoint to HF format#2

Open
360ZMEM wants to merge 2 commits intoStanford-ILIAD:mainfrom
360ZMEM:feat/support-hf-checkpoint-export
Open

Support converting Qwen2.5-based MiniVLA checkpoint to HF format#2
360ZMEM wants to merge 2 commits intoStanford-ILIAD:mainfrom
360ZMEM:feat/support-hf-checkpoint-export

Conversation

@360ZMEM
Copy link

@360ZMEM 360ZMEM commented Dec 22, 2025

Description:
This PR modifies file vla-scripts/extern/convert_openvla_weights_to_hf.py to enable the conversion of MiniVLA checkpoints (Qwen2.5 backbone) to standard HF format.

Changes:

  1. Qwen2.5 Support: Added logic to handle num_key_value_heads (GQA) which is present in Qwen2.5 configs but not in the default Llama mapping.
  2. Extra Tokens: Implemented logic to add <|extra_i|> tokens for models trained with the extra tokenizer configuration.
  3. Vocab Alignment: Ensured vocab_size in the config matches the padded vocabulary size used during Prismatic training.

Scope:

  • Verified: Tested with prism-qwen25-extra-dinosiglip-224px+0_5b (Standard Action Binning).
  • This PR does not consider VQActionTokenizer models.

@360ZMEM 360ZMEM marked this pull request as ready for review December 22, 2025 14:51
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant