support train qwen2.5-vl-32b eagle3 model#437
Merged
jiapingW merged 3 commits intosgl-project:mainfrom Jan 20, 2026
Merged
Conversation
Contributor
|
Warning You have reached your daily quota limit. Please wait up to 24 hours and I will start processing your requests again! |
cdd81c3 to
b0034f6
Compare
b0034f6 to
c3528ec
Compare
a5463b3 to
a8d9343
Compare
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Motivation
I noticed SpecForge currently supports Qwen2.5-VL-7B via Transformers. I think we should instead add systematic VL support through sglang, rather than a model-specific Transformers integration.
Modifications
Dataset preprocessing: Use transformers to generate pixel_values and image_grid_thw (already available).
Request packing (SGLang): Wrap the data into an sglang Request. When unpacking/splitting into per-request chunks, segment pixel_values by offset based on image_grid_thw, and ensure compatibility with mRoPE.
Draft model: During forward(), align mRoPE behavior with the main model.
Initialize mmCache: Set up the multimodal cache during initialization.
Related Issues
Fixes #403
Accuracy Test
TODO
Benchmark & Profiling
Checklist