Skip to content

[Feature] Support training Qwen2.5-VL-32B eagle model #403

@gerayking

Description

@gerayking

Checklist

Motivation

I noticed SpecForge currently supports Qwen2.5-VL-7B via Transformers. I think we should instead add systematic VL support through sglang, rather than a model-specific Transformers integration.

        if (
            args.is_vlm
            and draft_model_config.target_model_type == "qwen2_5_vl"
            and args.tp_size == 1
        ):
            from transformers import Qwen2_5_VLForConditionalGeneration

            target_model = (
                Qwen2_5_VLForConditionalGeneration.from_pretrained(
                    pretrained_model_name_or_path=args.target_model_path,
                    torch_dtype=torch.bfloat16,
                )
                .eval()
                .cuda()
            )

Related resources

No response

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions