Skip to content

Commit 1037b36

Browse files
committed
Use get_input_embeddings
Signed-off-by: Rahul Tuli <[email protected]>
1 parent 1c1d679 commit 1037b36

File tree

4 files changed

+13
-12
lines changed

4 files changed

+13
-12
lines changed

vllm/model_executor/models/llama_eagle3.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -248,8 +248,7 @@ def get_input_embeddings(
248248
multimodal_embeddings: Optional[NestedTensors] = None,
249249
is_multimodal: Optional[torch.Tensor] = None,
250250
) -> torch.Tensor:
251-
# The llama3 drafter only processes text embeddings
252-
return self.model.embed_tokens(input_ids)
251+
return self.model.get_input_embeddings(input_ids)
253252

254253
def forward(
255254
self,

vllm/model_executor/models/mllama4.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -64,8 +64,12 @@
6464
from vllm.sequence import IntermediateTensors
6565
from vllm.utils.tensor_schema import TensorSchema, TensorShape
6666

67-
from .interfaces import (MultiModalEmbeddings, SupportsEagle3,
68-
SupportsMultiModal, SupportsPP)
67+
from .interfaces import (
68+
MultiModalEmbeddings,
69+
SupportsEagle3,
70+
SupportsMultiModal,
71+
SupportsPP,
72+
)
6973
from .llama4 import Llama4ForCausalLM
7074
from .utils import AutoWeightsLoader, flatten_bn, maybe_prefix
7175
from .vision import run_dp_sharded_vision_model
@@ -773,7 +777,7 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
773777
def set_aux_hidden_state_layers(self, layers: tuple[int, ...]) -> None:
774778
"""Set which layers should output auxiliary hidden states for EAGLE3."""
775779
# Delegate to underlying language model (Llama4ForCausalLM)
776-
assert hasattr(self.language_model, 'set_aux_hidden_state_layers')
780+
assert hasattr(self.language_model, "set_aux_hidden_state_layers")
777781
self.language_model.set_aux_hidden_state_layers(layers)
778782

779783
def get_eagle3_aux_hidden_state_layers(self) -> tuple[int, ...]:
@@ -783,9 +787,7 @@ def get_eagle3_aux_hidden_state_layers(self) -> tuple[int, ...]:
783787
the speculative config if available, providing dynamic configuration.
784788
"""
785789
# Delegate to underlying language model (Llama4ForCausalLM)
786-
assert hasattr(
787-
self.language_model, "get_eagle3_aux_hidden_state_layers"
788-
)
790+
assert hasattr(self.language_model, "get_eagle3_aux_hidden_state_layers")
789791
return self.language_model.get_eagle3_aux_hidden_state_layers()
790792

791793
def _parse_and_validate_image_input(

vllm/transformers_utils/configs/speculators/algos.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,4 +34,5 @@ def update_eagle3(config_dict: dict, vllm_config: dict) -> None:
3434
vllm_config["architectures"] = ["Eagle3LlamaForCausalLM"]
3535
if config_dict.get("eagle_aux_hidden_state_layer_ids"):
3636
vllm_config["eagle_aux_hidden_state_layer_ids"] = config_dict[
37-
"eagle_aux_hidden_state_layer_ids"]
37+
"eagle_aux_hidden_state_layer_ids"
38+
]

vllm/v1/worker/gpu_model_runner.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3026,12 +3026,11 @@ def _get_eagle3_aux_layers_from_config(self) -> Optional[tuple[int, ...]]:
30263026
Tuple of layer indices if found in draft model config,
30273027
None otherwise.
30283028
"""
3029-
if not (self.speculative_config
3030-
and self.speculative_config.draft_model_config):
3029+
if not (self.speculative_config and self.speculative_config.draft_model_config):
30313030
return None
30323031

30333032
hf_config = self.speculative_config.draft_model_config.hf_config
3034-
if not hasattr(hf_config, 'eagle_aux_hidden_state_layer_ids'):
3033+
if not hasattr(hf_config, "eagle_aux_hidden_state_layer_ids"):
30353034
return None
30363035

30373036
layer_ids = hf_config.eagle_aux_hidden_state_layer_ids

0 commit comments

Comments
 (0)