|
44 | 44 | from vllm.distributed.parallel_state import (get_dp_group, get_pp_group,
|
45 | 45 | get_tp_group)
|
46 | 46 | from vllm.forward_context import get_forward_context
|
47 |
| -from vllm.inputs import INPUT_REGISTRY |
48 | 47 | from vllm.logger import logger
|
49 | 48 | from vllm.model_executor.layers.fused_moe import FusedMoE
|
50 | 49 | from vllm.model_executor.layers.rotary_embedding import MRotaryEmbedding
|
51 | 50 | from vllm.model_executor.model_loader import get_model
|
52 | 51 | from vllm.model_executor.models.interfaces import supports_transcription
|
53 | 52 | from vllm.model_executor.models.interfaces_base import (
|
54 | 53 | VllmModelForPooling, is_pooling_model, is_text_generation_model)
|
55 |
| -from vllm.multimodal import MULTIMODAL_REGISTRY |
56 | 54 | from vllm.multimodal.inputs import MultiModalKwargs, PlaceholderRange
|
57 | 55 | from vllm.multimodal.utils import group_mm_inputs_by_modality
|
58 | 56 | from vllm.pooling_params import PoolingParams
|
59 | 57 | from vllm.sampling_params import SamplingType
|
60 | 58 | from vllm.sequence import IntermediateTensors
|
61 | 59 | from vllm.utils import (STR_DTYPE_TO_TORCH_DTYPE, DeviceMemoryProfiler,
|
62 | 60 | LazyLoader, cdiv)
|
63 |
| -from vllm.v1.core.encoder_cache_manager import compute_encoder_budget |
64 | 61 | from vllm.v1.kv_cache_interface import (FullAttentionSpec, KVCacheConfig,
|
65 | 62 | KVCacheSpec)
|
66 | 63 | from vllm.v1.outputs import (EMPTY_MODEL_RUNNER_OUTPUT, LogprobsTensors,
|
@@ -169,13 +166,6 @@ def __init__(self, vllm_config: VllmConfig, device: torch.device):
|
169 | 166 | self.device = device
|
170 | 167 | self.dtype = self.model_config.dtype
|
171 | 168 | self.sampler = Sampler()
|
172 |
| - # Multi-modal data support |
173 |
| - self.input_registry = INPUT_REGISTRY |
174 |
| - self.mm_registry = MULTIMODAL_REGISTRY |
175 |
| - self.max_num_encoder_input_tokens, self.encoder_cache_size = compute_encoder_budget( |
176 |
| - model_config=self.model_config, |
177 |
| - scheduler_config=self.scheduler_config, |
178 |
| - mm_registry=self.mm_registry) |
179 | 169 |
|
180 | 170 | # Lazy initialization, these will be set after __init__
|
181 | 171 | self.kv_caches: List[torch.Tensor] = []
|
|
0 commit comments