Skip to content

Commit 3d9fa8d

Browse files
authored
[BugFix]: fix a lot of bug (#1565)
Signed-off-by: princepride <wangzhipeng628@gmail.com>
1 parent cd2234a commit 3d9fa8d

File tree

5 files changed

+18
-16
lines changed

5 files changed

+18
-16
lines changed

vllm_omni/diffusion/models/bagel/pipeline_bagel.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -128,13 +128,10 @@ def __init__(self, vision_model):
128128
else:
129129
self.vision_model = vision_model
130130

131-
# Configure weights for linear equivalent of patch embedding
132-
self.patch_embed_weight = self.vision_model.embeddings.patch_embedding.weight
133-
self.patch_embed_bias = self.vision_model.embeddings.patch_embedding.bias
134-
135131
def forward(self, packed_pixel_values, packed_flattened_position_ids, cu_seqlens, max_seqlen):
136-
w = self.patch_embed_weight.view(self.patch_embed_weight.shape[0], -1)
137-
x = F.linear(packed_pixel_values, w, self.patch_embed_bias)
132+
patch_embed = self.vision_model.embeddings.patch_embedding
133+
w = patch_embed.weight.view(patch_embed.weight.shape[0], -1)
134+
x = F.linear(packed_pixel_values, w, patch_embed.bias)
138135
pos = self.vision_model.embeddings.position_embedding(packed_flattened_position_ids)
139136
x = x + pos
140137
hidden_states = x.unsqueeze(0)
@@ -205,6 +202,9 @@ def __init__(self, *, od_config: OmniDiffusionConfig, prefix: str = ""):
205202
)
206203
vit_config_path = os.path.join(model_path, "vit_config.json")
207204
vit_conf = SiglipVisionConfig.from_json_file(vit_config_path)
205+
if vit_conf.num_hidden_layers == 27:
206+
vit_conf.num_hidden_layers = 26
207+
vit_conf.vision_use_head = False
208208
self.vit_model = SiglipVisionModel(vit_conf)
209209
self.image_processor = SiglipImageProcessor.from_pretrained(model_path, local_files_only=True)
210210

vllm_omni/diffusion/models/glm_image/pipeline_glm_image.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -871,8 +871,8 @@ def forward(self, req: OmniDiffusionRequest) -> DiffusionOutput:
871871

872872
def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]) -> set[str]:
873873
"""Load transformer weights."""
874-
# Filter weights for transformer only
875874
transformer_weights = (
876-
(name.replace("transformer.", ""), weight) for name, weight in weights if name.startswith("transformer.")
875+
(name.replace("transformer.", "", 1), weight) for name, weight in weights if name.startswith("transformer.")
877876
)
878-
return self.transformer.load_weights(transformer_weights)
877+
loaded = self.transformer.load_weights(transformer_weights)
878+
return {f"transformer.{name}" for name in loaded}

vllm_omni/distributed/omni_connectors/connectors/shm_connector.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ def _get_data_with_lock(self, lock_file: str, shm_handle: dict):
8686
return obj, int(shm_handle.get("size", 0))
8787
except Exception as e:
8888
logger.error(f"SharedMemoryConnector shm get failed for req : {e}")
89-
return None, 0
89+
return None
9090
finally:
9191
# If data has been received, delete lock_file.
9292
if obj and os.path.exists(lock_file):
@@ -105,32 +105,32 @@ def get(
105105
metadata = metadata.get(get_key)
106106

107107
if not isinstance(metadata, dict):
108-
return None, 0
108+
return None
109109

110110
if "inline_bytes" in metadata:
111111
try:
112112
obj = self.deserialize_obj(metadata["inline_bytes"])
113113
return obj, int(metadata.get("size", 0))
114114
except Exception as e:
115115
logger.error(f"SharedMemoryConnector inline get failed for req {get_key}: {e}")
116-
return None, 0
116+
return None
117117

118118
if "shm" in metadata:
119119
shm_handle = metadata["shm"]
120120
lock_file = f"/dev/shm/shm_{shm_handle['name']}_lockfile.lock"
121121
return self._get_data_with_lock(lock_file, shm_handle)
122122

123-
return None, 0
123+
return None
124124
shm = None
125125
try:
126126
shm = shm_pkg.SharedMemory(name=get_key)
127127
if shm is None or shm.size == 0:
128-
return None, 0
128+
return None
129129
lock_file = f"/dev/shm/shm_{get_key}_lockfile.lock"
130130
shm_handle = {"name": get_key, "size": shm.size}
131131
return self._get_data_with_lock(lock_file, shm_handle)
132132
except Exception:
133-
return None, 0
133+
return None
134134
finally:
135135
if shm:
136136
shm.close()

vllm_omni/distributed/omni_connectors/transfer_adapter/chunk_transfer_adapter.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -136,6 +136,8 @@ def _poll_single_request(self, request: Request):
136136
logger.error(f"SharedMemoryConnector get failed for req {connector_get_key}: {e}")
137137
return False
138138

139+
if result is None:
140+
return False
139141
payload_data, size = result
140142

141143
if payload_data:

vllm_omni/model_executor/models/glm_image/glm_image_ar.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,12 +36,12 @@
3636
GlmImageVQVAEConfig,
3737
)
3838
from transformers.models.glm_image.processing_glm_image import GlmImageProcessor
39-
from vllm.attention.layer import Attention
4039
from vllm.config import CacheConfig, MultiModalConfig, VllmConfig
4140
from vllm.config.multimodal import BaseDummyOptions
4241
from vllm.distributed import get_pp_group, get_tensor_model_parallel_world_size
4342
from vllm.distributed import utils as dist_utils
4443
from vllm.logger import init_logger
44+
from vllm.model_executor.layers.attention import Attention
4545
from vllm.model_executor.layers.attention.mm_encoder_attention import (
4646
MMEncoderAttention,
4747
)

0 commit comments

Comments
 (0)