Skip to content

Commit 24cde76

Browse files
authored
[Minor] Add comment on skipping rope caches (#2004)
1 parent 1aa1361 commit 24cde76

File tree

1 file changed

+4
-3
lines changed

1 file changed

+4
-3
lines changed

vllm/model_executor/models/llama.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -322,9 +322,10 @@ def load_weights(self,
322322
model_name_or_path, cache_dir, load_format, revision):
323323
if "rotary_emb.inv_freq" in name:
324324
continue
325-
if "rotary_emb.cos_cached" in name:
326-
continue
327-
if "rotary_emb.sin_cached" in name:
325+
if ("rotary_emb.cos_cached" in name
326+
or "rotary_emb.sin_cached" in name):
327+
# Models trained using ColossalAI may include these tensors in
328+
# the checkpoint. Skip them.
328329
continue
329330
for (param_name, weight_name, shard_id) in stacked_params_mapping:
330331
if weight_name not in name:

0 commit comments

Comments
 (0)