We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1aa1361 commit 24cde76Copy full SHA for 24cde76
vllm/model_executor/models/llama.py
@@ -322,9 +322,10 @@ def load_weights(self,
322
model_name_or_path, cache_dir, load_format, revision):
323
if "rotary_emb.inv_freq" in name:
324
continue
325
- if "rotary_emb.cos_cached" in name:
326
- continue
327
- if "rotary_emb.sin_cached" in name:
+ if ("rotary_emb.cos_cached" in name
+ or "rotary_emb.sin_cached" in name):
+ # Models trained using ColossalAI may include these tensors in
328
+ # the checkpoint. Skip them.
329
330
for (param_name, weight_name, shard_id) in stacked_params_mapping:
331
if weight_name not in name:
0 commit comments