Skip to content

Commit 74061f5

Browse files
authored
Avoid deepcopy megatron model config (#349)
Signed-off-by: Chenhan Yu <[email protected]>
1 parent c60baae commit 74061f5

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

modelopt/torch/opt/plugins/mcore_dist_checkpointing.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,9 @@ def _parse_transformer_config(transformer_config: dict) -> dict:
154154

155155
if dist.is_master():
156156
run_config_name = f"{checkpoint_name}/modelopt_run_config.yaml"
157-
config_dict = _parse_transformer_config(copy.deepcopy(model[0].config.__dict__))
157+
# We avoid deepcopy here since some attributes in Megatron-Bridge config cannot be
158+
# deepcopy.
159+
config_dict = _parse_transformer_config(model[0].config.__dict__)
158160
config_dict["nvidia_modelopt_version"] = modelopt.__version__
159161
with open(run_config_name, "w") as f:
160162
yaml.dump(config_dict, f, default_flow_style=False)

0 commit comments

Comments
 (0)