File tree Expand file tree Collapse file tree 2 files changed +0
-11
lines changed Expand file tree Collapse file tree 2 files changed +0
-11
lines changed Original file line number Diff line number Diff line change @@ -413,11 +413,6 @@ async def _load_tensor_parallel_state_dict(
413413 # setting explictly to llama3 for now as its our only use case
414414 sharding = VLLMSharding (self .tensor_parallel_size , self .rank )
415415
416- # key_str = ""
417- # for param_name in current_state_dict.keys():
418- # key_str += f" {self.state_dict_key}/{version}/{param_name}\n"
419- # logger.warning(f"############### policy get keys : {key_str}")
420-
421416 for param_name in current_state_dict .keys ():
422417 current_tensor = current_state_dict [param_name ]
423418 # Load the full tensor from torchstore
Original file line number Diff line number Diff line change @@ -304,12 +304,6 @@ async def push_weights(self) -> None:
304304 [gate , up ], dim = 0
305305 )
306306
307- # Remove before landing
308- # key_str = ""
309- # for key, _ in hf_state_dict.items():
310- # key_str += f" model_state_dict/{self.current_step}/{key}\n"
311- # logger.warning(f"rltrainer, put_state_dict keys : {key_str}")
312-
313307 await ts .put_state_dict (
314308 state_dict = hf_state_dict ,
315309 key = f"model_state_dict/{ self .current_step } " ,
You can’t perform that action at this time.
0 commit comments