@@ -105,7 +105,7 @@ def _save_epoch(self, ckpt: Any, name, epoch, steps, hps):
105105 opt ["weight" ][key ] = ckpt [key ].half ()
106106 opt ["config" ] = hps
107107 opt ["info" ] = "%sepoch_%siteration" % (epoch , steps )
108- # torch.save(opt, "%s/%s.pth" % (hps.save_weight_dir, name))
108+
109109 ckpt .save_with_torch (opt , os .path .join (hps .save_weight_dir , f"{ name } .pth" ))
110110 return "Success"
111111 except :
@@ -488,14 +488,14 @@ def _train_and_evaluate(
488488 )
489489 self .step += 1
490490 if epoch % hps .train .save_every_epoch == 0 and rank == 0 :
491- if hps .train .if_save_latest == 0 :
491+ if not hps .train .if_save_latest :
492492 ckpt .save_checkpoint (
493493 net_g ,
494494 optim_g ,
495495 hps .train .learning_rate ,
496496 epoch ,
497497 os .path .join (
498- hps .data . train_logs_dir , "G_{}.pth" . format ( self .step )
498+ hps .save_weight_dir , f"sovits_G_epoch { epoch } _step { self .step } .pth"
499499 ),
500500 )
501501 ckpt .save_checkpoint (
@@ -504,7 +504,7 @@ def _train_and_evaluate(
504504 hps .train .learning_rate ,
505505 epoch ,
506506 os .path .join (
507- hps .data . train_logs_dir , "D_{}.pth" . format ( self .step )
507+ hps .save_weight_dir , f"sovits_D_epoch { epoch } _step { self .step } .pth"
508508 ),
509509 )
510510 else :
@@ -514,7 +514,7 @@ def _train_and_evaluate(
514514 hps .train .learning_rate ,
515515 epoch ,
516516 os .path .join (
517- hps .data . train_logs_dir , "G_{} .pth" . format ( 233333333333 )
517+ hps .save_weight_dir , "sovits_G_latest .pth"
518518 ),
519519 )
520520 ckpt .save_checkpoint (
@@ -523,7 +523,7 @@ def _train_and_evaluate(
523523 hps .train .learning_rate ,
524524 epoch ,
525525 os .path .join (
526- hps .data . train_logs_dir , "D_{} .pth" . format ( 233333333333 )
526+ hps .save_weight_dir , "sovits_D_latest .pth"
527527 ),
528528 )
529529 if rank == 0 and hps .train .if_save_every_weights == True :
0 commit comments