Skip to content
Discussion options

You must be logged in to vote

This is not possible from the yaml file. If you need to customize the learning rate scheduler configuration, you can do so by overriding:

class MyLightningCLI(LightningCLI):
    @staticmethod
    def configure_optimizers(lightning_module, optimizer, lr_scheduler=None):
        if lr_scheduler is None:
            return optimizer
        return {
            "optimizer": optimizer,
            "lr_scheduler": {"scheduler": lr_scheduler, "interval": "step"},
        }

Replies: 2 comments 3 replies

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
3 replies
@ForJadeForest
Comment options

@carmocca
Comment options

@ForJadeForest
Comment options

Answer selected by carmocca
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment