diff --git a/src/lightning/fabric/connector.py b/src/lightning/fabric/connector.py index 55b4af2728e6f..ac3cc7c13851c 100644 --- a/src/lightning/fabric/connector.py +++ b/src/lightning/fabric/connector.py @@ -428,7 +428,7 @@ def _check_strategy_and_fallback(self) -> None: if strategy_flag in _DDP_FORK_ALIASES and "fork" not in torch.multiprocessing.get_all_start_methods(): raise ValueError( f"You selected `Fabric(strategy='{strategy_flag}')` but process forking is not supported on this" - f" platform. We recommed `Fabric(strategy='ddp_spawn')` instead." + f" platform. We recommend `Fabric(strategy='ddp_spawn')` instead." ) if ( strategy_flag in _FSDP_ALIASES or type(self._strategy_flag) is FSDPStrategy diff --git a/src/lightning/pytorch/callbacks/pruning.py b/src/lightning/pytorch/callbacks/pruning.py index 1517ef6920b0d..a55c21a5c6ed1 100644 --- a/src/lightning/pytorch/callbacks/pruning.py +++ b/src/lightning/pytorch/callbacks/pruning.py @@ -129,7 +129,7 @@ def __init__( - ``bool``. Always apply it or not. - ``Callable[[epoch], bool]``. For dynamic values. Will be called every epoch. - make_pruning_permanent: Whether to remove all reparametrization pre-hooks and apply masks + make_pruning_permanent: Whether to remove all reparameterization pre-hooks and apply masks when training ends or the model is saved. use_lottery_ticket_hypothesis: See `The lottery ticket hypothesis `_: