@@ -229,7 +229,7 @@ def optimizer_step(
229
229
\**kwargs: Keyword arguments to to ``optimizer.step``
230
230
"""
231
231
model = model or self .lightning_module
232
- # TODO(lite ): remove assertion once strategy's optimizer_step typing is fixed
232
+ # TODO(fabric ): remove assertion once strategy's optimizer_step typing is fixed
233
233
assert isinstance (model , pl .LightningModule )
234
234
return self .precision_plugin .optimizer_step (
235
235
optimizer , model = model , optimizer_idx = opt_idx , closure = closure , ** kwargs
@@ -241,19 +241,19 @@ def _setup_model_and_optimizers(self, model: Module, optimizers: List[Optimizer]
241
241
The returned objects are expected to be in the same order they were passed in. The default implementation will
242
242
call :meth:`_setup_model` and :meth:`_setup_optimizer` on the inputs.
243
243
"""
244
- # TODO: standardize this across all plugins in Lightning and Lite . Related refactor: #7324
244
+ # TODO: standardize this across all plugins in Lightning and Fabric . Related refactor: #7324
245
245
model = self ._setup_model (model )
246
246
optimizers = [self ._setup_optimizer (optimizer ) for optimizer in optimizers ]
247
247
return model , optimizers
248
248
249
249
def _setup_model (self , model : Module ) -> Module :
250
250
"""Performs setup for the model, e.g., by wrapping it by another class."""
251
- # TODO: standardize this across all plugins in Lightning and Lite . Related refactor: #7324
251
+ # TODO: standardize this across all plugins in Lightning and Fabric . Related refactor: #7324
252
252
return model
253
253
254
254
def _setup_optimizer (self , optimizer : Optimizer ) -> Optimizer :
255
255
"""Performs setup for the optimizer, e.g., by wrapping it by another class."""
256
- # TODO: standardize this across all plugins in Lightning and Lite . Related refactor: #7324
256
+ # TODO: standardize this across all plugins in Lightning and Fabric . Related refactor: #7324
257
257
return optimizer
258
258
259
259
def batch_to_device (self , batch : Any , device : Optional [torch .device ] = None , dataloader_idx : int = 0 ) -> Any :
0 commit comments