We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent bda6115 commit 2fd90afCopy full SHA for 2fd90af
litgpt/finetune/full.py
@@ -226,7 +226,6 @@ def fit(
226
state["iter_num"] * train.micro_batch_size * model.config.block_size * fabric.world_size
227
),
228
"learning_rate": scheduler.get_last_lr()[0],
229
- "val_loss": val_loss,
230
}
231
if isinstance(val_loss, torch.Tensor):
232
val_loss = f"{val_loss:.3f}"
litgpt/finetune/lora.py
@@ -247,7 +247,6 @@ def fit(
247
iter_num * train.micro_batch_size * model.config.block_size * fabric.world_size
248
249
250
251
252
253
0 commit comments