Skip to content

Commit 146e3bd

Browse files
MMathisLabstes
andauthored
Update docs/source/usage.rst
Co-authored-by: Steffen Schneider <[email protected]>
1 parent 81e9b4d commit 146e3bd

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/source/usage.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1184,7 +1184,7 @@ Improve model performance
11841184
🧐 Below is a (non-exhaustive) list of actions you can try if your embedding looks different from what you were expecting.
11851185

11861186
#. Assess that your model `converged <https://machine-learning.paperspace.com/wiki/convergence>`_. For that, observe if the training loss stabilizes itself around the end of the training or still seems to be decreasing. Refer to `Visualize the training loss`_ for more details on how to display the training loss.
1187-
#. Increase the number of iterations. It typically should be at least 10,000.
1187+
#. Increase the number of iterations. It typically should be at least 10,000. On small datasets, it can make sense to stop training earlier to avoid overfitting effects.
11881188
#. Make sure the batch size is big enough. It should be at least 512.
11891189
#. Fine-tune the model's hyperparameters, namely ``learning_rate``, ``output_dimension``, ``num_hidden_units`` and eventually ``temperature`` (by setting ``temperature_mode`` back to ``constant``). Refer to `Grid search`_ for more details on performing hyperparameters tuning.
11901190
#. To note, you should still be mindful of performing train/validation splits and shuffle controls to avoid `overfitting <https://developers.google.com/machine-learning/crash-course/overfitting/overfitting>`_.

0 commit comments

Comments
 (0)