Skip to content

Commit 1994161

Browse files
authored
user note added to usage.rst
1 parent f99530c commit 1994161

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

docs/source/usage.rst

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -179,7 +179,7 @@ We provide a set of pre-defined models. You can access (and search) a list of av
179179

180180
Then, you can choose the one that fits best with your needs and provide it to the CEBRA model as the :py:attr:`~.CEBRA.model_architecture` parameter.
181181

182-
As an indication the table below presents the model architecture we used to train CEBRA on the datasets presented in our paper (Schneider, Lee, Mathis, 2022).
182+
As an indication the table below presents the model architecture we used to train CEBRA on the datasets presented in our paper (Schneider, Lee, Mathis. Nature 2023).
183183

184184
.. list-table::
185185
:widths: 25 25 20 30
@@ -1186,9 +1186,10 @@ Improve model performance
11861186
🧐 Below is a (non-exhaustive) list of actions you can try if your embedding looks different from what you were expecting.
11871187

11881188
#. Assess that your model `converged <https://machine-learning.paperspace.com/wiki/convergence>`_. For that, observe if the training loss stabilizes itself around the end of the training or still seems to be decreasing. Refer to `Visualize the training loss`_ for more details on how to display the training loss.
1189-
#. Increase the number of iterations. It should be at least 10,000.
1189+
#. Increase the number of iterations. It typically should be at least 10,000.
11901190
#. Make sure the batch size is big enough. It should be at least 512.
11911191
#. Fine-tune the model's hyperparameters, namely ``learning_rate``, ``output_dimension``, ``num_hidden_units`` and eventually ``temperature`` (by setting ``temperature_mode`` back to ``constant``). Refer to `Grid search`_ for more details on performing hyperparameters tuning.
1192+
#. To note, you should still be mindful of performing train/validation splits and shuffle controls to avoid `overfitting <https://developers.google.com/machine-learning/crash-course/overfitting/overfitting>`_.
11921193

11931194

11941195

0 commit comments

Comments
 (0)