Skip to content

Commit c7beb6f

Browse files
minor docs fixes
1 parent 3ac56ac commit c7beb6f

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

docs/conf.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -62,4 +62,5 @@
6262
"onnxruntime",
6363
"wandb",
6464
"neptune",
65+
"git",
6566
]

docs/guide/overview.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,15 +69,15 @@ Extensions
6969
predictor network and a fixed target network. The implementation supports selecting dedicated observation groups for
7070
curiosity, optional state and reward normalization, and configurable weight schedules for annealing the intrinsic
7171
reward contribution over the training. This extension is compatible with the :class:`~rsl_rl.algorithms.ppo.PPO`
72-
algorithm. For more details, please check `this paper <https://proceedings.mlr.press/v229/schwarke23a.html>`_.
72+
algorithm. For more details, please check `this paper <https://proceedings.mlr.press/v229/schwarke23a.html>`__.
7373

7474
Symmetry
7575
Symmetry augments the collected environment interaction data with mirrored data using a user-provided symmetry
7676
function that defines how observations and actions are transformed. This can improve sample efficiency and promote
7777
symmetric behaviors for robots with structured morphology. Additionally, a mirror-loss regularization term can be
7878
added to the loss function to actively encourage symmetry in the policy. This extension is compatible with the
7979
:class:`~rsl_rl.algorithms.ppo.PPO` algorithm. For more details, please check
80-
`this paper <https://arxiv.org/abs/2403.04359>`_.
80+
`this paper <https://arxiv.org/abs/2403.04359>`__.
8181

8282
Loggers
8383
^^^^^^^

0 commit comments

Comments
 (0)