Skip to content

Commit 4279620

Browse files
author
Ervin T
authored
Added Migrating docs for 0.9 (#2347)
1 parent d26a502 commit 4279620

File tree

1 file changed

+21
-0
lines changed

1 file changed

+21
-0
lines changed

docs/Migrating.md

Lines changed: 21 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,26 @@
11
# Migrating
22

3+
## Migrating from ML-Agents toolkit v0.8 to v0.9
4+
5+
### Important Changes
6+
* We have changed the way reward signals (including Curiosity) are defined in the
7+
`trainer_config.yaml`.
8+
* When using multiple environments, every "step" as recorded in TensorBoard and
9+
printed in the command line now corresponds to a single step of a single environment.
10+
Previously, each step corresponded to one step for all environments (i.e., `num_envs` steps).
11+
12+
#### Steps to Migrate
13+
* If you were overriding any of these following parameters in your config file, remove them
14+
from the top-level config and follow the steps below:
15+
* `gamma` - Define a new `extrinsic` reward signal and set it's `gamma` to your new gamma.
16+
* `use_curiosity`, `curiosity_strength`, `curiosity_enc_size` - Define a `curiosity` reward signal
17+
and set its `strength` to `curiosity_strength`, and `encoding_size` to `curiosity_enc_size`. Give it
18+
the same `gamma` as your `extrinsic` signal to mimic previous behavior.
19+
See [Reward Signals](Training-RewardSignals.md) for more information on defining reward signals.
20+
* TensorBoards generated when running multiple environments in v0.8 are not comparable to those generated in
21+
v0.9 in terms of step count. Multiply your v0.8 step count by `num_envs` for an approximate comparison.
22+
You may need to change `max_steps` in your config as appropriate as well.
23+
324
## Migrating from ML-Agents toolkit v0.7 to v0.8
425

526
### Important Changes

0 commit comments

Comments
 (0)