File tree Expand file tree Collapse file tree 1 file changed +21
-0
lines changed Expand file tree Collapse file tree 1 file changed +21
-0
lines changed Original file line number Diff line number Diff line change 11# Migrating
22
3+ ## Migrating from ML-Agents toolkit v0.8 to v0.9
4+
5+ ### Important Changes
6+ * We have changed the way reward signals (including Curiosity) are defined in the
7+ ` trainer_config.yaml ` .
8+ * When using multiple environments, every "step" as recorded in TensorBoard and
9+ printed in the command line now corresponds to a single step of a single environment.
10+ Previously, each step corresponded to one step for all environments (i.e., ` num_envs ` steps).
11+
12+ #### Steps to Migrate
13+ * If you were overriding any of these following parameters in your config file, remove them
14+ from the top-level config and follow the steps below:
15+ * ` gamma ` - Define a new ` extrinsic ` reward signal and set it's ` gamma ` to your new gamma.
16+ * ` use_curiosity ` , ` curiosity_strength ` , ` curiosity_enc_size ` - Define a ` curiosity ` reward signal
17+ and set its ` strength ` to ` curiosity_strength ` , and ` encoding_size ` to ` curiosity_enc_size ` . Give it
18+ the same ` gamma ` as your ` extrinsic ` signal to mimic previous behavior.
19+ See [ Reward Signals] ( Training-RewardSignals.md ) for more information on defining reward signals.
20+ * TensorBoards generated when running multiple environments in v0.8 are not comparable to those generated in
21+ v0.9 in terms of step count. Multiply your v0.8 step count by ` num_envs ` for an approximate comparison.
22+ You may need to change ` max_steps ` in your config as appropriate as well.
23+
324## Migrating from ML-Agents toolkit v0.7 to v0.8
425
526### Important Changes
You can’t perform that action at this time.
0 commit comments