Replies: 1 comment 1 reply
-
It looks like input normalization is performed per-entity: https://github.com/entity-neural-network/incubator/blob/main/rogue_net/rogue_net/embedding_creator.py so that answers one question, the other question about how to "warm up" rare entities still stands I think. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I think there was a dicussion (that I can't find anymore) where people wre talking about input normalisation. I can probably move this discussion to the existing place. Or remove this if the question has been adequately answered and we have a solution.
I'm no expert in feature normalization I only know "its good and you should do it in PPO brrr", so maybe this question has already been answered
An example:
A particular environment may have several entities with "health" values that all have different max values.
During training it is very common to see an entity with low max health. So this health value is normalized around this "low health entity"
At some point the agent learns some policy which results in the introduction of an entity with significantly higher max health value.
I can see here two problems:
maybe we need to force environment wrappers to normalize all scalar values between 0-1?
normalization could also be seperate per entity... but then how do we "warm up" rare entities that are only seen in specific high trained policies?
Beta Was this translation helpful? Give feedback.
All reactions