You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* initial iq learn
* update Max entropy algorithms
* version using max not softmax with working grad
* working MaxEntropyDeepIRL
* refactor
* refactor demos
* final MaxEntropyDeep
The expert demonstrations for the Mountaincar-v0 are the same as used in [lets-do-irl](https://github.com/reinforcement-learning-kr/lets-do-irl/tree/master/mountaincar/maxent).
19
+
20
+
*Heatmap of Expert demonstrations with 400 states*:
[2][Wulfmeier, et al., "Maximum entropy deep inverse reinforcement learning." arXiv preprint arXiv:1507.04888 (2015).](https://arxiv.org/abs/1507.04888)
23
58
24
-
[3][Xi-liang Chen, et al., "A Study of Continuous Maximum Entropy Deep Inverse Reinforcement Learning", Mathematical Problems in Engineering, vol. 2019, Article ID 4834516, 8 pages, 2019. https://doi.org/10.1155/2019/4834516](https://www.hindawi.com/journals/mpe/2019/4834516/)
59
+
### Deep Maximum Entropy Inverse Reinforcement Learning
60
+
61
+
IRL using Deep Q-Learning with a Maximum Entropy update function.
0 commit comments