Skip to content

Commit e019fe1

Browse files
Update README.md
1 parent e0ad771 commit e019fe1

File tree

1 file changed

+16
-2
lines changed

1 file changed

+16
-2
lines changed

README.md

Lines changed: 16 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,23 @@
1-
# Popular Model-free Reinforcement Learning Algorithms [![Tweet](https://img.shields.io/twitter/url/http/shields.io.svg?style=social)](https://twitter.com/intent/tweet?text=State-of-the-art-Model-free-Reinforcement-Learning-Algorithms%20&url=hhttps://github.com/quantumiracle/STOA-RL-Algorithms&hashtags=RL)
1+
# Popular Model-free Reinforcement Learning Algorithms
2+
<!-- [![Tweet](https://img.shields.io/twitter/url/http/shields.io.svg?style=social)](https://twitter.com/intent/tweet?text=State-of-the-art-Model-free-Reinforcement-Learning-Algorithms%20&url=hhttps://github.com/quantumiracle/STOA-RL-Algorithms&hashtags=RL) -->
23

34

45
**PyTorch** and **Tensorflow 2.0** implementation of state-of-the-art model-free reinforcement learning algorithms on both Openai gym environments and a self-implemented Reacher environment.
56

6-
Algorithms include **Soft Actor-Critic (SAC), Deep Deterministic Policy Gradient (DDPG), Twin Delayed DDPG (TD3), Actor-Critic (AC/A2C), Proximal Policy Optimization (PPO), QT-Opt (including Cross-entropy (CE) Method)**, **PointNet**, **Transporter**, **Recurrent Policy Gradient**, **Soft Decision Tree**, **Probabilistic Mixture-of-Experts**, etc.
7+
Algorithms include:
8+
* **Actor-Critic (AC/A2C)**;
9+
* **Soft Actor-Critic (SAC)**;
10+
* **Deep Deterministic Policy Gradient (DDPG)**;
11+
* **Twin Delayed DDPG (TD3)**;
12+
* **Proximal Policy Optimization (PPO)**;
13+
* **QT-Opt (including Cross-entropy (CE) Method)**;
14+
* **PointNet**;
15+
* **Transporter**;
16+
* **Recurrent Policy Gradient**;
17+
* **Soft Decision Tree**;
18+
* **Probabilistic Mixture-of-Experts**;
19+
* **QMIX**
20+
* etc.
721

822
Please note that this repo is more of a personal collection of algorithms I implemented and tested during my research and study period, rather than an official open-source library/package for usage. However, I think it could be helpful to share it with others and I'm expecting useful discussions on my implementations. But I didn't spend much time on cleaning or structuring the code. As you may notice that there may be several versions of implementation for each algorithm, I intentionally show all of them here for you to refer and compare. Also, this repo contains only **PyTorch** Implementation.
923

0 commit comments

Comments
 (0)