CleanMARL provides single-file, clean, and educational implementations of Deep Multi-Agent Reinforcement Learning (MARL) algorithms in PyTorch, following the same philosophy of CleanRL.
-
Implementations of key MARL algorithms: VDN, QMIX, COMA, MADDPG, FACMAC, IPPO, and MAPPO.
-
A documentation for algorithms, code and training details.
-
We support parallel environments and recurrent policies.
-
Tensorboard and Weights & Biases logging.
We provide more details in our documentation.
You can find runs in Weights & Biases: Weights & Biases
Prerequisites:
- Python >=3.9
Installation:
git clone https://github.com/AmineAndam04/cleanmarl.git
cd cleanmarl
pip install .To run experiment you can run for example:
python cleanmarl/vdn.py --env_type="pz" --env_name="simple_spread_v3" --env_family="mpe" --use_wnb --wnb_project="cleanmarl-test" --wnb_entity="cleanmarl-test" --total_timesteps=1000000
python cleanmarl/mappo.py --env_type="smaclite" --env_name="3m" - Save the checkpoints.
- Support continuous actions (MAPPO, IPPO, MADDPG, FACMAC).
- Support individual rewards.
- Support Jax-based environments (Jaxmarl ..).
- Fully Jitted training for Jax-based environments.