(ICML 2022) PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
Original PyTorch implementation of PMIC from PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
PMIC is a MARL framework For more effective MI-driven collaboration. In PMIC, we use a new collaboration criterion measured by the MI between global states and joint actions. Based on the criterion, the key idea of PMIC is maximizing the MI associated with superior collaborative behaviors and minimizing the MI associated with inferior ones. The two MI objectives play complementary roles by facilitating learning towards better collaborations while avoiding falling into sub-optimal ones. Specifically, PMIC stores and progressively maintains sets of superior and inferior interaction experiences, from which dual MI neural estimators are established.
Known dependencies: Python (3.6.13), OpenAI gym (0.10.5), torch (1.8.1+cu102), numpy (1.19.5), Multi-agent Particle Environment
--env_name
: defines which environment in the MPE is to be used (default:"simple"
)--max_adv_c
coefficient of maximizing mutual information (i.e., \alpha in paper)--min_adv_c
coefficient of minimizing mutual information (i.e., \beta in paper)--gpu-no
: gpu num (default:-1
)--MI_update_freq
update frequency of MINE and CLUB (default:1
)--max_length
total time steps
run the run.sh
file directly
-
./algorithms/mpe_PMIC_MADDPG.py
: core code for the PMIC-MADDPG algorithm -
./run_PMIC_MADDPG.py
: code to run the PMIC-MADDPG algorithm -
./ma_utils.py
: replay buffer and DPCB -
./logs
: The folder where the logs are stored -
./tensorboard
: The folder where the tensorboards are stored -
./output
: The folder where the printed logs are stored -
./run.sh
: command-line file
If you use our method or code in your research, please consider citing the paper as follows:
@article{li2022pmic,
title={PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration},
author={Li, Pengyi and Tang, Hongyao and Yang, Tianpei and Hao, Xiaotian and Sang, Tong and Zheng, Yan and Hao, Jianye and Taylor, Matthew E and Wang, Zhen},
journal={ICML 2022},
year={2022}
}
PMIC is licensed under the MIT license. MPE are licensed under the Apache 2.0 license.