Sparse Adversarial Attack in Multi-agent Reinforcement Learning [article]

Yizheng Hu, Zhihua Zhang
2022 arXiv   pre-print
Cooperative multi-agent reinforcement learning (cMARL) has many real applications, but the policy trained by existing cMARL algorithms is not robust enough when deployed. There exist also many methods about adversarial attacks on the RL system, which implies that the RL system can suffer from adversarial attacks, but most of them focused on single agent RL. In this paper, we propose a sparse adversarial attack on cMARL systems. We use (MA)RL with regularization to train the attack policy. Our
more » ... periments show that the policy trained by the current cMARL algorithm can obtain poor performance when only one or a few agents in the team (e.g., 1 of 8 or 5 of 25) were attacked at a few timesteps (e.g., attack 3 of total 40 timesteps).
arXiv:2205.09362v2 fatcat:n2jjhat2mzcdtc7t7wfrchj4fq