What is it about?

Recent advancements in multi-agent reinforcement learning (MARL) have opened up vast application prospects, such as swarm control of drones, collaborative manipulation by robotic arms, and multitarget encirclement. However, potential security threats during the MARL deployment need more attention and thorough investigation.

Featured Image

Why is it important?

Recent research reveals that attackers can rapidly exploit the victim’s vulnerabilities, generating adversarial policies that result in the failure of specific tasks. For instance, reducing the winning rate of a superhuman-level Go AI to around 20%. Existing studies predominantly focus on two-player competitive environments, assuming attackers possess complete global state observation.

Perspectives

In this study, we unveil, for the first time, the capability of attackers to generate adversarial policies even when restricted to partial observations of the victims in multi-agent competitive environments. Specifically, we propose a novel black-box attack (SUB-PLAY) that incorporates the concept of constructing multiple subgames to mitigate the impact of partial observability and suggests sharing transitions among subpolicies to improve attackers’ exploitative ability. Extensive evaluations demonstrate the effectiveness of SUBPLAY under three typical partial observability limitations. Visualization results indicate that adversarial policies induce significantly different activations of the victims’ policy networks. Furthermore, we evaluate three potential defenses aimed at exploring ways to mitigate security threats posed by adversarial policies, providing constructive recommendations for deploying MARL in competitive environments.

Oubo Ma
Zhejiang University

Read the Original

This page is a summary of: SUB-PLAY: Adversarial Policies against Partially Observed Multi-Agent Reinforcement Learning Systems, December 2024, ACM (Association for Computing Machinery),
DOI: 10.1145/3658644.3670293.
You can read the full text:

Read

Resources

Contributors

The following have contributed to this page