background pattern

Quantifying and Mitigating Emerging Risks in Multi-Agent Collaboration

This project investigates critical safety challenges in large-scale deployments of AI agents, focusing on privacy leakage and collusion risks in multi-agent environments. As agents collaborate and negotiate across complex tasks, they may unintentionally expose sensitive information or coordinate in ways that misalign with human values. The research develops a simulation testbed to analyse these behaviours, introduces dynamic privacy protocols, and explores how scaling agent interactions amplifies risk. Outcomes include a taxonomy of collusion patterns, mitigation strategies, and design principles for safer, transparent, and trustworthy multi-agent systems—informing future AI safety standards and governance.

This research is conducted via The Agentic AI Research and Innovation (AARI) Initiative which focuses on the next frontier of agentic systems through Grand Challenges with the academic community and Microsoft Research.

People

Portrait of Jianxun Lian

Jianxun Lian

Principal Researcher

Portrait of Beibei Shi

Beibei Shi

Principal Research PM

Portrait of Yule Wen

Yule Wen

Undergraduate

Tsinghua University.

Portrait of Xing Xie

Xing Xie

Assistant Managing Director

Portrait of Diyi  Yang

Diyi Yang

Assistant Professor

Stanford University

Portrait of Xiaoyuan Yi

Xiaoyuan Yi

Researcher

Portrait of Yanzhe Zhang

Yanzhe Zhang

PhD Student

Stanford University