Coordination Risks¶
When multiple AI agents interact, coordination can be beneficial (cooperation) or harmful (collusion). SWARM studies the boundary between the two — and provides governance mechanisms to keep coordination constructive. See Soft-Label Governance for Distributional Safety in Multi-Agent Systems for the formal framework; see also Distributional AGI Safety.
Why Coordination Becomes Risky¶
Individual agents acting independently produce risks that scale linearly. Coordinated agents produce risks that scale combinatorially. Three failure patterns dominate:
1. Collusion¶
Two or more agents coordinate to extract value at the expense of others. In SWARM, this appears as correlated exploitation patterns:
from swarm.governance import GovernanceConfig
config = GovernanceConfig(
collusion_detection=True,
collusion_threshold=0.8, # flag pairs with >80% correlation
collusion_window=20, # over 20 interactions
)
Detection signal: Unusually high correlation between agent pairs' exploitation timing.
2. Information Cascades¶
Agents copy each other's behavior rather than acting on private signals. When the first few agents make a mistake, the entire population follows:
| Phase | Behavior | Risk |
|---|---|---|
| Seed | 2-3 agents adopt strategy | Low |
| Cascade | Population copies without evaluation | Growing |
| Lock-in | Wrong strategy becomes consensus | High |
Detection signal: Sudden homogenization of agent strategies within 1-2 epochs.
3. Coordinated Exploitation¶
A group of agents systematically targets specific counterparties or exploits governance gaps that only work with multiple participants.
Detection signal: Subgroup of agents with consistently high payoffs while specific counterparties suffer.
Measuring Coordination Risk¶
SWARM provides metrics for coordination health:
from swarm.metrics.soft_metrics import SoftMetrics
metrics = SoftMetrics()
# Check for pairwise exploitation correlation
for pair in agent_pairs:
correlation = metrics.pairwise_correlation(interactions, pair)
if correlation > 0.8:
print(f"Potential collusion: {pair} (r={correlation:.3f})")
Governance Countermeasures¶
| Mechanism | What it addresses | Configuration |
|---|---|---|
| Collusion detection | Coordinated exploitation | collusion_threshold, collusion_window |
| Transaction tax | Reduces volume of coordinated interactions | transaction_tax |
| Random audits | Probabilistic detection of any pattern | audit_probability |
| Reputation decay | Prevents coordinated trust accumulation | reputation_decay |
The Cooperation-Collusion Boundary¶
Not all coordination is harmful. The challenge is distinguishing:
| Cooperation (beneficial) | Collusion (harmful) |
|---|---|
| Improves system welfare | Extracts from system welfare |
| Transparent signaling | Concealed coordination |
| Positive quality gap | Negative quality gap |
| Others can participate | Exclusive to in-group |
SWARM's quality gap metric helps distinguish these: when coordinated agents produce a negative quality gap, the system is selecting for harm.
See also¶
- Governance Mechanisms — Collusion detection and other countermeasures
- Deception — When coordination involves misrepresentation
- Governance Simulation — Test coordination scenarios before deployment
- Red-Teaming Guide — Adversarial coordination attack patterns