- The paper presents a comprehensive survey of LLM-driven AI agent communication protocols, highlighting security vulnerabilities and recommending defense strategies.
- It analyzes diverse protocols across client-server, peer-to-peer, and hybrid architectures, detailing functional and security trade-offs in each model.
- It emphasizes actionable countermeasures including registration verification, load balancing, and multi-source isolation to mitigate identified risks.
A Survey of LLM-Driven AI Agent Communication: Protocols, Security Risks, and Defense Countermeasures (2506.19676)
Introduction
The paper presents a comprehensive survey on LLM-driven AI agent communication, focusing on the protocols that facilitate these interactions, the security risks involved, and potential defense strategies. As AI agents evolve to perform complex tasks through collaboration with other agents and tools, understanding the protocol designs and security challenges becomes pivotal for safe and efficient deployment in real-world scenarios.
Protocols for Agent Communication
Agent communication protocols are classified into several types based on architecture: client-server (CS), peer-to-peer (P2P), hybrid, and unspecified architectures. Each architecture has distinct characteristics influencing discovery process, flexibility, interoperability, and security posture.
CS-Based Communication Protocols
- ACP-IBM enables client-server interaction with multi-turn state preservation and robust agent discovery mechanisms.
- ACP-AGNTCY provides thread management and stateful operation, allowing scalable communication within centralized architectures.
P2P-Based Communication Protocols
- ACP-AgentUnion and ACN facilitate decentralized communication, enhancing scalability and end-to-end security.
- Agora dynamically switches modes to optimize task handling, enabling fluid experimentation across different agent networks.
Hybrid and Others
- LMOS offers flexible discovery for both centralized and decentralized setups, expanding potential deployment scenarios.
- A2A allows asynchronous priority scheduling, adapting discovery and execution schemes to varying contexts.
Security Risks in Agent Communication
Security risks in agent communication vary according to the architecture and include both specific and universal threats.
CS-Based Architecture Risks
- Registration Pollution: Overloads or hijacks agent registration leading to scheduling issues.
- Description Poisoning: Alters agent behavior via misleading capability descriptions.
- Task Flooding: Exhausts system resources through excessive task submissions.
- SEO Poisoning: Manipulates agent ranking via keyword stuffing, impacting discovery.
P2P-Based Architecture Risks
- Non-convergence: Task loops or stalling due to decentralized orchestration.
- Man-in-the-middle (MITM): Message interception, exploiting weak encryption.
Universal Risks
- Agent Spoofing: Impersonates trusted agents for injection or info-theft.
- Agent Exploitation/Trojan: Uses compromised/malicious agents as attack vectors.
- Agent Bullying: Disrupts logic via negative feedback loops.
- Privacy Leakage: Uncontrolled information spread between agents.
- Responsibility Evasion: Difficult fault attribution in multi-agent interactions.
- Denial of Service (DoS): Task overloads induce resource exhaustion.
Defense Countermeasure Prospect
To mitigate these risks, several strategies are proposed:
CS-Based Defense Strategies
- Registration Verification: Enforce authentication and monitor dynamic behavior.
- Capability Verification: Use benchmarks to confirm agent capabilities.
- Load Balancing: Implement rate limiting and intelligent scheduling.
- Anti-Manipulation Optimization: Apply adversarial training and randomization to search algorithms.
P2P-Based Defense Strategies
- Lifecycle Monitoring: Coordination for early termination of non-convergent tasks.
- Encryption Enhancements: Routine updates for security patches against MITM.
Universal Defense Strategies
- Identity Authentication: MFA and blockchain DIDs for robust verification.
- Agent Behavior Auditing: Log interaction history, detect procedural anomalies.
- Access Control: Fine-grained permissions ensure safe data access.
- Multi-Source Isolation: Prevent cross-agent contamination.
- Attack Modeling and Testing: Use adversarial frameworks to stress-test agent systems.
- Agent Orchestration: Optimize task scheduling and resource allocation dynamically.
Conclusion
The survey highlights the critical need for comprehensive security frameworks to protect agent-agent and agent-environment communication from both established and nascent threats. As AI systems grow more interconnected, maintaining trust, security, and reliability becomes increasingly challenging, necessitating adaptable defense strategies and rigorous protocol evaluation.