Submitted:
23 October 2025
Posted:
28 October 2025
You are already at the latest version
Abstract
Keywords:
1. The Agentic Security Paradigm Shift
2. Mapping Threats to the Five-Layer Model
3. Layer 1: Prompt & Reasoning Integrity
3.1. Attack Vectors
- Direct Prompt Injection - Malicious prompts directly alter agent behavior through explicit instructions
- Indirect Prompt Injection - Hidden instructions embedded in external data sources that the agent processes
- Tool/Service Misuse ("Confused AI") - Agent uses correct tools in wrong contexts due to semantic manipulation
- MCP Sampling Manipulation - Exploiting MCP’s sampling feature to extract data through crafted completion requests
3.2. Attack Vector Analysis




3.3. Defensive Strategies
- Provenance-Aware Processing: Implement strict source tagging where all external data is wrapped in explicit delimiters like `[EXTERNAL_DATA_START]...[EXTERNAL_DATA_END]` that the LLM is trained to treat as non-executable content
- Semantic Firewall: Deploy a secondary LLM specifically trained to detect prompt injection patterns, analyzing both the content and metadata of inputs before they reach the primary agent
- Tool Call Validation: Implement a policy engine that validates tool calls against a formal specification, checking that parameters don’t contain values derived directly from external sources without sanitization
- Sampling Request Analysis: For MCP sampling, implement automated analysis of sampling requests to detect patterns indicative of data extraction, with special scrutiny for requests mentioning environment variables, credentials, or other sensitive information
- Context Isolation: Maintain separate contexts for different trust levels, ensuring that data from untrusted sources cannot influence tool selection or parameter construction for sensitive operations
4. Layer 2: Tool & Supply Chain Security
4.1. Attack Vectors
- Tool Poisoning - Malicious instructions are embedded in a tool’s description or metadata. This attack doesn’t alter the tool’s code but instead tricks the AI’s reasoning, causing it to misuse an otherwise legitimate tool for a malicious purpose.
- Tool Shadowing - An attacker creates a fake tool that impersonates a legitimate one by using a similar name and functionality. This "shadow" tool intercepts calls intended for the trusted tool, allowing the attacker to steal data or manipulate operations.
- Package Squatting - This supply chain attack uses typographical errors or similar names to a legitimate package (e.g., reqeusts instead of requests). Systems or developers may mistakenly install the malicious version, compromising the environment.
- Rug Pull Attacks - This attack involves a betrayal of trust over time. A developer publishes a safe, useful tool, waits for it to gain widespread adoption, and then pushes a malicious update. Users who have come to trust the tool are then compromised.
- Composability Chaining - In this sophisticated vector, the attack is hidden in the interaction between multiple tools. Each tool in the chain may appear benign individually, but when they execute in sequence, their combined actions result in a malicious outcome.
4.2. Attack Vector Analysis




4.3. Defensive Strategies
- Cryptographic Tool Verification - Implement a PKI system where all tools must be signed by verified publishers, with certificate transparency logs for audit trails
- Behavioral Runtime Analysis - Deploy machine learning models trained on normal tool behavior patterns to detect anomalies in real-time execution
- Tool Dependency Mapping - Maintain a complete graph of tool dependencies and communication patterns, alerting on unexpected connections or data flows
- Gradual Trust Building - New tools operate in restricted sandboxes with limited capabilities until they build trust through consistent benign behavior over time, post which they are subject to behavioral runtime analysis.
- Community Threat Intelligence - Establish industry-wide threat intelligence sharing for tool vulnerabilities, with automated blocklist distribution
- Immutable Tool Execution - Tools run in read-only containers with no ability to modify their own code or download additional components
- Supply Chain Transparency - Require tools to provide Software Bill of Materials (SBOM) with all dependencies clearly documented and verified
5. LAYER 3: EXECUTION & CONFIGURATION SECURITY
5.1. Attack Vectors
- Command Injection - Unvalidated input executed in system shells
- Sandbox Escape - Tools break isolation boundaries to access unauthorized resources
- Configuration Drift - Gradual security policy degradation through incremental changes
- Schema Confusion - Exploiting mismatches between expected and actual data formats
- Resource Exhaustion - Denial of service through computational or memory overflow
5.2. Attack Vector Analysis






5.3. Defensive Strategies
- Parameterized Execution - Never construct commands through string concatenation; use parameter arrays and prepared statements exclusively
- Capability-Based Security - Tools declare required capabilities at registration; runtime enforces these boundaries with mandatory access controls
- Configuration Immutability - Critical security configurations stored in append-only logs with cryptographic verification; changes require multi-party authorization
- Schema Enforcement Gateways - Type-check and validate all data at tool boundaries using strict schema validation with no type coercion
- Execution Provenance Tracking - Complete audit trail of all tool executions including parameters, environment state, and results with tamper-proof logging
- Drift Detection Systems - Machine learning models trained on normal configuration patterns detect anomalous changes and alert security teams
- Resource Quotas - Hard limits on CPU, memory, disk, and network usage per tool with automatic termination on violation
6. Layer 4: Protocol & Network Security
6.1. Attack Vectors
- MCP Rebinding - DNS rebinding attacks redirecting local MCP servers to attacker control
- Man-in-the-Middle - Intercepting and modifying MCP protocol messages
- Protocol Downgrade - Forcing connections to use weaker security protocols
- Certificate Spoofing - Impersonating legitimate MCP servers through certificate manipulation
- WebSocket Hijacking - Taking control of persistent WebSocket connections
6.2. Attack Vector Analysis

6.3. Defensive Strategies
- Mutual TLS (mTLS) Enforcement - Require certificate-based authentication for all MCP connections with regular certificate rotation
- DNS Security Extensions (DNSSEC) - Implement DNSSEC to prevent DNS spoofing and cache poisoning attacks
- WebSocket Security Headers - Enforce strict Origin validation, implement frame masking, and require secure WebSocket (wss://) connections
- Protocol Version Pinning - Prevent downgrade attacks by enforcing minimum protocol versions and rejecting legacy handshakes
- Connection State Validation - Implement nonce-based anti-replay mechanisms and validate connection state at protocol level
- Network Segmentation - Isolate MCP servers in dedicated network segments with strict firewall rules and IDS/IPS monitoring
- Rate Limiting and Throttling - Implement aggressive rate limiting to prevent scanning and brute force attacks
7. Layer 5: Data & Telemetry Security
7.1. Attack Vectors
- Inference Attacks - Extracting sensitive information from aggregated telemetry
- Consent Fatigue Exploitation - Overwhelming users with approval requests to hide malicious actions
- Token Leakage - Credentials exposed in logs, errors, or telemetry
- Privacy Violation - Inadequate PII protection in operational data
- Telemetry Poisoning - Injecting false data to corrupt monitoring and decision-making
7.2. Attack Vector Analysis



7.3. Defensive Strategies
- Differential Privacy - Add calibrated noise to telemetry data to prevent inference attacks while maintaining statistical utility
- Automated PII Detection - Deploy ML models to identify and redact PII in logs and telemetry before storage
- Consent UX Improvements - Implement progressive disclosure, risk scoring, and visual differentiation for different request types
- Token Rotation and Encryption - Automatic rotation of credentials with encryption at rest and in transit
- Telemetry Minimization - Collect only essential operational data with automatic expiration and deletion policies
- Anomaly Detection - ML-based detection of unusual patterns in consent requests and telemetry data
- Audit Trail Integrity - Cryptographic signing of audit logs with tamper-evident storage
8. Strands Agents: Implementing Layered Security
8.0.1. Step 1: Observability and Telemetry Configuration

8.0.2. Step 2: Guardrails and Content Safety

8.0.3. Step 3: PII Protection Implementation

8.0.4. Step 4: Tool Security and Responsible AI Principles
- Least Privilege - Tools receive minimum necessary permissions
- Input Validation - Rigorous validation prevents injection attacks
- Clear Documentation - Purpose and behavior clearly documented
- Error Handling - Graceful failures without exposing sensitive information
- Audit Logging - Security-relevant operations logged for review


8.0.5. Step 5: Secure Agent Integration


8.1. Enterprise Security Features Summary
- Observability - Complete audit trails via OpenTelemetry integration
- Content Safety - Amazon Bedrock guardrails for automated filtering
- PII Protection - Third-party library integration for specialized detection
- Tool Security - Developer-enforced best practices with validation and logging
- Access Control - AWS IAM integration and least privilege principles
- Input Validation - Comprehensive validation at multiple layers
- Error Handling - Secure error management without information disclosure
9. Future Challenges and Research Directions
9.1. Multi-Agent Coordination Attacks
- Byzantine Agents - Compromised agents that provide subtly incorrect information to influence collective decisions
- Coordination Protocol Exploitation - Attacks on consensus mechanisms and distributed decision-making
- Emergent Behavior Manipulation - Exploiting unexpected behaviors from agent interactions
9.2. Temporal and Persistent Threats
- Slow Poisoning - Gradual corruption of agent behavior over extended periods
- Memory Manipulation - Attacks on agent memory and context retention systems
- Behavioral Drift - Unintended changes in agent behavior through continuous learning
9.3. Cross-Modal Security
- Modality Confusion - Hiding attacks in one modality while appearing benign in others
- Semantic Gaps - Exploiting differences in how agents interpret different data types
- Multimodal Injection - Coordinated attacks across text, images, audio, and video
10. Conclusion
References
- Model Context Protocol, “What is the Model Context Protocol (MCP)?” modelcontextprotocol.io, 2024. [Online]. Available: https://modelcontextprotocol.io/docs/getting-started/intro.
- SOC Prime, “CVE-2025-32711 Vulnerability: ’EchoLeak’ Flaw in Microsoft 365 Copilot Could Enable a Zero-Click Attack on an AI Agent,” Jun. 2025. [Online]. Available: https://socprime.com/blog/cve-2025-32711-zero-click-ai-vulnerability/.
- JFrog Security Research, “Critical RCE Vulnerability in mcp-remote: CVE-2025-6514 Threatens LLM Clients,” Jul. 2025. [Online]. Available: https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/.
- HackTheBox, “Inside CVE-2025-32711 (EchoLeak): Prompt injection meets AI exfiltration,” Jun. 2025. [Online]. Available: https://www.hackthebox.com/blog/cve-2025-32711-echoleak-copilot-vulnerability.
- CyberArk Labs, “Is your AI safe? Threat analysis of MCP (Model Context Protocol),” 2025. [Online]. Available: https://www.cyberark.com/resources/threat-research-blog/is-your-ai-safe-threat-analysis-of-mcp-model-context-protocol.
- GitHub Security Lab, “Analysis of the MCP Tool Ecosystem: Security Challenges and Recommendations,” Aug. 2025. [Online]. Available: https://github.blog/security/vulnerability-research/mcp-ecosystem-security-analysis/.
- Invariant Labs, “MCP Security Notification: Tool Poisoning Attacks,” Apr. 2025. [Online]. Available: https://invariantlabs.ai/blog/mcp-security-notification-tool-poisoning-attacks.
- Equixly, “MCP Server: New Security Nightmare,” Mar. 2025. [Online]. Available: https://equixly.com/blog/2025/03/29/mcp-server-new-security-nightmare/.
- GitHub Security Lab, “DNS rebinding attacks explained,” Jun. 2025. [Online]. Available: https://github.blog/security/application-security/dns-rebinding-attacks-explained-the-lookup-is-coming-from-inside-the-house/.
- Varonis, “Understanding DNS rebinding threats to MCP servers,” Aug. 2025. [Online]. Available: https://www.varonis.com/blog/model-context-protocol-dns-rebind-attack.
- Palo Alto Networks, “MCP Security Exposed: What You Need to Know Now,” May 2025. [Online]. Available: https://live.paloaltonetworks.com/t5/community-blogs/mcp-security-exposed-what-you-need-to-know-now/ba-p/1227143.
- AWS Open Source Blog, “Introducing Strands Agents, an open source AI agents SDK,” May 2025. [Online]. Available: https://aws.amazon.com/blogs/opensource/introducing-strands-agents-an-open-source-ai-agents-sdk/.
- Strands Agents SDK, “Responsible AI,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/responsible-ai/.
- Strands Agents SDK, “Guardrails,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/guardrails/.
- Strands Agents SDK, “Observability,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/observability-evaluation/observability/.
- Strands Agents SDK, “PII Redaction,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/pii-redaction/.
- AWS Machine Learning Blog, “Strands Agents SDK: A technical deep dive into agent architectures and observability,” 2025. [Online]. Available: https://aws.amazon.com/blogs/machine-learning/strands-agents-sdk-a-technical-deep-dive-into-agent-architectures-and-observability/.
- AWS Blog, “Introducing Amazon Bedrock AgentCore: Securely deploy and operate AI agents at any scale,” Jul. 2025. [Online]. Available: https://aws.amazon.com/blogs/aws/introducing-amazon-bedrock-agentcore-securely-deploy-and-operate-ai-agents-at-any-scale/.
- Microsoft Security Blog, “Understanding and mitigating security risks in MCP implementations,” Apr. 2025. [Online]. Available: https://techcommunity.microsoft.com/blog/microsoft-security-blog/understanding-and-mitigating-security-risks-in-mcp-implementations/4404667.
- Model Context Protocol, “Security Best Practices,” draft specification, 2025. [Online]. Available: https://modelcontextprotocol.io/specification/draft/basic/security_best_practices.
- OWASP Foundation, “OWASP Top 10 for Large Language Model Applications,” version 2.0, 2025. [Online]. Available: https://owasp.org/www-project-top-10-for-large-language-model-applications/.
- National Institute of Standards and Technology, “AI Risk Management Framework (AI RMF 2.0),” Jan. 2025. [Online]. Available: https://www.nist.gov/itl/ai-risk-management-framework.
- International Organization for Standardization, “ISO/IEC 27561:2025 - Information Security for Artificial Intelligence Systems,” Geneva, Switzerland: ISO, 2025.
- V. Pendyala, R. Raja, A. Vats, R. Para, D. Krishnamoorthy, U. Kumar, S. R. Narra, S. Bharadwaj, D. Nagasubramanian, P. Roy, D. Roy, D. Pant, and S. Lohani, “The Cognitive Nexus, Vol. 1, Issue 2: Advances in AI Methodology, Infrastructure, and Governance,” IEEE Computational Intelligence Society, Santa Clara Valley Chapter, Oct. 2025. Magazine issue editorial. Available at: https://www.researchgate.net/publication/396179773_The_Cognitive_Nexus_Vol_1_Issue_2_Advances_in_AI_Methodology_Infrastructure_and_Governance.
- V. Pendyala, R. Raja, A. Vats, N. Krishnan, L. Yerra, A. Kar, N. Kalu-Mba, M. Venkatram, and S. R. Bolla, “The Cognitive Nexus, Vol. 1, Issue 1: Computational Intelligence for Collaboration, Vision–Language Reasoning, and Resilient Infrastructures,” IEEE Computational Intelligence Society, Santa Clara Valley Chapter, July 2025. Magazine issue editorial. Available at: https://www.researchgate.net/publication/396179779_The_Cognitive_Nexus_Vol_1_Issue_1_Computational_Intelligence_for_Collaboration_Vision-Language_Reasoning_and_Resilient_Infrastructures.
- V. Pendyala, R. Raja, A. Vats, R. Para, D. Krishnamoorthy, U. Kumar, S. R. Narra, S. Bharadwaj, D. Nagasubramanian, P. Roy, D. Roy, D. Pant, and S. Lohani, “The Cognitive Nexus, Vol. 1, Issue 2: Advances in AI Methodology, Infrastructure, and Governance,” Preprints, Oct. 2025. [CrossRef]
Short Biography of Authors
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
