Preprint
Article

This version is not peer-reviewed.

Securing Agentic AI: A Comprehensive Threat Analysis of Model Context Protocol Systems with Layered Defense Strategies

Submitted:

23 October 2025

Posted:

28 October 2025

You are already at the latest version

Abstract
The rise of agentic artificial intelligence, powered by the Model Context Protocol (MCP), creates a complex and evolving security landscape. This article presents a comprehensive analysis of the MCP threat ecosystem through a structured five-layer architectural model: Prompt \& Reasoning, Tool \& Supply Chain, Execution \& Configuration, Protocol \& Network, and Data \& Telemetry. We examine over twenty distinct attack vectors ranging from well-documented prompt injection to novel threats like composability chaining and MCP sampling manipulation, and their manifestations across each architectural layer. Drawing from recent security research by Microsoft, CyberArk, Palo Alto Networks, and others, we analyze real-world vulnerabilities including CVE-2025-32711 (EchoLeak) and CVE-2025-6514 to demonstrate how traditional attack vectors evolve in agentic environments. We present actionable defense strategies, highlighting how modern frameworks like Strands Agents provide essential security controls through identity management, comprehensive observability, and zero-trust tool execution architectures for building secure, observable, and resilient agentic systems.
Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  
Artificial intelligence is evolving from a predictive tool into an autonomous workforce. These new "agentic" systems, powered by large language models (LLMs) and equipped with tool-calling capabilities through the Model Context Protocol (MCP), can independently use software tools, query databases, and interact with APIs to accomplish complex goals. MCP, functioning as a "USB-C port for AI applications," provides a universal interface between an AI’s reasoning core and the digital world [1].
While a simple LLM can generate text, an agent can act. This distinction is the crux of the security challenge. An agent with access to email tools, databases, and payment APIs exposes a vastly greater attack surface than the LLM alone. Recent critical vulnerabilities, including the CVE-2025-32711 "EchoLeak" zero-click vulnerability in Microsoft 365 Copilot and the CVE-2025-6514 remote code execution flaw in mcp-remote affecting over 437,000 installations, demonstrate the urgent reality of these threats [2,3].
This article provides a structured analysis of the MCP threat landscape, moving beyond LLM-centric views to focus on the unique risks of agentic tool-using systems. We map known and emerging MCP threats to a five-layer model, treating vectors like prompt injection not as standalone threats, but as mechanisms that cause agents to misuse their capabilities, leading to concrete risks like unauthorized tool execution and data exfiltration.

1. The Agentic Security Paradigm Shift

The security implications of agentic AI extend far beyond those of traditional language models. While a conventional LLM’s failure modes are typically limited to generating inappropriate text, an agentic system’s compromised behavior can result in data exfiltration, system compromise, financial fraud, and operational disruption across interconnected enterprise systems.
Consider the attack surface evolution: a traditional chatbot processes user input and generates responses within a controlled sandbox. In contrast, an agentic AI system may access email systems, cloud storage, databases, payment processors, and development environments. The compromise of such a system doesn’t merely risk inappropriate text generation, it potentially exposes the entire digital infrastructure the agent can reach.
This paradigm shift necessitates a fundamental reconceptualization of AI security, moving from content filtering and output sanitization to comprehensive workflow security, tool access control, and multi-layer defense strategies. The stakes encompass the full spectrum of enterprise cybersecurity risks, as demonstrated by recent high-profile incidents that validated theoretical concerns about agentic AI security.

2. Mapping Threats to the Five-Layer Model

The diverse MCP threat landscape can be organized into our five-layer architectural model, providing a clear framework for understanding how different attack vectors compromise specific parts of an agentic system. Through extensive analysis of current security research, we have identified over twenty attack vectors that map to these layers.
This structured approach recognizes that sophisticated attacks often span multiple layers, with initial compromise at one layer enabling escalation and persistence across others. Each layer presents distinct vulnerability classes requiring specialized defensive measures.

3. Layer 1: Prompt & Reasoning Integrity

This foundational layer concerns the integrity of the agent’s core logic and decision-making process. Attacks at this layer target the semantic understanding of the LLM itself, with the primary risk materializing not as improper text generation, but as the execution of unintended, often malicious, actions through the agent’s available tools.

3.1. Attack Vectors

The following attack vectors target the reasoning layer:
  • Direct Prompt Injection - Malicious prompts directly alter agent behavior through explicit instructions
  • Indirect Prompt Injection - Hidden instructions embedded in external data sources that the agent processes
  • Tool/Service Misuse ("Confused AI") - Agent uses correct tools in wrong contexts due to semantic manipulation
  • MCP Sampling Manipulation - Exploiting MCP’s sampling feature to extract data through crafted completion requests

3.2. Attack Vector Analysis

1. Indirect Prompt Injection: The EchoLeak Evolution
The CVE-2025-32711 "EchoLeak" vulnerability demonstrates the sophistication of modern indirect prompt injection. Unlike traditional prompt injection that requires direct user interaction, EchoLeak exploits how agents process external documents, emails, and web content. The attack embeds instructions in multiple vectors simultaneously: HTML metadata, comments, hidden divs, and even Unicode characters that are invisible to humans but processed by LLMs [4].
Preprints 182022 i001
The attack becomes particularly dangerous when combined with social engineering. Attackers send seemingly legitimate business documents that contain hidden payloads, exploiting the trust relationship between the document source and the organization. When the agent processes these documents for routine tasks like summarization or data extraction, it executes the embedded commands.
2. Tool/Service Misuse Through Semantic Manipulation
The "Confused AI" attack goes beyond simple naming confusion. Sophisticated attackers manipulate the semantic understanding of tools by crafting descriptions that exploit the LLM’s tendency to follow contextual cues. This attack leverages the agent’s inability to distinguish between legitimate operational guidance and malicious misdirection.
Preprints 182022 i002
3. MCP Sampling Exploitation
MCP’s sampling capability represents a powerful feature that enables servers to request LLM completions through clients, essentially allowing tools to "think" using the agent’s reasoning capabilities. However, this creates a bidirectional flow of control that attackers can exploit [5].
The attack works through a multi-stage process. First, the attacker creates a malicious MCP server that appears to offer legitimate functionality. When connected, the server sends sampling requests that seem benign but contain carefully crafted prompts designed to extract sensitive information. The sophistication lies in how these prompts bypass user scrutiny:
Preprints 182022 i003a
Preprints 182022 i003b

3.3. Defensive Strategies

Protecting reasoning integrity requires a multi-faceted approach that addresses both technical and human factors:
  • Provenance-Aware Processing: Implement strict source tagging where all external data is wrapped in explicit delimiters like `[EXTERNAL_DATA_START]...[EXTERNAL_DATA_END]` that the LLM is trained to treat as non-executable content
  • Semantic Firewall: Deploy a secondary LLM specifically trained to detect prompt injection patterns, analyzing both the content and metadata of inputs before they reach the primary agent
  • Tool Call Validation: Implement a policy engine that validates tool calls against a formal specification, checking that parameters don’t contain values derived directly from external sources without sanitization
  • Sampling Request Analysis: For MCP sampling, implement automated analysis of sampling requests to detect patterns indicative of data extraction, with special scrutiny for requests mentioning environment variables, credentials, or other sensitive information
  • Context Isolation: Maintain separate contexts for different trust levels, ensuring that data from untrusted sources cannot influence tool selection or parameter construction for sensitive operations

4. Layer 2: Tool & Supply Chain Security

This layer addresses the integrity of the agent’s capabilities—the tools themselves. With over 15,000 publicly available MCP tools and approximately 200 new tools published weekly, the ecosystem’s rapid growth creates a massive attack surface. The sophistication of attacks at this layer has evolved from simple malicious tools to complex supply chain compromises that can affect thousands of deployments simultaneously [6].

4.1. Attack Vectors

The tool ecosystem presents five primary attack categories, distinguished by how and where the deception occurs:
  • Tool Poisoning - Malicious instructions are embedded in a tool’s description or metadata. This attack doesn’t alter the tool’s code but instead tricks the AI’s reasoning, causing it to misuse an otherwise legitimate tool for a malicious purpose.
  • Tool Shadowing - An attacker creates a fake tool that impersonates a legitimate one by using a similar name and functionality. This "shadow" tool intercepts calls intended for the trusted tool, allowing the attacker to steal data or manipulate operations.
  • Package Squatting - This supply chain attack uses typographical errors or similar names to a legitimate package (e.g., reqeusts instead of requests). Systems or developers may mistakenly install the malicious version, compromising the environment.
  • Rug Pull Attacks - This attack involves a betrayal of trust over time. A developer publishes a safe, useful tool, waits for it to gain widespread adoption, and then pushes a malicious update. Users who have come to trust the tool are then compromised.
  • Composability Chaining - In this sophisticated vector, the attack is hidden in the interaction between multiple tools. Each tool in the chain may appear benign individually, but when they execute in sequence, their combined actions result in a malicious outcome.

4.2. Attack Vector Analysis

1. Advanced Tool Poisoning with Behavioral Manipulation
Modern tool poisoning has evolved into sophisticated psychological manipulation that exploits both LLM reasoning patterns and human cognitive biases. Attackers craft tool descriptions that appear to enhance security or compliance while actually creating vulnerabilities:
Preprints 182022 i004
2. Composability Chaining
Composability chaining represents the most sophisticated attack vector in the tool ecosystem. It exploits MCP’s powerful server-to-server communication feature to create multi-hop attack chains that are extremely difficult to detect and attribute:
Preprints 182022 i005a
Preprints 182022 i005b
This attack is particularly dangerous because each server in the chain can pass individual security audits. Malicious behavior only activates under specific conditions. The attack can remain dormant for extended periods before activation
3. Supply Chain Rug Pull with Behavioral Analysis Evasion
Rug pull attacks have evolved from simple malicious updates to sophisticated operations that actively evade detection:
Preprints 182022 i006

4.3. Defensive Strategies

Securing the tool ecosystem requires comprehensive approaches addressing multiple aspects of supply chain security:
  • Cryptographic Tool Verification - Implement a PKI system where all tools must be signed by verified publishers, with certificate transparency logs for audit trails
  • Behavioral Runtime Analysis - Deploy machine learning models trained on normal tool behavior patterns to detect anomalies in real-time execution
  • Tool Dependency Mapping - Maintain a complete graph of tool dependencies and communication patterns, alerting on unexpected connections or data flows
  • Gradual Trust Building - New tools operate in restricted sandboxes with limited capabilities until they build trust through consistent benign behavior over time, post which they are subject to behavioral runtime analysis.
  • Community Threat Intelligence - Establish industry-wide threat intelligence sharing for tool vulnerabilities, with automated blocklist distribution
  • Immutable Tool Execution - Tools run in read-only containers with no ability to modify their own code or download additional components
  • Supply Chain Transparency - Require tools to provide Software Bill of Materials (SBOM) with all dependencies clearly documented and verified

5. LAYER 3: EXECUTION & CONFIGURATION SECURITY

This layer concerns the runtime environment where tools execute and how system configurations evolve over time. Even with secure tools, vulnerabilities arise from the execution context, configuration management, and the dynamic nature of agentic systems that can modify their own operational parameters.

5.1. Attack Vectors

Execution environment vulnerabilities include:
  • Command Injection - Unvalidated input executed in system shells
  • Sandbox Escape - Tools break isolation boundaries to access unauthorized resources
  • Configuration Drift - Gradual security policy degradation through incremental changes
  • Schema Confusion - Exploiting mismatches between expected and actual data formats
  • Resource Exhaustion - Denial of service through computational or memory overflow

5.2. Attack Vector Analysis

1. Command Injection in MCP Tools
Research by Equixly demonstrated how MCP tools that shell out directly to system commands are vulnerable to injection attacks [8].
Preprints 182022 i007
2. Advanced Sandbox Escape Techniques
Modern sandbox escape attacks in MCP environments exploit the complex interaction between the agent, tools, and the underlying execution environment. Unlike traditional sandbox escapes that rely on memory corruption or kernel exploits, MCP-specific escapes abuse the semantic layer and tool interoperability:
Preprints 182022 i008
3. Configuration Drift Through Agent Self-Modification
Configuration drift in agentic systems is particularly dangerous because agents can modify their own operational parameters. Attackers exploit this through gradual changes that individually appear benign but collectively compromise security:
Preprints 182022 i009a
Preprints 182022 i009b
4. Schema Confusion Attacks
Schema confusion exploits the mismatch between what tools expect and what they receive, particularly dangerous in dynamically typed environments where MCP operates:
Preprints 182022 i010a
Preprints 182022 i010b

5.3. Defensive Strategies

Protecting execution environments requires defense-in-depth approaches:
  • Parameterized Execution - Never construct commands through string concatenation; use parameter arrays and prepared statements exclusively
  • Capability-Based Security - Tools declare required capabilities at registration; runtime enforces these boundaries with mandatory access controls
  • Configuration Immutability - Critical security configurations stored in append-only logs with cryptographic verification; changes require multi-party authorization
  • Schema Enforcement Gateways - Type-check and validate all data at tool boundaries using strict schema validation with no type coercion
  • Execution Provenance Tracking - Complete audit trail of all tool executions including parameters, environment state, and results with tamper-proof logging
  • Drift Detection Systems - Machine learning models trained on normal configuration patterns detect anomalous changes and alert security teams
  • Resource Quotas - Hard limits on CPU, memory, disk, and network usage per tool with automatic termination on violation

6. Layer 4: Protocol & Network Security

This layer addresses the security of communication channels between MCP components, including client-server interactions, inter-tool communications, and agent-to-service connections. The distributed nature of MCP systems creates numerous attack vectors at the protocol level.

6.1. Attack Vectors

Protocol-layer vulnerabilities include:
  • MCP Rebinding - DNS rebinding attacks redirecting local MCP servers to attacker control
  • Man-in-the-Middle - Intercepting and modifying MCP protocol messages
  • Protocol Downgrade - Forcing connections to use weaker security protocols
  • Certificate Spoofing - Impersonating legitimate MCP servers through certificate manipulation
  • WebSocket Hijacking - Taking control of persistent WebSocket connections

6.2. Attack Vector Analysis

1. Advanced MCP Rebinding Attacks
MCP rebinding attacks have evolved beyond simple DNS rebinding to exploit the complex trust relationships in distributed MCP deployments:
Preprints 182022 i011

6.3. Defensive Strategies

Protocol security requires comprehensive measures at multiple levels:
  • Mutual TLS (mTLS) Enforcement - Require certificate-based authentication for all MCP connections with regular certificate rotation
  • DNS Security Extensions (DNSSEC) - Implement DNSSEC to prevent DNS spoofing and cache poisoning attacks
  • WebSocket Security Headers - Enforce strict Origin validation, implement frame masking, and require secure WebSocket (wss://) connections
  • Protocol Version Pinning - Prevent downgrade attacks by enforcing minimum protocol versions and rejecting legacy handshakes
  • Connection State Validation - Implement nonce-based anti-replay mechanisms and validate connection state at protocol level
  • Network Segmentation - Isolate MCP servers in dedicated network segments with strict firewall rules and IDS/IPS monitoring
  • Rate Limiting and Throttling - Implement aggressive rate limiting to prevent scanning and brute force attacks

7. Layer 5: Data & Telemetry Security

This layer addresses the protection of sensitive information in operational data, including logs, metrics, traces, and user interactions. The extensive telemetry generated by agentic systems creates unique privacy and security challenges.

7.1. Attack Vectors

Data layer vulnerabilities include:
  • Inference Attacks - Extracting sensitive information from aggregated telemetry
  • Consent Fatigue Exploitation - Overwhelming users with approval requests to hide malicious actions
  • Token Leakage - Credentials exposed in logs, errors, or telemetry
  • Privacy Violation - Inadequate PII protection in operational data
  • Telemetry Poisoning - Injecting false data to corrupt monitoring and decision-making

7.2. Attack Vector Analysis

1. Advanced Inference Attacks on Telemetry
Modern telemetry systems collect vast amounts of operational data that, when aggregated, can reveal sensitive information not apparent in individual data points:
Preprints 182022 i012
2. Consent Fatigue Attack with Psychological Manipulation
This sophisticated attack exploits human psychology and UI/UX patterns to manipulate users into approving malicious actions:
Preprints 182022 i013a
Preprints 182022 i013b

7.3. Defensive Strategies

Protecting data and telemetry requires comprehensive privacy-preserving approaches:
  • Differential Privacy - Add calibrated noise to telemetry data to prevent inference attacks while maintaining statistical utility
  • Automated PII Detection - Deploy ML models to identify and redact PII in logs and telemetry before storage
  • Consent UX Improvements - Implement progressive disclosure, risk scoring, and visual differentiation for different request types
  • Token Rotation and Encryption - Automatic rotation of credentials with encryption at rest and in transit
  • Telemetry Minimization - Collect only essential operational data with automatic expiration and deletion policies
  • Anomaly Detection - ML-based detection of unusual patterns in consent requests and telemetry data
  • Audit Trail Integrity - Cryptographic signing of audit logs with tamper-evident storage

8. Strands Agents: Implementing Layered Security

The Strands Agents SDK, developed by AWS as an open-source framework, demonstrates how modern agentic platforms can implement comprehensive security across all five layers while maintaining operational flexibility [12].
The Strands Agents SDK provides a production-ready framework for building secure agentic systems. It implements security through a combination of native features and integrations, focusing on observability, content safety, and developer-enforced best practices. The following demonstrates a technically accurate approach to building secure agents with Strands.

8.0.1. Step 1: Observability and Telemetry Configuration

Enterprise-grade security begins with comprehensive monitoring. Strands uses OpenTelemetry to provide deep insights into agent behavior, essential for forensic analysis and security monitoring.
Preprints 182022 i014
This setup enables distributed tracing, allowing security teams to reconstruct the exact sequence of an agent’s reasoning, tool calls, and API interactions for forensic analysis after security incidents.

8.0.2. Step 2: Guardrails and Content Safety

Strands integrates with Amazon Bedrock’s native guardrails system rather than providing its own implementation. Guardrails enforce content policies, deny harmful topics, and filter prompts and responses.
Preprints 182022 i015

8.0.3. Step 3: PII Protection Implementation

Strands SDK does not provide native PII redaction [16]. Instead, it recommends wrapping agent invocations with third-party libraries for flexibility and specialized functionality.
Preprints 182022 i016

8.0.4. Step 4: Tool Security and Responsible AI Principles

Strands emphasizes developer responsibility for tool security [13]. Tools must implement these security principles:
  • Least Privilege - Tools receive minimum necessary permissions
  • Input Validation - Rigorous validation prevents injection attacks
  • Clear Documentation - Purpose and behavior clearly documented
  • Error Handling - Graceful failures without exposing sensitive information
  • Audit Logging - Security-relevant operations logged for review
Preprints 182022 i017a
Preprints 182022 i017b

8.0.5. Step 5: Secure Agent Integration

The final step integrates all security components into a comprehensive secure agent. Authentication is inherited from the execution environment (AWS IAM roles, environment variables).
Preprints 182022 i018a
Preprints 182022 i018b

8.1. Enterprise Security Features Summary

This layered approach provides comprehensive security through:
  • Observability - Complete audit trails via OpenTelemetry integration
  • Content Safety - Amazon Bedrock guardrails for automated filtering
  • PII Protection - Third-party library integration for specialized detection
  • Tool Security - Developer-enforced best practices with validation and logging
  • Access Control - AWS IAM integration and least privilege principles
  • Input Validation - Comprehensive validation at multiple layers
  • Error Handling - Secure error management without information disclosure
This comprehensive approach addresses security threats by combining native model safety features, complete observability for monitoring and forensics, and strong emphasis on developer-enforced best practices for secure tool creation and agent deployment.

9. Future Challenges and Research Directions

The rapid evolution of agentic AI systems introduces emerging security challenges requiring continued research:

9.1. Multi-Agent Coordination Attacks

As organizations deploy multiple specialized agents that collaborate, new attack vectors emerge:
  • Byzantine Agents - Compromised agents that provide subtly incorrect information to influence collective decisions
  • Coordination Protocol Exploitation - Attacks on consensus mechanisms and distributed decision-making
  • Emergent Behavior Manipulation - Exploiting unexpected behaviors from agent interactions

9.2. Temporal and Persistent Threats

Long-running agentic systems face unique temporal security challenges:
  • Slow Poisoning - Gradual corruption of agent behavior over extended periods
  • Memory Manipulation - Attacks on agent memory and context retention systems
  • Behavioral Drift - Unintended changes in agent behavior through continuous learning

9.3. Cross-Modal Security

As agents process diverse data types, security must address:
  • Modality Confusion - Hiding attacks in one modality while appearing benign in others
  • Semantic Gaps - Exploiting differences in how agents interpret different data types
  • Multimodal Injection - Coordinated attacks across text, images, audio, and video

10. Conclusion

Recent vulnerabilities like EchoLeak and mcp-remote have transformed theoretical risks into practical realities with significant potential for organizational harm. The sophistication of attacks from semantic sandbox escapes to multi-stage consent fatigue campaigns highlights the evolution of cybersecurity threats in the age of autonomous AI. Traditional security models prove insufficient for systems where the boundary between data and instructions becomes fluid and where semantic interpretation enables novel attack vectors.
However, frameworks like Strands Agents demonstrate that these challenges can be addressed through thoughtful security architecture. By implementing defense-in-depth strategies across all five layers from reasoning integrity to telemetry protection, organizations can deploy agentic AI systems that are both powerful and secure. Organizations deploying MCP-based systems must recognize that security cannot be an afterthought. The autonomous nature of agentic systems, combined with their extensive tool access and semantic reasoning capabilities, requires security-by-design approaches integrated into every aspect of system architecture and operation. Only through such comprehensive security measures can we realize the transformative potential of agentic AI while maintaining the trust and safety that modern enterprise environments demand.

References

  1. Model Context Protocol, “What is the Model Context Protocol (MCP)?” modelcontextprotocol.io, 2024. [Online]. Available: https://modelcontextprotocol.io/docs/getting-started/intro.
  2. SOC Prime, “CVE-2025-32711 Vulnerability: ’EchoLeak’ Flaw in Microsoft 365 Copilot Could Enable a Zero-Click Attack on an AI Agent,” Jun. 2025. [Online]. Available: https://socprime.com/blog/cve-2025-32711-zero-click-ai-vulnerability/.
  3. JFrog Security Research, “Critical RCE Vulnerability in mcp-remote: CVE-2025-6514 Threatens LLM Clients,” Jul. 2025. [Online]. Available: https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/.
  4. HackTheBox, “Inside CVE-2025-32711 (EchoLeak): Prompt injection meets AI exfiltration,” Jun. 2025. [Online]. Available: https://www.hackthebox.com/blog/cve-2025-32711-echoleak-copilot-vulnerability.
  5. CyberArk Labs, “Is your AI safe? Threat analysis of MCP (Model Context Protocol),” 2025. [Online]. Available: https://www.cyberark.com/resources/threat-research-blog/is-your-ai-safe-threat-analysis-of-mcp-model-context-protocol.
  6. GitHub Security Lab, “Analysis of the MCP Tool Ecosystem: Security Challenges and Recommendations,” Aug. 2025. [Online]. Available: https://github.blog/security/vulnerability-research/mcp-ecosystem-security-analysis/.
  7. Invariant Labs, “MCP Security Notification: Tool Poisoning Attacks,” Apr. 2025. [Online]. Available: https://invariantlabs.ai/blog/mcp-security-notification-tool-poisoning-attacks.
  8. Equixly, “MCP Server: New Security Nightmare,” Mar. 2025. [Online]. Available: https://equixly.com/blog/2025/03/29/mcp-server-new-security-nightmare/.
  9. GitHub Security Lab, “DNS rebinding attacks explained,” Jun. 2025. [Online]. Available: https://github.blog/security/application-security/dns-rebinding-attacks-explained-the-lookup-is-coming-from-inside-the-house/.
  10. Varonis, “Understanding DNS rebinding threats to MCP servers,” Aug. 2025. [Online]. Available: https://www.varonis.com/blog/model-context-protocol-dns-rebind-attack.
  11. Palo Alto Networks, “MCP Security Exposed: What You Need to Know Now,” May 2025. [Online]. Available: https://live.paloaltonetworks.com/t5/community-blogs/mcp-security-exposed-what-you-need-to-know-now/ba-p/1227143.
  12. AWS Open Source Blog, “Introducing Strands Agents, an open source AI agents SDK,” May 2025. [Online]. Available: https://aws.amazon.com/blogs/opensource/introducing-strands-agents-an-open-source-ai-agents-sdk/.
  13. Strands Agents SDK, “Responsible AI,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/responsible-ai/.
  14. Strands Agents SDK, “Guardrails,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/guardrails/.
  15. Strands Agents SDK, “Observability,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/observability-evaluation/observability/.
  16. Strands Agents SDK, “PII Redaction,” 2025. [Online]. Available: https://strandsagents.com/latest/documentation/docs/user-guide/safety-security/pii-redaction/.
  17. AWS Machine Learning Blog, “Strands Agents SDK: A technical deep dive into agent architectures and observability,” 2025. [Online]. Available: https://aws.amazon.com/blogs/machine-learning/strands-agents-sdk-a-technical-deep-dive-into-agent-architectures-and-observability/.
  18. AWS Blog, “Introducing Amazon Bedrock AgentCore: Securely deploy and operate AI agents at any scale,” Jul. 2025. [Online]. Available: https://aws.amazon.com/blogs/aws/introducing-amazon-bedrock-agentcore-securely-deploy-and-operate-ai-agents-at-any-scale/.
  19. Microsoft Security Blog, “Understanding and mitigating security risks in MCP implementations,” Apr. 2025. [Online]. Available: https://techcommunity.microsoft.com/blog/microsoft-security-blog/understanding-and-mitigating-security-risks-in-mcp-implementations/4404667.
  20. Model Context Protocol, “Security Best Practices,” draft specification, 2025. [Online]. Available: https://modelcontextprotocol.io/specification/draft/basic/security_best_practices.
  21. OWASP Foundation, “OWASP Top 10 for Large Language Model Applications,” version 2.0, 2025. [Online]. Available: https://owasp.org/www-project-top-10-for-large-language-model-applications/.
  22. National Institute of Standards and Technology, “AI Risk Management Framework (AI RMF 2.0),” Jan. 2025. [Online]. Available: https://www.nist.gov/itl/ai-risk-management-framework.
  23. International Organization for Standardization, “ISO/IEC 27561:2025 - Information Security for Artificial Intelligence Systems,” Geneva, Switzerland: ISO, 2025.
  24. V. Pendyala, R. Raja, A. Vats, R. Para, D. Krishnamoorthy, U. Kumar, S. R. Narra, S. Bharadwaj, D. Nagasubramanian, P. Roy, D. Roy, D. Pant, and S. Lohani, “The Cognitive Nexus, Vol. 1, Issue 2: Advances in AI Methodology, Infrastructure, and Governance,” IEEE Computational Intelligence Society, Santa Clara Valley Chapter, Oct. 2025. Magazine issue editorial. Available at: https://www.researchgate.net/publication/396179773_The_Cognitive_Nexus_Vol_1_Issue_2_Advances_in_AI_Methodology_Infrastructure_and_Governance.
  25. V. Pendyala, R. Raja, A. Vats, N. Krishnan, L. Yerra, A. Kar, N. Kalu-Mba, M. Venkatram, and S. R. Bolla, “The Cognitive Nexus, Vol. 1, Issue 1: Computational Intelligence for Collaboration, Vision–Language Reasoning, and Resilient Infrastructures,” IEEE Computational Intelligence Society, Santa Clara Valley Chapter, July 2025. Magazine issue editorial. Available at: https://www.researchgate.net/publication/396179779_The_Cognitive_Nexus_Vol_1_Issue_1_Computational_Intelligence_for_Collaboration_Vision-Language_Reasoning_and_Resilient_Infrastructures.
  26. V. Pendyala, R. Raja, A. Vats, R. Para, D. Krishnamoorthy, U. Kumar, S. R. Narra, S. Bharadwaj, D. Nagasubramanian, P. Roy, D. Roy, D. Pant, and S. Lohani, “The Cognitive Nexus, Vol. 1, Issue 2: Advances in AI Methodology, Infrastructure, and Governance,” Preprints, Oct. 2025. [CrossRef]

Short Biography of Authors

Dewank Pant is an AI Security Engineer and Senior Member of IEEE specializing in adversarial testing, AI supply chain security, and secure agent architecture design. He received an M.S. in Security Informatics from Johns Hopkins University and has extensive experience in vulnerability research and red team operations. His current research focuses on the security implications of agentic AI systems, Model Context Protocol implementations, and the development of defensive frameworks for autonomous AI deployments. He is also a contributor to the OWASP LLM Top 10 project. His work and CVEs have been cited in major security advisories and has contributed to industry best practices for secure AI deployment. Contact him at dewankpant@gmail.com.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated