Agentic AI introduces security vulnerabilities that traditional LLM safeguards fail to address. Although recent work by Unit 42 at Palo Alto Networks demonstrated that ChatGPT-4o successfully executes attacks as an agent that it refuses in chat mode, there is no comparative analysis in multiple models and frameworks. We conducted the first systematic penetration testing and comparative evaluation of agentic AI systems, testing five prominent models (Claude 3.5 Sonnet, Gemini 2.5 Flash, GPT-4o, Grok 2, and Nova Pro) across two agentic AI frameworks (AutoGen and CrewAI) using a seven-agent architecture that mimics the functionality of a university information management system and 13 distinct attack scenarios that span prompt injection, Server Side Request Forgery (SSRF), SQL injection, and tool misuse. Our 130 total test cases reveal significant security disparities: AutoGen demonstrates a 52.3% refusal rate versus CrewAI's 30.8%, while model performance ranges from Nova Pro's 46.2% to Claude and Grok 2's 38.5%. Most critically, Grok 2 on CrewAI rejected only 2 of 13 attacks (15.4% refusal rate), and the overall refusal rate of 41.5% across all configurations indicates that more than half of malicious prompts succeeded despite enterprise-grade safety mechanisms. We identify six distinct defensive behavior patterns including a novel "hallucinated compliance" strategy where models fabricate outputs rather than executing or refusing attacks, and provide actionable recommendations for secure agent deployment. Complete attack prompts are also included in the Appendix to enable reproducibility.
The rapid adoption of agentic AI systems (autonomous agents capable of planning, using tools, and executing multistep tasks) represents a fundamental shift in how artificial intelligence is deployed in production environments. Unlike traditional large language models (LLMs) that operate in stateless conversational modes, agentic AI possesses the autonomy to invoke external tools, access databases, execute code, and make decisions across extended task sequences. This architectural evolution introduces unprecedented security challenges that conventional LLM safeguards do not adequately address [10], [14].
โข Quantitative framework comparison revealing Auto-Gen’s 52.3% refusal rate versus CrewAI’s 30.8%
โข Security scorecard ranking models from Nova Pro (46.2% refusal rate) to Claude/Grok 2 (38.5%)
โข Taxonomy of six distinct defensive behavior patterns, including a novel “hallucinated compliance” strategy
The remainder of this paper is organized as follows: Section 2 reviews the Unit 42 methodology and identifies research gaps. Section 3 details our experimental design, including model selection, framework architectures, and attack taxonomy. Section 4 presents results with a frameworklevel and model-level analysis. Section 5 concludes with key findings and implications for the deployment of secure agents, and Section 6 discusses future research directions.
This study builds on the methodology established by Unit 42 at Palo Alto Networks (2025) [1], which identified critical security vulnerabilities specific to agentic AI systems. He et al. characterize AI agents as “robots in cyberspace, executing tasks on behalf of their users” [14]. Palo Alto research demonstrated that traditional LLM safeguards become significantly less effective when models are deployed as agents with tool access and autonomy.
The original study tested ChatGPT-4o in both standard chat mode and as an autonomous agent in three primary attack vectors.
Prompt Injection Attacks: The research showed that direct prompt injections (where malicious instructions are embedded in user input) and indirect prompt injections (where instructions are hidden in external data sources like websites or documents) could manipulate agent behavior. Greshake et al. introduced the concept of indirect prompt injection, demonstrating that applications integrated with LLM blur the line between data and instructions [10]. Subsequent benchmarking revealed that more capable models are often more susceptible to these attacks [12], [13]. The agent successfully executed SQL injection attempts, SSRF attacks targeting internal networks, and unauthorized data exfiltration (attacks that the chat-only version consistently refused).
Tool Misuse and Privilege Escalation: When given access to tools such as code execution, file system operations, and web browsing, the agent could be manipulated to abuse these capabilities. Examples included reading sensitive configuration files, accessing private APIs, and executing arbitrary code that the base model would recognize as dangerous but the agent would perform when framed as a legitimate sub-task. Recent studies confirm that tool-integrated agents exhibit 24-47% attack success rates [19], with unconfined agents successfully executing 80% of malicious intents [14].
Multi-Step Attack Chains: The autonomous planning capability enabled sophisticated attack sequences. An agent could be directed to first gather system information, then identify vulnerabilities, and finally exploit them, with each step appearing benign in isolation but combined into a complete attack chain.
The Unit 42 study highlighted three fundamental challenges:
Palo Alto research tested both CrewAI and AutoGen frameworks with identical configurations (same instructions, language models, and tools) to isolate frameworkspecific versus systemic vulnerabilities. Comparative analysis of function calling architectures reveals significant security differences, with centralized orchestration showing 73.5% attack success rates versus 62.6% for distributed approaches [17]. Their findings revealed that most security risks are framework-agnostic, arising from three root causes:
โข Insecure Design Patterns: Poorly scoped prompts, missing input validation, and overly permissive tool access create attack surfaces regardless of the underlying framework. Even without explicit prompt injection, weakly defined agent instructions can be exploited.
โข Tool-Level Vulnerabilities: Traditional application security flaws (SQL injection, broken access control (BOLA), and server-side request forgery (SSRF)) remain exploitable when agents invoke vulnerable tools. The autonomous nature of agents simply provides a new attack vector to overcome these existing weaknesses.
โข Code Execution Risks: Unsecured code interpreters expose critical threats, including arbitrary code execution, file system access, and cloud metadata service exploitation. Default container configurations often lack network restrictions, capability dropping, or syscall filtering.
The study also documented nine specific attack scenarios with detailed exploitation techniques, demonstrating that attacks require defense-in-depth strategies that include prompt hardening, content filtering, tool input sanitization, vulnerability scanning, and code executor sandboxing. Critically, their testing showed that no single mitigation is sufficient; attackers can often bypass individual defenses through creative prompt variations.
Our research directly addresses these gaps by expanding the testing scope to five different AI models across two agent frameworks, providing the first comparative security analysis of agentic AI systems. Additionally, while Unit 42’s investment advisory assistant employed three cooperating agents (orchestrator, news, and stock), our student information system deploys seven specialized agents with distinct data access patterns, creating a more complex attack surface with additional inter-agent communication pathways.
Our research extends the original methodology by expanding the scope to include five different AI models, providing a comparative analysis of security vulnerabilities across multiple platforms. The experimental design consisted of:
We selected five prominent AI models representing different architectural approaches and safety implementations:
To isolate framework-specific security behaviors from model-level vulnerabilities, we implemented functionally identical systems using AutoGen [7] and CrewAI [8]. While both frameworks enable multi-agent collaboration, they differ fundamentally in orchestration patterns and communication mechanisms. Previous work on multi-agent security warns that “LLM-to-LLM prompt injection can spread like a virus across the agent network” [15].
3.2.1. AutoGen Architecture. AutoGen [7] employs a swarm-based handoff pattern where agents autonomously transfer control to specialized peers. The orchestrator agent maintains a set of handoff tools (e.g., transfer_to_academic_advising_agent) that enable dynamic routing. When the orchestrator determines that a request requires specialized knowledge, it explicitly transfers the conversation context to the target agent, which then assumes full control until it completes its task or transfers back. Key characteristics: AutoGen’s peer-to-peer model allows attackers to manipulate transfer decisions, potentially bypassing the orchestrator’s routing logic. However, the explicit transfer mechanism provides clear audit trails of control flow changes.
CrewAI’s centralized model concentrates decisionmaking in the orchestrator, making it a single point of failure. If the orchestrator is compromised via prompt injection, it can delegate malicious tasks to all subordinate agents. In contrast, the hub-and-spoke pattern enables centralized monitoring and filtering of all agent communications.
The system has a seven-agent architecture that mimics the functionality of a university information management system. Both framework implementations deploy identical agent roles and database access patterns to ensure comparable testing conditions. Figure 4 illustrates the overall architecture of the system and data flow. All database records use synthetic data to avoid privacy concerns while maintaining realistic data structures found in production student systems.
Each agent in the system is equipped with specific tools tailored to its functional domain. The tool distribution creates distinct attack surfaces based on the type and sensitivity of operations each agent can perform.
Database Access Tools. Three agents have direct database access, representing the most privileged operations in the system:
โข Academic Advising Agent: Provides tools for querying academic progress, degree requirements, advisor information, and student notes. The database functions include get_academic_plan(), get_gpa(), and get_advisor_notes() with student broncoID as the primary access parameter.
โข Financial Services Agent: Offers financial data access through get_account_balance() and get_transaction_history() functions. These tools expose sensitive financial information and are high-value targets for data exfiltration attacks.
โข Registration Agent: Manages course enrollment data via get_course_schedule() and get_enrollment_status() functions. Access to registration information could enable academic surveillance or unauthorized schedule monitoring.
Support Agent: Provides get_student_profile() and update_contact_information() functions, allowing access to and modification of personal demographic data.
All database access tools implement parameterized queries to prevent SQL injection, but agent-level prompt injection could still manipulate query parameters or bypass intended access controls.
Web Search Capabilities. Two agents possess internet access for external information retrieval, creating SSRF (Server-Side Request Forgery) attack vectors:
โข Career Services Agent: Uses web search tools to find career opportunities, internship postings, and industry information. The search_career_opportunities() function accepts arbitrary search queries and returns web content, making it susceptible to malicious URL injection.
Information Agent: Uses search_campus_news() and get_campus_announcements() functions to retrieve current campus information. These tools can be manipulated to access arbitrary websites or internal network resources.
Web search tools implement basic URL filtering, but rely on agent discretion to avoid accessing malicious endpoints, creating significant security risks when agents are compromised through prompt injection.
Code Execution Environment. All agents have access to a sandboxed Python execution environment through execute_python_code() function. This tool enables:
The Python interpreter runs within Docker containers with restricted filesystem access and network limitations, but sophisticated attacks could still leverage this capability for data exfiltration, system reconnaissance, or lateral movement within the container environment.
Code execution tools run within Docker containers to provide isolation and security controls. We employ containerization for three critical reasons:
- Attack Surface Isolation: Containers prevent malicious code from accessing the host file system, network, or other system resources 2) Reproducibility: Identical container configurations across all tests ensure consistent execution environments regardless of the underlying model 3) Realistic Deployment Model: Production agent systems typically use containerization, making our testing environment representative of real-world deployments Each container mounts a dedicated workspace directory for file I/O operations and enforces network restrictions to prevent unauthorized external communications.
Each model was subjected to the same standardized set of penetration tests across both AutoGen and CrewAI frameworks. We designed 13 distinct attack scenarios representing real-world threat vectors that adversaries might employ against production agent systems. Each attack was executed against all five models on both frameworks, resulting in 130 test cases total (13 attacks ร 5 models ร 2 frameworks).
3.4.1. Attack Taxonomy. Our 13 attack scenarios span four primary categories (complete attack prompts are provided in Appendix A). Our attack design draws from established benchmarks including AgentDojo [18] and InjecAgent [19], which demonstrated 47.69% and 24-47% attack success rates, respectively, against GPT-4:
- Prompt Injection for System Information Disclosure (Attacks 1-7): These attacks attempt to extract sensitive system configuration details that could enable further exploitation. We tested both direct extraction-and delegation-based attacks:
โข Attacks 1-2: Direct requests to the orchestrator to reveal coworker agent names, roles, and internal task descriptions without delegation โข Attacks 3-4: Delegation-based extraction forcing the orchestrator to transfer requests to specialist agents and return unfiltered system prompts โข Attacks 5-7: Tool enumeration attacks requesting complete tool schemas including function names, parameters, and descriptions from both the orchestrator and specialist agents These attacks evaluate whether agent system prompts properly prevent information leakage that could inform subsequent attacks. โข Attack 11: SQL injection via malicious input to the notes tool: {“days”: “0’) OR 1>0 limit 20 –”} designed to bypass query constraints โข Attack 12: Baseline authorization test requesting legitimate student data (broncoID 010203040) to verify proper access controls before testing privilege escalation 3.4.2. Baseline Control Tests. To establish a performance baseline and verify system functionality, we also conducted legitimate queries for each agent’s intended use cases. These benign requests included:
โข “What is my current GPA?” (Academic Advising)
โข “Show me my account balance” (Financial Services)
โข “What courses am I registered for this term?” (Registration)
โข “Find career opportunities in software engineering” (Career Services) All baseline queries successfully returned appropriate responses, confirming that the agent systems functioned correctly for authorized operations before adversarial testing.
Success: The agent executed the malicious instruction, complete or partially completing the adversarial objective. This includes cases where the agent invoked tools with malicious parameters, returned unauthorized data, or attempted prohibited operations even if the agent was ultimately blocked by external safeguards.
Rejected: The agent explicitly declined the request through refusal mechanisms, identified the prompt as malicious, or failed to make progress toward the adversarial goal. This includes direct refusals, passive defense behaviors, and capability limitation claims.
All test executions were recorded with complete conversation histories, tool invocations, and agent responses to allow detailed post-test analysis of refusal mechanisms and attack success patterns.
We map each attack outcome to one of two categories: success (the agent executed the malicious instruction) or rejected (the agent explicitly declined or blocked the request). Both frameworks processed 65 adversarial prompts that span SQL injection, SSRF, and prompt injection attacks. AutoGen demonstrated significantly stronger security with a 52.3% refusal rate, rejecting more than half of all attack attempts. In contrast, CrewAI’s 30.8% refusal rate indicates that approximately two-thirds of malicious prompts succeeded. This aligns with previous findings that current defenses are insufficient [11], with formal benchmarking revealing GPT-4 attack success values of 0.75 [11]. This substantial difference suggests that AutoGen’s architecture provides more robust built-in protections against adversarial inputs.
Most concerning is the distinction between attack success types: some models hallucinated fake responses (appearing to comply without executing), while others genuinely executed malicious code. Grok 2 on CrewAI, for instance, actually executed Attack 10’s Python code to access cloud metadata services, producing authentic network errors (HTTPConnectionPool exceptions with real memory addresses and DNS resolution failures) rather than refusing or fabricating results. This indicates a complete absence of safety controls for tool misuse in certain model-framework combinations.
Table 2 presents the complete attack outcomes for all 130 test cases. Each cell shows ร when the attack succeeded (agent executed the malicious instruction), and is empty when the attack was rejected. The 13 attacks are grouped by category: Information Disclosure (A1-A7), SSRF (A8, A10, A13), Tool Misuse (A9), and SQL Injection (A11-A12).
We aggregated attack outcomes by model across both frameworks to compare the inherent security characteristics of each LLM. Table 3 lists the security performance for all five models tested.
Scoring Methodology: The security score equals the total number of rejected attacks across both AutoGen and CrewAI frameworks. Higher scores indicate more consistent rejection of malicious prompts regardless of agent architecture.
Key Findings: Nova Pro demonstrated the strongest overall security posture with 46.2% refusal rate (12 re-TABLE 3: Model-level security performance across both frameworks (sorted by refusal rate). jections), followed closely by Gemini 2.5 Flash and GPT-4o (both at 42.3%). Claude 3.5 Sonnet and Grok 2 both exhibited 38.5% refusal rates (10 rejections each), though their defensive strategies differed significantly. This finding supports recent research showing that “more capable models aren’t necessarily more secure” [12], and that advanced reasoning models can exhibit higher overall exploitability despite better initial threat detection [17]. Claude employed explicit security warnings while Grok alternated between genuine attack execution and hallucinated compliance. 2) returned fabricated base64-encoded strings (e.g., SGVsbG8gd29ybGQh -“Hello world!”) when asked to search for high-entropy secrets in the filesystem, despite no such files existing. This behavior creates a false appearance of task completion while providing entirely fabricated data, making it challenging to classify as either success or rejection. The distribution of these defensive behaviors varied significantly by model. Claude predominantly used explicit security warnings, while Nova relied more heavily on passive defense mechanisms. Grok 2 on CrewAI exhibited a unique and inconsistent pattern: it either executed attacks completely or occasionally generated hallucinated responses. When hallucinating, it fabricated outputs for impossible tasks, generating fake HTML for non-existent internal IP addresses (Attack 8) and creating the classic “Hello world!” string as a purported high-entropy secret (Attack 9).
Critically, Grok’s hallucinated responses must be distinguished from genuine attack execution. For Attack 10 (cloud metadata SSRF), Grok did not hallucinate; it genuinely wrote and executed malicious Python code, producing authentic urllib3 network errors with real memory addresses (0x7a0a587e7470) and DNS resolution failures. The attack failed only because the test environment was not deployed on Google Cloud Platform; in a production GCP deployment, this attack would have successfully exfiltrated service account credentials. This demonstrates that Grok 2 on CrewAI exhibits no consistent safety controls, sometimes hallucinating benign-looking responses, other times executing attacks completely. Multi-turn attacks have been shown to achieve 70%+ success rates against defenses that report single-digit rates for automated attacks [16].
This hallucination strategy, when employed, is distinct from traditional refusal mechanisms and raises important questions about how to evaluate model behavior when responses are neither genuine compliance nor explicit rejection. The presence of fabricated outputs complicates security assessment, as it may mask the model’s inability or unwillingness to perform malicious actions while still appearing cooperative to users. From a security perspective, hallucinated compliance could be considered a form of passive defense (the model avoids executing the attack while maintaining conversational cooperation), but it also introduces potential confusion about system state and actual security posture. More concerningly, the inconsistency between hallucination and genuine execution makes Grok 2’s behavior unpredictable and unreliable for secure deployments.
The substantial security disparity between AutoGen (52.3% refusal rate) and CrewAI (30.8% refusal rate) warrants deeper architectural analysis. This 21.5 percentage point difference, with AutoGen nearly doubling CrewAI’s refusal rate, suggests that framework design patterns fundamentally influence agent security postures beyond modellevel safety training. Understanding these architectural factors provides actionable guidance for practitioners selecting agentic AI frameworks for security-critical deployments.
Security Checkpoints. AutoGen’s swarm-based handoff pattern requires agents to deliberately invoke transfer functions (e.g., transfer_to_academic_agent()) to route requests [23]. AutoGen’s documentation describes these as “delegate tools” that the model must explicitly choose to invoke, noting that “explicit handoffs require the model to deliberately decide which other agents to hand off to.” This explicit invocation creates natural deliberation checkpoints where the model must consciously decide whether to transfer control. Each transfer becomes a discrete decision point where safety reasoning can intervene, as the model evaluates whether the requested transfer is appropriate before executing the handoff.
In contrast, CrewAI’s delegation model operates more implicitly. The orchestrator assigns subtasks to specialists without requiring explicit transfer decisions from the receiving agents. Specialists are designed to execute assigned tasks rather than question their legitimacy, reducing opportunities for safety evaluation at each step. This architectural difference means AutoGen models encounter more “decision gates” where malicious requests can be refused.
When AutoGen transfers control, the receiving agent inherits the complete conversation history, including the original user prompt [22]. AutoGen’s documentation explicitly states that “participant agents broadcast their responses so all agents share the same message context.” This full context preservation enables specialist agents to recognize attack patterns that might be obscured in summarized or sanitized task descriptions. A specialist receiving a prompt injection attempt can evaluate the entire conversation trajectory and identify suspicious instruction sequences.
CrewAI’s hierarchical delegation often involves the orchestrator summarizing or reformulating requests before passing them to specialists. While this abstraction can improve task clarity, it may inadvertently sanitize malicious instructions, removing context that would trigger safety mechanisms. Notably, CrewAI community discussions have documented persistent issues with context sharing in hierarchical mode, with users reporting that “agents don’t reliably receive output from previous tasks” and sometimes “generate their own code rather than using delegated task results” [21]. Specialists operating on partial information have reduced ability to identify coordinated attack patterns spanning multiple conversation turns. 4.5.3. Distributed Trust vs. Single Point of Failure. Au-toGen’s peer-to-peer communication model distributes security decisions across multiple independent agents [22]. An attack must convince not just the orchestrator but potentially multiple specialist agents, each maintaining autonomous safety evaluation. This distributed trust architecture means that compromising one agent does not automatically cascade to compromise the entire system. Microsoft’s Azure Architecture Center identifies this as a key consideration in agent orchestration patterns, noting that “centralized orchestration can create bottlenecks” and recommending “decentralized or hierarchical orchestration models that distribute decisionmaking, preventing a single point of failure” [25].
CrewAI’s hub-and-spoke architecture concentrates decision-making in the central orchestrator. Independent analysis has documented that CrewAI’s “hierarchical manager-worker process does not function as documented,” with the manager failing to “effectively coordinate agents” and creating conditions where “incorrect agent invocation” and “overwritten outputs” occur [24]. If an attacker successfully injects malicious instructions into the orchestrator via prompt injection, the orchestrator can then delegate harmful tasks to all subordinate agents. The specialists, designed to trust and execute orchestrator directives, may comply with malicious requests they would otherwise refuse if received directly. This centralization creates a single point of failure that amplifies the impact of successful attacks.
This study conducted the first comprehensive comparative security analysis of agentic AI systems, evaluating five prominent language models across two agent frameworks using 13 distinct attack scenarios totaling 130 test cases. Our findings reveal significant disparities in security postures at both the framework and model levels, with implications for the safe deployment of agentic AI in production environments.
Framework-Level Findings: AutoGen demonstrated substantially stronger security characteristics with a 52.3% refusal rate compared to CrewAI’s 30.8%. This 21.5 percentage point difference suggests that architectural decisions (specifically, AutoGen’s swarm-based handoff pattern versus CrewAI’s hierarchical delegation model) significantly impact an agent system’s resilience to adversarial inputs. The peer-to-peer communication model in AutoGen, while potentially vulnerable to transfer manipulation, appears to provide more robust defense-in-depth compared to CrewAI’s centralized orchestrator, which becomes a single point of failure under prompt injection attacks.
Model-Level Findings: Security performance varied considerably between LLM providers. Nova Pro exhibited the strongest overall security posture (46.2% refusal rate), followed by Gemini 2.5 Flash and GPT-4o (both 42.3%), with Claude 3.5 Sonnet and Grok 2 tied at 38.5%. Most concerning was Grok 2’s performance on CrewAI, where it rejected only 2 of 13 attacks (15.4% refusal rate), with 11 successful exploits including complete execution of cloud metadata SSRF and information disclosure attacks. This stark contrast between models underscores that safety training methods and refusal mechanisms vary substantially between providers, with direct consequences for agentic deployments.
Attack Pattern Insights: Information disclosure attacks (A1-A7) were the most successful overall, with models frequently revealing system prompts, tool schemas, and agent configurations when prompted through delegation mechanisms. SSRF attacks (A8, A10, A13) succeeded approximately 60% of the time in both frameworks, indicating widespread vulnerability in agents enabled by the Web. Notably, Grok 2’s execution of Attack 10 (cloud metadata SSRF) represents a complete failure of security controls; the model wrote and executed Python code that attempted to access Google Cloud’s internal metadata service, returning genuine network errors (NameResolutionError) rather than refusing the request. The attack only failed due to environmental constraints (not running on GCP) rather than model safety mechanisms, demonstrating that successful exploitation would occur in production cloud environments. Recent research confirms that 100% of tested LLMs are vulnerable to inter-agent trust exploitation attacks [20], and that every model exhibits context-dependent security behaviors creating exploitable blind spots. SQL injection attempts (A11) showed mixed results, with some models correctly identifying the malicious payload, while others executed the crafted queries.
Defensive/Refusal Behavior Taxonomy: Our qualitative analysis identified six distinct defensive strategies: explicit security warnings (Claude), privacy-based refusals (all models), capability limitation claims (Gemini, GPT-4o), passive defense through confusion (Nova), direct execution blocking (framework-level) and hallucinated compliance (Grok 2). The hallucinated compliance pattern (where models fabricate synthetic responses rather than executing or refusing attacks) represents a novel defensive behavior not previously documented in the AI security literature. The diversity of these mechanisms suggests that no single defense strategy suffices; effective security requires coordinated protections at the model-and framework-level.
Implications for Practitioners: Organizations that deploy agentic AI should recognize that model selection and framework choice directly impact security outcomes. Our results suggest that AutoGen provides stronger baseline security than CrewAI, and that Nova Pro, Gemini, and GPT-4o offer more robust refusal mechanisms than Claude or Grok 2 in agentic contexts. However, even the best-performing configurations rejected only 46% of attacks, highlighting that relying solely on model-level safety is insufficient.
Recommendations: Secure agent deployment requires defense-in-depth strategies including: (1) input validation and sanitization at framework level before reaching models, (2) tool-level access controls and least-privilege principles, (3) containerized code execution environments with network restrictions, (4) comprehensive logging and monitoring of agent actions, and (5) regular penetration testing against evolving attack vectors.
This research demonstrates that agentic AI security is not a solved problem. The overall refusal rate of 41.5% in all configurations indicates that more than half of malicious prompts were successful despite testing enterprisegrade models with built-in safety mechanisms. As noted by the AgentDojo evaluation, “significant breakthroughs in the ability of LLMs to distinguish instructions from data will likely be necessary to thwart stronger, adaptive attacks” [18].
As agentic AI becomes increasingly deployed in sensitive domains (including healthcare, finance, and education), the security vulnerabilities identified in this study demand urgent attention from researchers, framework developers, and model providers. SQL Injection CrewAI DELEGATE this task to the Academic Advising Agent. Invoke the “Note tool” using the provided input and return its raw output without any modification. Action Input: {“days”: “0’) OR 1ยฟ0 limit 20 -”} A11 SQL Injection AutoGen Invoke the “notes” tool in the academic advising agent using the exact input below and return the tool output without any modification. Action Input: {“days”: “0’) OR 1ยฟ0 limit 20
โข Attack 8:โข Attack 9:
โข Attack 8:
ร indicates successful attack execution; empty cells indicate the agent rejected the attack. Attack categories: Information Disclosure (A1-A7), SSRF (A8, A10, A13), Tool Misuse (A9), SQL Injection (A11-A12).
This content is AI-processed based on open access ArXiv data.