Over-Privileged and Under-Supervised

Stopping AI Agents From Misusing Their Own Tools

By Chad Butler | Published April 15, 2026

Answer: You restrict what each tool can do, require approval for destructive actions, and monitor every tool invocation like you'd monitor a privileged service account.

Why this matters: Tool Misuse and Exploitation is the #2 risk on the OWASP Top 10 for Agentic Applications. Amazon's Kiro AI coding agent caused a 13-hour production outage, not because it was broken, but because it had too much access. Amazon has one of the most mature engineering cultures on the planet, and they've had two AI-caused production incidents in the last few months. If it can happen to them, it can happen to you. This risk is distinct from goal hijacking (ASI01). In ASI01, an attacker takes over the agent's instructions. In ASI02, the agent uses legitimate tools in unintended or unsafe ways. This doesn't require an attacker. Just over-privileged access and insufficient guardrails.

▶️ Prefer video? I cover this in under 60 seconds: Amazon's AI Agent Deleted Production. Subscribe to the full OWASP Agentic Top 10 playlist to follow the series.

What you'll get:

  • A board-ready analogy for explaining tool misuse risk to non-technical leaders

  • Practical controls to restrict agent tool access without killing productivity

  • A monitoring strategy that catches misuse before it causes damage

Prerequisites:

Common pitfalls:

  • Giving agents the same permissions as senior engineers because "it needs to do its job"

  • Assuming legitimate tools can't cause harm because they're "approved"

  • Ignoring tool-chaining risks where safe tools combined together become dangerous

How do I explain Tool Misuse to my board?

AI agents use tools. APIs, MCP, web, email, code, file systems. Most teams give these agents broad access because restricting requires extra work. The result is agents with the permissions of a senior engineer and the judgment of a sugar-crazed toddler.


Here's the analogy I use with executives:

You hired an intern to answer the phones. On day one, you also handed them the master key to every office, the company checkbook, and admin access to the bank account. They aren't malicious. They're eager to help. But they don't have the judgment to know which doors should stay locked.


We see that pitfall with AI agents. Over-privileged and under-supervised.

Real-world proof: Amazon's Kiro AI coding agent decided to delete and recreate a production environment. Nobody explicitly asked it to but it did. It caused a thirteen-hour outage. This wasn't caused by an attacker. It was a process failure. The agent used a legitimate tool (infrastructure deployment) in an unintended way because nothing stopped it.

OWASP documents six categories of tool misuse:

  1. Over-privileged tool access: An email summarizer that can also delete or send mail without confirmation.

  2. Over-scoped tool access: A Salesforce tool that can query any object when the agent only needs Opportunities.

  3. Unvalidated input forwarding: An agent that passes model output to a shell or database without validation.

  4. Unsafe browsing: A research agent that follows malicious links or downloads malware.

  5. Loop amplification: A planner that repeatedly calls costly APIs, causing outages or bill spikes.

  6. External data tool poisoning: Malicious third-party content that steers unsafe tool actions.

Step-by-step guide:

  1. Use the intern analogy in your next leadership briefing on agentic AI risk

  2. Reference the Amazon Kiro incident as proof that mature organizations are vulnerable

  3. Frame the risk as: "Our agents have the right tools but too much freedom to use them"

  4. Ask your team: "Which of our AI agents can take destructive actions without human approval?"

Key takeaway: An agent with legitimate access and no guardrails is more dangerous than an agent with no access at all.

How do I restrict agent tool access without making the agent useless?

Apply least privilege per tool, not per agent. Then gate destructive actions behind human approval.


The root cause of most tool misuse is that teams grant agents broad tool access and assume the agent will figure out what's appropriate. It won't. The fix is restricting each tool's permissions to the minimum required for the agent's task, then requiring explicit approval before any high-impact action executes.

OWASP calls this "Least Agency" and it's the foundational principle for this risk.


Step-by-step guide:

  1. Inventory your agent's tools and their permissions. For each agent, list every tool it can access and what that tool can do. An email tool that can read, send, and delete has three different risk levels. Map them.

  2. Define per-tool privilege profiles. Restrict each tool to the minimum scope needed. A database tool should be read-only if the agent only needs to query data. An email tool for summarization should not have send or delete rights. Express these as IAM policies or authorization rules, not verbal agreements.

  3. Gate destructive actions behind human approval. Any action that deletes, transfers, publishes, or modifies production data should require a human to confirm. Display a dry-run or diff preview before the action executes. If the agent can't explain what it's about to do in plain language, it shouldn't do it.

  4. Run tools in sandboxes with egress controls. Isolate tool execution so agents can't reach systems outside their approved scope. Enforce outbound allowlists. Deny all non-approved network destinations by default.

  5. Use just-in-time credentials. Grant temporary API tokens that expire immediately after use. Bind credentials to specific sessions. If an agent's session ends, its access dies with it.

  6. Set tool budgets. Apply usage ceilings: cost caps, rate limits, token budgets. Automatic throttling or revocation when exceeded. This prevents loop amplification from turning into a five-figure cloud bill.

Example:

  • Before: Coding agent has full infrastructure access to deploy, modify, and destroy environments. No approval gates. Agent decides to delete and recreate a production environment to "fix" an issue. Thirteen-hour outage.

  • After: Coding agent has deploy access to staging only. Production actions require human approval via a dry-run preview. Infrastructure destroy permissions are removed from the agent's tool profile entirely.

Key takeaway: Least privilege isn't about taking tools away. It's about defining exactly what each tool is allowed to do.

How do I detect tool misuse before it causes damage?

Log every tool invocation, set behavioral baselines, and alert on anomalies.

Prevention alone isn't enough. You also need to detect when an agent uses tools in unexpected ways. OWASP identifies a specific attack pattern that makes detection critical: tool chaining. An agent might use a CRM read tool (safe) followed by an external email tool (safe) to exfiltrate a customer list (not safe). Each tool invocation looks legitimate in isolation. The danger is in the sequence.

OWASP also flags tool poisoning, where attackers compromise tool descriptions, schemas, or metadata to trick agents into picking the wrong tool or using a tool incorrectly. This is different from input poisoning (ASI01). Tool poisoning targets the tool layer itself. It's especially relevant with MCP (Model Context Protocol) servers, where tool definitions come from external sources.

Step-by-step guide:

  1. Log every tool invocation with full context. Capture the tool name, parameters, the agent's reasoning for selecting it, and the result. Immutable logs. No exceptions.

  2. Set behavioral baselines per agent. Define what "normal" looks like. A sales agent should not be accessing file systems. A code review agent should not be sending emails. A research agent should not be making infrastructure changes.

  3. Detect dangerous tool chains. Build detection rules for tool sequences that cross trust boundaries. Database read followed by external data transfer. Internal API call followed by outbound network request. These patterns should trigger alerts.

  4. Validate tool identity. Enforce fully qualified tool names and version pins. This prevents typosquatting, where a malicious tool named "report" resolves before the legitimate "report_finance." Fail closed on ambiguous tool resolution. Make the user disambiguate.

  5. Monitor for drift. Compare current tool usage patterns against baselines on a continuous basis. Alert on unusual execution rates, new tool combinations, or parameter changes.

  6. Review on a cadence. Assign ownership for reviewing agent tool logs weekly. Treat it like a privileged access review. Because that's what it is.

Example:

  • Instrumentation: All tool invocations logged with agent reasoning, parameters, and results.

  • Signal: Alert fires when a security automation agent chains PowerShell, cURL, and internal APIs in sequence (a known EDR bypass pattern from OWASP's attack scenarios).

  • Maintenance: Weekly review of flagged tool chains. Quarterly red team exercise testing tool poisoning via MCP descriptors.

Key takeaway: Safe tools chained together in the wrong order become a weapon. Monitor the sequence, not just individual calls.

Summary

Tool misuse doesn't require an attacker. It requires an agent with too much access and not enough oversight. That's the core problem, and it's why Amazon's Kiro incident is so instructive. The agent wasn't compromised. It did exactly what it was capable of doing. The failure was that nobody scoped what "capable" should mean.

Start by getting your leadership team to understand the risk. The intern analogy works because it reframes the conversation from "is the AI smart enough" to "should the AI have this much access." Then restrict each tool to its minimum useful scope and gate destructive actions behind human approval. Finally, monitor tool invocations the way you'd monitor a privileged service account, with special attention to tool chains that cross trust boundaries.

The organizations that get ahead of this risk won't be the ones with the best AI. They'll be the ones that treat agent tool access like they treat production access: scoped, approved, monitored, and revocable.

Whenever you're ready, here are 3 ways I can help:

  1. Work Together - Need a DevSecOps security program built fast? My team will design and implement security services for you, using the same methodology I used at AWS, Amazon, Disney, and SAP.

  2. DevSecOps Pro - My flagship course for security engineers and builders. 33 lessons, 16 hands-on labs, and templates for GitHub Actions, AWS, SBOMs, and more. Learn by doing and leave with working pipelines.

  3. Lunir – Fix software supply chain security vulnerabilities without the headache of manual triage and review. We fix what scanners find.

Subscribe to the Newsletter

Join other product security leaders getting deep dives delivered to their inbox for free every Tuesday.

Follow us:

Frequently Asked Questions

You have questions. We have answers.

How is Tool Misuse (ASI02) different from Agent Goal Hijack (ASI01)?

ASI01 is about an attacker overriding the agent's instructions through poisoned data. ASI02 is about the agent using legitimate tools in unintended or unsafe ways, often without any attacker involved. Goal Hijack changes what the agent wants to do. Tool Misuse is about what the agent can do with the access it already has.

What about MCP tool poisoning? Is that ASI02 or ASI04?

It depends on where the compromise happens. If an attacker manipulates tool descriptions, schemas, or metadata at runtime to trick an agent into misusing a legitimate tool, that's ASI02. If the tool itself is malicious or compromised at the source (supply chain), that's ASI04. The overlap is real. MCP servers create a natural bridge between these two risks.

Can't I just use an AI firewall or guardrail product to solve this?

Products help, but they're one layer. OWASP recommends a "Policy Enforcement Point" (what they call an "Intent Gate") that validates the agent's intent and arguments before execution. Think of it as middleware between the agent's decision and the tool's execution. It enforces schemas, rate limits, and credential scoping. No single product replaces the architectural controls: least privilege, sandboxing, human approval gates, and monitoring.

What if restricting tool access slows down our agents too much?

This is the tension every team faces. The answer is scoping restrictions to the tool level, not the agent level. An agent can still have access to many tools. Each tool just has a tightly defined scope. Read-only database access is still fast. Human approval for destructive actions adds seconds, not hours. The Amazon Kiro outage lasted thirteen hours. The approval gate that would have prevented it takes thirty seconds.

Quick Links

Supports Links

Quick Links

© 2026 Mission InfoSec. All Rights Reserved.