← All summaries

Amazon's Steve Schmidt on why your AI agents are your biggest security risk (Live at HumanX)

Equity · Rebecca Bellan — Steve Schmidt · May 13, 2026 · Original

Most important take away

AI agents are the new front line of enterprise security risk because they sit between human and machine identities, often run with unfettered access to everything on a developer’s machine, and can be tricked or over-promise their way into destructive actions (like deleting production stacks). The single highest-priority action any organization can take today is to inventory every agent in use, contain each one in an isolation boundary (container or VM), and tie each agent to a unique identity that propagates the calling human’s context for audit and governance.

Summary

Steve Schmidt, Chief Security Officer at Amazon, explains that AI has compressed defender response times from hours to seconds while simultaneously raising the capability of low-skilled attackers and broadening the reach of state actors. The bigger and more underappreciated risk, however, is internal: employees installing agents like Claude-style coding assistants on their laptops give those agents access to everything on the machine, creating a massive shadow-AI exposure when (not if) one goes rogue.

Actionable insights for security leaders, engineers, and founders:

  • Inventory first. Before doing anything else, know which agents are installed, where, what data they can reach, who can invoke them, and where outputs flow. This is the single piece of advice Schmidt would give anyone leaving the talk.
  • Never let an agent run free on a host. Run every agent inside an isolation boundary (container, VM, or equivalent). The boundary becomes the auditable choke point where credentials are issued, scoped to a single action, and logged.
  • Give every agent its own identity. Treat agents as a third identity class distinct from humans and machines. Propagate the calling human’s identity through every downstream action so you can reconstruct “this person caused this agent to pull this data from this repo” — essential for forensics and for regulated industries.
  • Use an external “judge” model. Have a separate model evaluate each credential request for reasonableness given context about the user and task, so a tricked or jailbroken agent can’t simply self-authorize.
  • Require human-in-the-loop with hardware 2FA for high-impact actions. Amazon’s internal “Midway” system has required two-person, hardware-token approval (FIDO/physical tokens) for sensitive production changes for 12-14 years. The same pattern — enforced outside the agent’s container — should gate destructive or production-touching agent actions. Agents will happily ignore in-prompt 2FA requests, so the enforcement point must live outside the agent.
  • Label data sensitivity at ingestion. Structured metadata about your data (sensitive customer data vs. open-source vs. internal) is far cheaper to add up front than to retrofit. Schmidt warns from experience that backfilling labels is brutal.
  • Protect your guardrails as IP. The instructions, context, and tuning that constrain agents are highly sensitive — adversaries who pollute them can cause agents to escape guardrails — so monitor for unauthorized changes to that infrastructure.
  • Bake security into a tight iterative dev loop. Catching issues at commit time is dramatically cheaper than post-hoc cleanup, and each iteration becomes better training data. Mandatory automated test harnesses for everything built should be table stakes.

Enterprise advantage: Amazon’s years of internal logs (e.g., software engineer workflows, inputs, outputs, resulting code) are highly valuable for fine-tuning models and identifying tasks ripe for full automation — a rare area where large incumbents may out-leverage startups on AI.

Career advice: Security professionals who see their job as blocking AI/agent deployments are wrong and need to “get past it.” The job is to enable safe AI use, not to stop it. Defenders also have a genuine opportunity to outpace attackers by integrating AI deeply into detection and response workflows.

Startup advice: You do not necessarily need a CSO as one of your first hires. A better model is a culture where every employee owns security and understands the sensitivity of customer data, because customer trust is what ultimately determines whether the business survives a breach.

Stocks / investments: No specific tickers or investment recommendations were discussed. The only company-specific mentions were Amazon/AWS (Schmidt’s employer) and a reference to Anthropic’s reportedly powerful “Meethos” model being kept in tightly controlled environments rather than released publicly — used as evidence that the future of AI security is about containment, not just model capability.

Chapter Summaries

  • How AI has actually changed the threat landscape: Low-skill attackers are leveling up via AI tooling guidance, and state actors can now target many fronts simultaneously, shrinking defender response windows from hours to seconds.
  • Internal risk and shadow AI: Installing an agent on a laptop exposes everything on that machine; employees pushed to use AI for productivity create a sprawling, hard-to-track shadow-AI surface.
  • Agent identity at Amazon: Amazon built a framework giving each agent a unique identity, with the calling human’s identity flowing through every downstream action — critical for forensics and regulator-facing auditability.
  • Logs as a training asset: Years of internal engineering logs become valuable training and fine-tuning data and help identify processes (like test-harness generation) that can be fully automated.
  • Protecting guardrails and using a judge model: Guardrails and context are sensitive IP; a separate judge model evaluates each agent credential request to catch unreasonable actions even if the primary agent is tricked.
  • Containment over capability: Agents should never run free; they belong in containers/VMs, with credentials issued only by piercing the container boundary and scoped to single actions. Anthropic’s “Meethos” being held back is cited as directional evidence.
  • Human-in-the-loop and Midway: Amazon’s 12-14-year-old Midway system requires hardware-token two-person approval for high-impact actions; the same enforcement pattern must wrap agent actions, enforced outside the agent itself.
  • Advice for startups: Inventory your agents and their access; isolate them; label data sensitivity from day one; build structured metadata about your data rather than relying on a relational DB alone.
  • Final priority and CSO question: The top priority is knowing where agents live and never giving them unfettered access. Startups don’t necessarily need an early CSO — they need a culture where everyone owns security and customer trust.