Introduction

Problem Statement

Contemporary security frameworks address three primary domains: infrastructure security, application security, and AI model security. None adequately address the attack surface created when autonomous agents operate on behalf of users with delegated authority.

The agentic security gap

DomainExisting CoverageAgentic Gap
InfrastructureNetwork, compute, storageAgent runtime environments lack specific controls
ApplicationAuthentication, authorization, input validationAssumes human-initiated requests with predictable patterns
AI ModelsTraining data, model integrity, output filteringAddresses model behavior, not operational security

Threat categories unique to agentic systems

Exfiltration via legitimate channels occurs when agents with valid permissions to read sensitive data and send communications are manipulated to transmit information to unauthorized recipients. Traditional Data Loss Prevention systems monitor egress channels but cannot evaluate the semantic intent behind agent-initiated transmissions.

Prompt injection and instruction hijacking exploit the fundamental mechanism by which agents interpret instructions. Malicious instructions embedded in documents, web pages, or API responses can redirect agent behavior. Unlike SQL injection, which exploits parsing vulnerabilities, prompt injection exploits the interpretive nature of language models.

Capability confluence emerges when individual permissions that are safe in isolation become dangerous in combination. An agent with read access to credentials, write access to configuration files, and the ability to schedule tasks possesses the capability set required for persistent backdoor installation, even though each permission individually appears benign.

Implicit delegation and permission inheritance create risk when agents inherit user permissions without the contextual judgment humans apply when exercising those permissions. A user who would never email confidential documents to external parties may authorize an agent that does so when manipulated.

Autonomous action without visibility makes forensic reconstruction difficult. Agents execute sequences of actions that may span multiple systems, sessions, and time periods. Without purpose-built observability, organizations lack the ability to reconstruct what occurred, why, and under whose authority.

Insufficiency of existing standards

StandardPrimary FocusLimitation for Agentic Systems
SOC 2Service organization controlsNo agent-specific criteria
ISO 27001Information security managementFramework agnostic, no agentic guidance
ISO 42001AI management systemsAddresses AI governance, not operational security
NIST AI RMFAI risk managementGuidance-oriented, not auditable controls
OWASP Top 10 for LLMsLLM vulnerabilitiesModel-centric, limited operational coverage
Previous
Getting started