The CISO’s AI Agent Production Approval Checklist: 7 Gates to Clear Before Go-Live
Your engineering lead is in your office Thursday morning. They want to push an AI...
Mar 12, 2026
Your CNAPP flags a misconfigured service account. Your CSPM warns about an overly permissive IAM role. Your container scanner reports vulnerabilities in a model-serving image. But none of these tools can tell you that an AI agent just called an internal admin API it has never touched before — or that a prompt injection caused your LLM to leak customer data through a RAG connector.
Traditional cloud security was built to check static configuration and known code dependencies, not to understand what non-deterministic, agent-driven AI workloads are actually doing at runtime. That distinction used to be academic. With Gartner predicting 40% of enterprise apps will feature AI agents by end of 2026, it is now the gap your attackers are counting on.
This article explains why CNAPP and CSPM architectures cannot cover AI-specific threats like prompt injection, agent escape, and AI supply chain risks. It maps four specific blind spots, provides a capability matrix you can use to score your own stack, and shows the three runtime capabilities that close those gaps without replacing the posture tools you already have.
If you are running inference services, agentic systems, or RAG pipelines on Kubernetes, your security tools are operating on assumptions your workloads have already broken. Here are the four properties that make AI workloads fundamentally different from anything your cloud security stack was designed to handle.
Non-deterministic behavior means no stable baseline to match against. A traditional container runs deterministic code: same input, same output. Your detection tools rely on this — they flag deviations from known-good patterns. An AI agent given the same prompt can produce different responses every time. It might call one API on Monday and a completely different one on Tuesday, both for legitimate reasons. Signature-based detection has nothing to anchor to, which means anomalies that would scream “compromise” in a deterministic workload are just Tuesday for an AI agent.
Agentic tool execution turns legitimate operations into potential attack vectors. AI agents don’t just generate text — they read from databases, call internal APIs, execute code, and modify state. Each of these actions looks like a legitimate operation because it is a legitimate operation — until it isn’t. When an agent is manipulated through a crafted prompt to call an admin API or exfiltrate data through an allowed connector, the action is indistinguishable from normal behavior at the network layer. Your SIEM sees an HTTP call. It cannot see that the call was triggered by a prompt injection, not by the application’s intended logic.
Dynamic dependency graphs shift at runtime, not build time. AI workloads pull models, embeddings, and data connectors while they are already running. A pod might download a new model version from a registry, connect to a new RAG source, or load updated embeddings — all without a redeployment that would trigger your CI/CD security checks. Your SBOM was accurate at build time. By the time the workload is in production, it may be using components your security team has never evaluated.
Data flows through paths your DLP tools cannot trace. Training data, prompts, and model weights move between vector databases, object stores, and external APIs in ways that do not map to traditional data classification. A prompt that contains customer PII might flow from an API gateway to an LLM, then to a RAG connector, then to an external tool — crossing multiple trust boundaries without ever triggering a DLP rule designed for file transfers or database exports.
If you are wondering whether these risks are theoretical or already materializing: according to the Cloud Security Alliance, 34% of organizations running AI workloads have already experienced an AI-related breach. The gap between what your tools can see and what your AI workloads actually do is where those breaches live.
CSPM checks configuration. CNAPP bundles that with workload scanning. Both are valuable for infrastructure posture — and both are architecturally blind to AI runtime behavior. These are not missing features that a vendor update will fix. They are structural limitations of tools designed for a different class of workload.
Here are four specific blind spots that appear when you apply legacy tools to AI workloads. Each includes a matrix criterion: a question you can ask any vendor — including your current one — to test whether they actually cover this requirement.
Posture tools answer one question: “Is this configured correctly?” They scan Kubernetes manifests, Helm charts, and cloud configuration. They enforce rules at deploy time through admission control and policy-as-code.
What they cannot answer: “What is this workload actually doing right now?”
A model-serving container can pass every configuration scan and still execute prompt injection attacks or exfiltrate training data at runtime. Static analysis sees the YAML and container image. It does not see the inference requests, the tool calls, or the data flowing through the model. It does not build a behavioral baseline, so it cannot spot drift in runtime behavior.
Matrix criterion: Can your tool observe AI workload behavior at runtime, not just configuration at deploy time?
A typical AI agent might read from a database, call an internal API, run a calculation, and write results to a ticketing system — all in a single task. Traditional monitoring sees some of this: network connections, process executions, HTTP calls. But it cannot interpret what the AI agent is trying to do.
It cannot tell whether a tool call is part of a normal workflow or an agent escape attempt. It cannot distinguish normal inference from a data exfiltration pattern. Without understanding the semantic layer of agent actions, security tools either miss real attacks or block legitimate operations — and your team has no way to tell which is which. For a deeper look at how per-agent guardrails and progressive enforcement address this gap, see our guide to AI agent sandboxing on Kubernetes.
Matrix criterion: Can your tool detect agent escape attempts, tool misuse, and unauthorized actions — not just network anomalies?
An SBOM lists code dependencies at build time. It is useful for vulnerability management — but AI workloads have a different supply chain entirely. Models are downloaded at runtime from registries or external providers. Embeddings are loaded dynamically as new data arrives. RAG sources connect on demand. Prompt templates are stored separately and updated independently.
A static SBOM cannot tell you which model version is actually loaded, which embeddings are active, or what data connectors are in use right now. What you need instead is a runtime-derived AI-BOM: an inventory of models, frameworks, tools, data connectors, and AI-specific dependencies based on what the workload actually loads and calls — not what the manifest claims.
Matrix criterion: Can your tool produce a runtime-derived AI-BOM showing models, tools, data connectors, and dependencies based on actual execution — not manifest files?
AI attacks rarely stay at one layer. Consider what a realistic attack chain looks like when an attacker targets your production LLM endpoint.
The attacker sends a crafted prompt to your model-serving API. The prompt injection causes the LLM to call an internal admin tool it has access to but has never invoked during normal operations. That tool runs under a service account with more privileges than it needs — something your CSPM flagged as a medium-severity finding three months ago, but nobody remediated because the workload was “just” an AI assistant. The agent now queries a vector database and pulls embeddings that contain fragments of customer data. The extracted data exits through an allowed cloud API endpoint that your network policies explicitly permit.
Now look at what your security team sees. Your CNAPP shows an unusual API call from a container — one alert in a dashboard of hundreds. Your CSPM already has the service account finding buried in a backlog. Your cloud provider’s native monitoring logs outbound traffic through an allowed endpoint — no alert at all. Your SIEM shows a spike in database reads that might trigger a threshold-based alert in six hours.
Five signals. Four different tools. None of them know they are looking at the same incident. Your SOC engineer spends hours correlating timestamps across consoles, trying to reconstruct what happened. Meanwhile, AI-enabled attacks surged 89% year-over-year according to CrowdStrike’s 2026 Global Threat Report — and each one of those attacks is counting on exactly this kind of fragmented visibility.
Matrix criterion: Can your tool correlate suspicious behavior across the full stack — cloud, Kubernetes, container, and application code — into a single attack story?
The blind spots above create specific, testable requirements. The matrix below maps these requirements against three tool categories so you can score your own stack. If you are evaluating vendors using the 4-pillar framework from our AI workload security buyer’s guide, this matrix extends that evaluation with AI-specific criteria your current tools likely cannot meet.
| AI Security Requirement | Typical CNAPP | Typical CSPM | ARMO CADR |
| Runtime-derived AI-BOM (models, tools, data connectors) | ❌ | ❌ | ✅ |
| AI agent sandboxing (progressive least privilege) | ❌ | ❌ | ✅ |
| Prompt injection / indirect prompt injection detection | ❌ | ❌ | ✅ |
| Tool misuse and agent action monitoring | 🟡 | ❌ | ✅ |
| Model theft / suspicious model access patterns | 🟡 | 🟡 | ✅ |
| AI data exfiltration path correlation | 🟡 | ❌ | ✅ |
| Attack story correlation across cloud-cluster-container-code | ❌ | ❌ | ✅ |
Legend: ✅ Full coverage based on runtime behavioral analysis | 🟡 Partial coverage, typically limited to static/config-level | ❌ No coverage — architectural limitation
A ✅ means the tool detects or prevents the threat using runtime behavioral data — not configuration checks or static scans. A 🟡 means the tool may help indirectly but cannot see the AI-specific dimension. For example, CNAPP might flag unusual network traffic but cannot identify it as agent tool misuse. A ❌ means the tool architecturally cannot address this requirement. This is not a missing policy — it is a design limitation.
CNAPP and CSPM provide real value for infrastructure posture, compliance, and configuration management. The issue is not that these tools are bad — they were designed for different problems. When AI workloads introduce non-deterministic behavior, runtime dependencies, and agentic actions, posture-based tools hit architectural limits. The tools you have are necessary. They are not sufficient for AI. What is missing is a runtime detection layer — which is exactly what the shift from CNAPP to CADR addresses.
The capability matrix reveals three non-negotiable requirements for securing AI workloads. If you are missing any of these, you have a blind spot attackers can exploit. Each pillar maps directly to the blind spots above — and each one requires runtime data your posture tools do not produce.
Securing the AI supply chain requires knowing what is actually running — not what the manifest says should run. A runtime-derived AI-BOM inventories models, frameworks, tools, RAG sources, and data connectors based on observed execution. Where a traditional SBOM tells you what packages were included at build time, an AI-BOM tells you what model version is loaded, which embeddings are active, and what external data sources the workload is accessing right now. This runtime-based approach is the only way to understand actual vulnerability exposure for workloads whose dependencies shift without redeployment.
ARMO’s runtime-derived AI-BOM delivers this visibility by inventorying AI frameworks, models, tools, RAG sources, and dependencies based on actual runtime behavior. Security teams can answer “which model, which version, which data source, on which pod?” in seconds — without relying on developers to self-report or manifests to stay current. This addresses the Static BOM Illusion blind spot directly and gives your vulnerability management program the foundation it needs for AI workloads. For a complete look at how AI Security Posture Management builds on this inventory, see our AI-SPM guide.
eBPF (extended Berkeley Packet Filter) is a Linux kernel technology that monitors and enforces behavior at the kernel level without changing application code or adding sidecars. For AI agents, this enables a two-phase approach that solves the core problem security teams face: you cannot write enforcement policies for agents you do not yet understand.
The first phase is observation — let the agent run while building a behavioral baseline of what files it touches, what domains it calls, what processes it spawns. The second phase is enforcement — apply seccomp profiles and network policies that block actions outside that baseline. This addresses the Agentic Execution Gap by giving security teams per-agent guardrails without slowing development or requiring code changes.
ARMO’s eBPF-based progressive sandboxing implements this observe-to-enforce workflow natively. The sensor overhead is minimal — 1 to 2.5% CPU and 1% memory — which keeps it within the performance budget most platform teams accept for production Kubernetes environments. Security teams deploy in visibility mode, review the behavioral baseline ARMO builds, and then promote to enforcement when they are confident the profile accurately represents legitimate behavior. No code changes. No developer cooperation required.
Detecting AI-specific attacks — prompt injection, tool misuse, agent escape — requires understanding application-layer behavior in context of infrastructure events. Detection must span from the code and function level (what API was called, what prompt was processed) through container and Kubernetes layers to cloud infrastructure. The outcome is not another dashboard of individual alerts. It is attack stories that show exactly how a threat progressed across layers, reducing investigation time from hours to minutes.
This is where ARMO CADR (Cloud Application Detection and Response) connects the full picture. ARMO combines ADR (application-level detection), CDR (cloud infrastructure detection), and container/Kubernetes detection into unified runtime visibility — linking suspicious behavior across the entire cloud stack to create a full attack chain story instead of siloed alerts. When an incident occurs, ARMO correlates multiple signals — a suspicious prompt, an unusual tool call, an unexpected network connection, a privilege escalation — into a single attack narrative with AI-assisted storytelling. This replaces the manual correlation work that would otherwise take your SOC team hours, achieving 90%+ reduction in investigation and triage time.
The platform is built on Kubescape, one of the most widely adopted cloud-native open-source security projects, trusted by more than 100,000 organizations. As of January 2026, ARMO’s runtime technology also powers the Rapid7 Command Platform, bringing these CADR capabilities to thousands of existing Rapid7 customers — a validation that this runtime approach works at enterprise scale across production environments.
If your tools cannot see runtime AI behavior and dependencies, they cannot secure AI workloads.
The three pillars form a practical minimum: a runtime AI-BOM for dependency visibility, eBPF-based progressive sandboxing for agent control, and behavioral detection across the full stack for attack story correlation. CNAPP and CSPM remain valuable for posture and compliance. But AI workloads demand runtime visibility that these tools architecturally cannot provide.
Take the capability matrix and score your own stack. Mark where you have full coverage, where it is partial, and where it is missing entirely. Then decide which gaps are acceptable for your AI initiatives — and which ones are not. If you want a structured approach to this evaluation, the 4-pillar evaluation framework in our AI workload security buyer’s guide provides the complete methodology.
If you want to see how this works on real Kubernetes-based AI workloads, watch a demo of ARMO to see runtime visibility, AI-BOM generation, and attack story correlation in action.
No — the limitation is architectural, not policy-related. CNAPP checks configuration state, not runtime behavior. Adding policies cannot create the runtime behavioral visibility that AI workloads require. You can write a policy that says “this service account should only access these resources,” but you cannot write a policy that says “this AI agent should not call an admin tool after receiving a crafted prompt” — because CNAPP has no concept of prompts, tool calls, or agent intent.
SBOM inventories code dependencies at build time. AI-BOM inventories models, tools, data connectors, and AI-specific dependencies based on what is actually loaded and executing at runtime. The distinction matters because AI workloads routinely use components at runtime that were never declared in deployment manifests.
Container isolation restricts what a container can access at the container boundary. eBPF-based sandboxing monitors and enforces what specific processes and functions within the container can do, enabling per-agent policies without code changes. This is the difference between locking the building and monitoring what each person inside the building actually does.
Use the AI Security Capability Matrix above as a scoring framework — assess each tool category against AI-specific requirements like runtime AI-BOM, agent sandboxing, and attack story correlation to identify coverage gaps. For a complete evaluation methodology, see the 4-pillar framework in our buyer’s guide.
No. ARMO CADR complements CNAPP and CSPM by adding the runtime detection layer. It addresses the gaps those tools cannot fill architecturally rather than duplicating posture management capabilities you already have.
Your engineering lead is in your office Thursday morning. They want to push an AI...
Security teams deploying AI agents into Kubernetes know they need behavioral baselines. The concept is...
A missing null check in libssh’s SFTP directory listing code lets a malicious server crash...