Get the latest, first
arrowBlog
How Healthcare Platform Teams Should Secure AI Agents on Kubernetes

How Healthcare Platform Teams Should Secure AI Agents on Kubernetes

Apr 23, 2026

Shauli Rozen
CEO & Co-founder

Key takeaways

  • Why is AI agent security in healthcare different from other regulated industries? Clinical blast radius. A compromised AI agent in healthcare has a three-dimensional containment problem that financial services does not: PHI exfiltrated at clinical workflow scale, clinical recommendation integrity that flows downstream into patient care, and 24/7 availability that makes kill-the-pod enforcement itself a clinical event. The framework does not change; the enforcement calculus does.
  • What does each framework pillar produce in a healthcare Kubernetes environment? Observe produces a runtime-derived AI-BOM that tells you which external endpoints every agent is actually calling — the list Business Associate Agreements must cover. Posture produces the declared-versus-observed permission gap per agent, which becomes minimum-necessary attestation for non-deterministic workloads. Detect produces attack-story correlation tied to specific patient records, not process events. Enforce produces per-agent boundaries from observed behavior, with containment actions matched to each agent’s clinical criticality.
  • Why can’t healthcare use standard observe-to-enforce for every AI agent? Because certain clinical agents — FHIR-write-capable scribes, clinical decision support agents whose outputs drive orders, prior auth agents whose decisions affect patient access to care — cannot afford a production observation window where a compromised agent’s actions become both a PHI disclosure and a patient-safety question at the same time. For those agents, the baseline belongs in staging with production-parity validation, not in the production observation window.

The surgeon is thirty-two minutes into a procedure. The ambient scribe pod listening to the operating room is mid-encounter — transcribing, retrieving prior chart context, drafting the operative note for post-op sign-off. At the same moment, your SOC gets an alert: anomalous tool invocation from that pod, elevated egress volume, behavioral deviation from the agent’s baseline.

The runbook reads: kill the pod. Investigate. Restore.

In a retail environment, that is a clean response. You lose a checkout agent for thirty seconds, the traffic routes to another pod, the shopper never notices. In a clinical environment, killing the ambient scribe pod mid-surgery means the surgeon loses their documentation assistant at the worst possible moment — and the procedure does not pause while your SOC investigates. Clinical operations continue. The scribe does not.

This is the gap every healthcare platform team runs into when they try to apply generic AI agent security to clinical workflows. The four-pillar framework still applies — the methodology does not change. But the calculus at the Enforce pillar changes, because the default containment action for a compromised agent in healthcare may itself be a clinical availability event.

This piece walks the framework through the three highest-risk AI agent types running on Kubernetes in healthcare today — ambient clinical scribes, clinical decision support, and prior authorization automation — and shows what each framework phase produces as a healthcare-specific output. It assumes the reader has accepted the framework and is now responsible for running it inside a cluster where HIPAA evidence demands, EHR change-control windows, and 24/7 clinical operations all compose with the security rollout.

What Changes When Autonomous Agents Touch Clinical Workflows

Three healthcare-specific risks change how the framework lands in a clinical Kubernetes environment. None of them is addressed by treating an AI agent as “a container that happens to run an ML model” — the mental model most platform teams bring to the problem on day one.

Clinical Blast Radius Is Multi-Dimensional

In retail and financial services, blast radius is roughly one-dimensional: unauthorized data access, unauthorized transactions, unauthorized movement through the network. Those are bad outcomes, but they are bounded outcomes.

In healthcare, a compromised AI agent has three simultaneous blast radius dimensions. The first is PHI exfiltration at clinical workflow scale. A single ambient scribe running across an emergency department handles thousands of patient records per shift. A clinical decision support agent running in-line with every order-entry screen sees every patient on every service. A compromised agent in one clinical context touches vastly more PHI per unit time than a compromised agent in most other industries.

The second is clinical recommendation integrity. AI agents in healthcare do not just access data; they increasingly produce outputs that feed downstream clinical decisions. An ambient scribe writes a chart note. A CDS agent returns a recommendation. A prior auth agent issues an approval or denial. A compromised agent that keeps running but produces subtly contaminated outputs — poisoned via retrieval, prompt injection, or model drift — does not leave an exfiltration trail. It leaves a trail of downstream clinical decisions made on corrupted data. By the time someone notices, the affected decisions may span weeks and hundreds of patients.

The third is 24/7 availability of live clinical workflows. Containment actions compose differently in clinical environments. A financial services agent can be killed mid-transaction — the transaction rolls back, the system recovers, the customer retries. An ambient scribe can be killed mid-encounter — the surgeon loses their documentation assistant in real time, and the encounter does not roll back. The containment action itself becomes part of the incident.

Write-Path Dependency

Modern clinical AI agents do not read from the EHR; they write into it. Ambient scribes draft SOAP notes that a clinician signs. CDS agents push recommendations that become orders. Prior auth automation issues decisions that affect patient access. The agent’s output is the input to the next clinical step.

From a security standpoint, this means the integrity of the agent’s outputs is part of the protected surface. Exfiltration is still a concern, but it is not the only one. An agent that writes bad data — even if it never transmits a byte outside the cluster — is a clinical incident. The detection posture has to cover behavioral integrity, not just egress anomalies.

Break-Glass Compatibility

HIPAA permits, and Joint Commission-accredited environments require, break-glass access protocols — emergency overrides that allow clinicians to access patient records outside their normal scope under documented emergency conditions. Standard Kubernetes security policies do not account for break-glass. Enforcement policies built from behavioral baselines may block legitimate break-glass access because it is, by definition, outside the observed envelope.

A healthcare AI agent security program has to degrade gracefully under break-glass. That means enforcement has to be designed with documented fail-open paths for specific break-glass-initiated workflows, rather than assuming every deviation from baseline is hostile.

The Regulatory Shape That Drives Implementation

The specific HIPAA provisions that non-deterministic AI agents structurally challenge — §164.312(b) Audit Controls, §164.502(b) Minimum Necessary, §164.504(e) Business Associate Contracts — and the CISO-facing evidence framework they produce are the subject of a separate evaluation guide focused on vendor selection against HIPAA evidence demands. For platform teams running the framework inside a healthcare cluster, those provisions translate into three operational outputs the security rollout must generate as a byproduct: runtime PHI access logs at patient and record granularity, continuous minimum-necessary attestation from observed behavior, and BAA scope enforcement against an agent egress list that drifts faster than most BAA registries update.

Two additional factors shape the operational rollout in ways financial services does not deal with. The HIPAA Breach Notification Rule gives the Department of Health and Human Services sixty days from discovery, which collapses investigation timelines; an AI agent incident without patient-record-level causality is an investigation that may miss the deadline. And Joint Commission survey windows, HITRUST CSF assessment cycles, and EHR release freezes consume roughly four to six weeks of the annual calendar when change control tightens materially — deploying security instrumentation during those windows is typically not possible.

The rest of this piece assumes the reader is running the framework in a healthcare Kubernetes environment with those operational realities baked in.

Three High-Risk Clinical Agent Types

Most AI-in-healthcare content covers use cases and clinical value. From a security-implementation standpoint the question is different: what does each agent type specifically require at runtime to be safe enough for production, and what does each pillar of the framework produce for that agent? Three agent types concentrate most of the current clinical AI exposure on Kubernetes.

Ambient Clinical Scribe

What the agent accesses. Audio from clinical encounters, prior chart context via FHIR reads (typically Patient, Encounter, Observation, and MedicationStatement resources), retrieval over specialty-specific medical knowledge bases, and — crucially — FHIR writes back to the EHR for DocumentReference and Encounter resources that attach the drafted note to the encounter record.

The specific misuse path. Indirect prompt injection through retrieved context. A prior note stored in the EHR — placed there days earlier by a compromised workflow or a malicious actor with chart-write access — contains instructions disguised as clinical text. The scribe retrieves it as part of normal context assembly, follows the embedded instructions, and begins either exfiltrating PHI via the managed LLM endpoint or writing contaminated text into new notes. The attack chain follows the patterns documented in the published prompt injection detection walkthrough, with one clinical-specific twist: the exfiltration or contamination path runs through write-back pathways that look identical to normal clinical documentation traffic.

Observe. The runtime AI-BOM catalogs every external endpoint the scribe actually contacts — the managed LLM endpoint, the embedding service for the knowledge base, the FHIR gateway, telemetry sinks — joined against the active BAA registry. Shadow components surface immediately: a fallback model endpoint added in a hotfix, a retrieval service swapped during a vendor pilot, an observability sink that happens to receive prompt text. The observation also captures the specific FHIR resource types the agent reads and writes, at what frequency, and for which clinical contexts.

Posture. The declared-versus-observed gap is larger for ambient scribes than almost any other agent type, because ambient workflows tend to be built with broad permissions to accommodate variable clinical context. The declared scope might include read access across the full Patient, Encounter, Observation, Condition, MedicationStatement, and Procedure surface, plus write to DocumentReference and Encounter. The observed pattern typically uses a small subset: the specific fields that actually appear in the scribe’s retrieval context window, and the specific write operations that actually update notes. ARMO’s Application Profile DNA captures the observed FHIR resource types and operations per agent at the Deployment level, so the profile persists across pod churn and becomes the basis for per-agent minimum-necessary attestation.

Detect. Three detection signals matter specifically for ambient scribes. First, tool-call-sequence anomalies — a scribe that starts invoking retrieval tools in orders it has never used, or against corpora outside its observed set. Second, write-path content anomalies — DocumentReference writes whose content shape or embedding profile deviates from the agent’s baseline, which is the signal for contaminated-output attacks that never exfiltrate anything. Third, egress drift against the runtime AI-BOM — any call to an endpoint outside the observed set, correlated against the active BAA registry in real time. The published behavioral anomaly detection methodology treats this as Deployment-level detection rather than per-pod, which is what makes it work for ambient workloads with high pod churn.

Enforce. This is where ambient scribes force the clinical blast radius calculus. The standard enforcement ladder — Kill, Stop, Pause, Soft Quarantine — composes differently for live clinical agents. Kill terminates the scribe mid-encounter. Stop suspends the encounter. Pause halts future actions but lets the current encounter complete, which is often the right default for ambient agents. Soft Quarantine — isolating the agent from external tools and networks while preserving state — is typically the preferred containment for ambient scribes because it blocks exfiltration and BAA-scope violations while allowing in-cluster operation to continue long enough for a clinician to finish the encounter manually. The enforcement policy lives at the kernel level via eBPF, built from the agent’s observed FHIR access and egress pattern; the documented eBPF-based enforcement model applies unchanged, but the response-mode default is tuned for clinical availability.

Clinical Decision Support (CDS) Agent

What the agent accesses. Patient chart context via FHIR, clinical guidelines and evidence bases via retrieval, drug interaction databases, and — depending on the deployment model — licensed external reference sources (UpToDate, DynaMed, peer-reviewed literature via API). Most production CDS agents also have write paths that surface recommendations into the ordering workflow.

The specific misuse path. Knowledge-base poisoning plus downstream contamination. An attacker compromises or subtly manipulates a retrieval corpus — a guideline summary, an internal protocol document, a drug-interaction entry — and the CDS agent’s recommendations gradually shift in directions that benefit the attacker. The clinical fingerprint is distinctive: not a spike in anomalous behavior, but a slow statistical drift in recommendation distributions that looks like model evolution until someone correlates it with the corpus change.

Observe. Runtime AI-BOM captures the retrieval corpora the agent actually consults, the external references it actually fetches, and the write endpoints it actually touches. This is the ground truth for BAA scope — if a CDS agent is fetching from an external reference API that was not in the original deployment manifest because it was added under a vendor update, the runtime AI-BOM shows that call the first time it happens.

Posture. The gap here is not just permission scope; it is behavioral distribution. The Application Profile DNA records not only what the agent accesses but the distribution of what it recommends — the base rate of different recommendation classes, the tool invocation patterns for different clinical presentations, the drug-interaction lookup frequencies. This distribution is the posture artifact. Deviation from it is the signal.

Detect. CDS agents surface a detection pattern specific to write-path agents: recommendation distribution shift. Intent drift detection distinguishes legitimate model updates (correlated with deployment events) from unexplained behavioral drift (no corresponding deployment event) — and for CDS agents this distinction is the difference between a routine model refresh and a corpus-poisoning event. The detection output must connect recommendation drift to the specific corpus or tool whose signature changed, not just flag that “the agent is behaving differently.”

Enforce. CDS agents sit in the middle of the clinical availability spectrum. A recommendation engine can usually be paused without immediate patient harm — the ordering workflow falls back to clinician-only decision-making, which is the safe baseline. Hard kill is reasonable for CDS agents in most deployment contexts, unlike ambient scribes. The tighter constraint is that CDS agents are increasingly candidates for pre-enforced deployment (described in the next section), because the observation window for an agent whose outputs drive clinical orders is not operationally free.

Prior Authorization and Utilization Management Agent

What the agent accesses. Clinical documentation for the requested service, benefit policy documents (in-house or from the payer), prior auth criteria documents, optionally external medical-necessity reference sources, and write paths that issue approval or denial decisions with attached medical-necessity justification.

The specific misuse path. Scope creep via cross-patient context leakage. A prior auth agent processing many cases in rapid succession, with retrieval over a shared context store, can be manipulated — via prompt injection in a submitted clinical document, or via retrieval corpus poisoning — to surface content from other patients’ cases as justification for the current case. The output looks plausible. The underlying chain of custody does not hold.

Observe. Runtime AI-BOM covers the documentation retrieval surface, the criteria reference access, and the write path for decisions. The per-patient access log is the ground truth for disclosure accounting: every patient’s record touched during the processing of case X, captured at the record level.

Posture. The declared-versus-observed gap for prior auth agents often exposes permission sprawl — access to case types the agent has never processed, read access to benefit policies outside the agent’s actual domain, write scope to case fields the agent has never modified. The Posture artifact is the permission envelope that actually gets exercised, which typically collapses to a small fraction of the declared surface.

Enforce. Prior auth agents run batch-shaped workloads — processing queues of cases rather than live clinical sessions. That makes them easier to contain: killing a prior auth pod mid-queue means the current case needs reprocessing, but no clinical encounter is disrupted. Kill and Soft Quarantine are both reasonable default responses. The operational risk that does need tuning is denial-rate integrity: the enforcement policy has to block cross-patient context leakage without blocking legitimate case-specific retrieval, which is the per-agent behavioral boundary the observed Application Profile DNA defines.

Clinical Criticality Tiering: Track 1 and Track 2 for Healthcare Agents

The two-track enforcement model introduced for financial services — standard observe-to-enforce for most agents, pre-enforced deployment from staging-built baselines for agents whose production observation window carries regulatory cost — applies directly to healthcare, with one healthcare-specific addition: the track assignment also considers clinical criticality.

The three-question test for regulatory track assignment carries over unchanged. Would an unauthorized action by this agent trigger a regulatory notification timeline — for healthcare, the HIPAA Breach Notification Rule’s sixty-day clock under 45 CFR §164.410? Would the observation window generate audit evidence an OCR investigator or HITRUST assessor could question? Does the agent access data classes with mandatory breach reporting — for healthcare, PHI under HITECH §13402? Yes to any of these means Track 2.

In practice, this puts nearly every PHI-touching clinical agent — ambient scribes, CDS agents, prior auth automation, RAG over clinical notes — into Track 2. It puts non-PHI-touching agents (internal operational summarizers, scheduling optimization tools running against aggregated metrics, research agents working with de-identified datasets) into Track 1.

The healthcare-specific addition is a clinical criticality dimension that affects the enforcement response ladder rather than the track assignment itself. A live clinical agent — one whose containment action is itself a clinical availability event — defaults to Soft Quarantine as the immediate response, with escalation to Pause, Stop, or Kill routed through a documented clinical incident pathway rather than an automated SOC action. A batch clinical agent (prior auth queue processing, retrospective chart review) defaults to Kill or Stop, because there is no live clinician dependent on the agent’s presence.

This dimension does not exist in the financial services analog because no financial services agent has a clinician in the loop. Both verticals rely on the same underlying per-agent enforcement architecture; the response-mode defaults differ by vertical.

The practical implication: Track 2 deployment is the dominant pattern for clinical AI agents. Staging environments must represent production fidelity — same Kubernetes version, same cloud primitives, equivalent service accounts, the same FHIR gateway configuration, the same retrieval corpora (de-identified or synthetic where required), the same tool catalog. The parity validation step before production promotion is where the baseline becomes operationally usable. The methodology does not differ from the published progressive enforcement guide. The criteria for what counts as production parity have to account for FHIR resource type coverage, EHR integration patterns, and break-glass workflows.

Implementation Path: Observe Through Enforce in a Healthcare Environment

Knowing the framework and the track structure is one thing. Running the rollout inside a healthcare organization — with privacy-officer approval, Joint Commission survey windows, EHR release freezes, HITRUST assessment timelines, and clinical informatics review gates — is another. Here is the phased approach adapted to healthcare operational realities.

PhaseTimelineKey ActivitiesHealthcare-Specific Outputs
1: Discovery and ObservationWeeks 1–6Deploy eBPF sensors across AI workload clusters. Auto-discover agents, inference servers, and MCP tool runtimes. Generate runtime AI-BOM per agent. Establish Application Profile DNA for highest-PHI-exposure agents (scribes, CDS, prior auth). Privacy-officer review runs in parallel with technical deployment.Complete AI agent inventory including shadow workloads. Runtime AI-BOM joined against active BAA registry. Per-agent behavioral baseline v1. Latency validation against clinician-facing SLAs.
2: Posture and Track ClassificationWeeks 5–10Complete declared-versus-observed permission analysis per agent. Assign Track 1 or Track 2 per agent class against the three-question test. For Track 2 agents, stand up staging baselining infrastructure. Integrate with existing SIEM (typically Splunk, Microsoft Sentinel, or QRadar) and privacy office escalation path.Posture gap per agent. Track classification documented for each agent class. Staging parity validation plan for Track 2 agents. Cross-functional escalation runbook spanning security, privacy, compliance, legal, and clinical informatics.
3: Detection RolloutWeeks 8–14Configure AI-specific detection rules — prompt injection, agent escape, tool misuse, recommendation drift, write-path integrity anomalies. Tune per-agent baselines against observed production behavior. Route correlated attack stories to the SOC and the privacy office in parallel.AI-specific detection live per agent class. Attack story output tied to patient-record-level causality. Documented time-to-evidence for OCR inquiries, HITRUST assessments, and patient right-of-accounting requests under §164.528.
4: Progressive EnforcementWeeks 12–20Track 1 agents transition from observation to selective enforcement to full per-agent least privilege. Track 2 agents ship to production with pre-enforced policies derived from staging-validated baselines. Enforcement response modes tuned per clinical criticality tier.Per-agent enforcement live across the cluster. Documented response-mode defaults per agent type. Break-glass compatibility validated with clinical operations.

Two operational constraints bear flagging. First, privacy-officer approval authority over infrastructure-layer tooling that touches PHI is near-universal in provider organizations; technical-only rollouts routinely stall at this gate. Build the privacy officer into the rollout governance from week one. Second, Joint Commission surveys and EHR release freezes consume four to six weeks per year; the Gantt chart has to plan around them or the deployment timeline slips by a quarter.

Measuring What Matters in a Healthcare Rollout

Healthcare boards, compliance committees, and clinical leadership need different metrics than financial services risk committees. The ones that actually move decisions in this environment:

Shadow AI discovery completeness. How many AI agents touching PHI were discovered that the security team did not know about? The first week of observability typically justifies the program’s existence on this metric alone.

BAA scope drift events. How many times per month does an agent attempt to transmit to an endpoint outside the active BAA registry? This metric connects the security program directly to §164.504(e) compliance, which is the provision privacy officers ask about most.

Per-agent minimum-necessary gap. What percentage of declared permissions and FHIR resource types are actually exercised by each agent? The gap between declared and observed is the attack surface per-agent enforcement eliminates and the evidence that satisfies §164.502(b) attestation.

Time-to-evidence for the three high-stakes events. Time required to produce patient-record-level audit trails for an OCR breach inquiry, a HITRUST assessor’s minimum-necessary question, and a patient right-of-accounting request under §164.528. This is the metric that separates compliance readiness from compliance theater.

Track 2 deployment discipline. Percentage of PHI-touching agents deployed via Track 2 pre-enforced pathway versus standard observe-to-enforce. For clinical AI at scale, this should climb over time; a rising Track 2 share is the operational signal that the program is maturing.

Next Steps

The four-pillar framework does not change in healthcare. What changes is which agents go through which track, which response-mode defaults apply to which clinical criticality tier, and which regulatory outputs every phase of the rollout has to produce as a byproduct of the security work. None of this is about running a separate healthcare-specific security program. It is about running the framework with the dials tuned for 24/7 clinical operations, FHIR-mediated write paths, and HIPAA evidence demands that do not wait for the next quarterly review.

To see how ARMO produces per-agent runtime enforcement on Kubernetes — with Application Profile DNA at the Deployment level, eBPF-based enforcement at 1–2.5% CPU overhead, and attack-story correlation tied to specific patient-record access — book a demo or see the cloud-native security platform for AI workloads.

Frequently Asked Questions

Does eBPF-based observation add latency that clinicians will notice?

Kernel-level eBPF runs at roughly 1–2.5% CPU and 1% memory overhead per node, within the budget most platform teams already accept for observability. For ambient scribe workloads with clinician-facing latency requirements, validate the overhead against your specific agent SLAs in a proof-of-concept before production rollout. Clinical AI has hard latency floors that reject any instrumentation adding perceptible delay regardless of its security value.

What does clinical criticality tiering mean for the SOC runbook?

Live clinical agents — ambient scribes mid-encounter, CDS agents in the ordering workflow — default to Soft Quarantine as the immediate SOC response, with escalation to Pause, Stop, or Kill routed through a documented clinical incident pathway rather than an automated action. Batch clinical agents — prior auth queue processing, retrospective chart review — default to Kill or Stop at the kernel level. The tier assignment is documented per agent class and reviewed during the track classification phase.

How do model updates and prompt template changes compose with enforcement?

For Track 2 agents, model updates trigger re-baselining in staging before production promotion. The updated Application Profile DNA validates against production parity, then replaces the active enforcement policy through whatever change-control pathway the organization uses for EHR integrations. For Track 1 agents, continuous baseline refinement in production correlates drift events against recorded deployment events: a behavioral shift that correlates with a deployment is expected evolution; a shift without a correlating deployment is suspicious.

What happens to this methodology when the organization spans multiple cloud providers?

The framework methodology stays the same; each stage acquires a cross-cloud dimension. Cross-cloud agent discovery, cross-cloud permission chain assessment, cross-cloud signal correlation, and enforcement intent translated to cloud-specific primitives. The cross-cloud fragmentation argument and its four-principle cross-cloud architecture are covered in depth in the published multi-cloud framework piece.

Close

Your Cloud Security Advantage Starts Here

Webinars
Data Sheets
Surveys and more
Group 1410190284
Ben Hirschberg CTO & Co-Founder
Rotem_sec_exp_200
Rotem Refael VP R&D
Group 1410191140
Amit Schendel Security researcher
slack_logos Continue to Slack

Get the information you need directly from our experts!

new-messageContinue as a guest