Skip to content
Orion Intelligence Agency logo
ORION
INTELLIGENCE AGENCY
← Back to Insights

SOC 2 AI Controls — What Auditors Actually Require

Sat Feb 21 2026

SOC 2 AI controls require runtime enforcement evidence — not policy documents. Map the four-layer governance stack to audit-ready compliance.

SOC 2 auditors are no longer satisfied with policy documents and governance committee meeting minutes. As AI workloads move into production — executing transactions, accessing sensitive data, invoking external tools — the Trust Services Criteria now require evidence that governance is enforced at runtime, not merely documented in a binder.

The expansion is structural. Traditional SOC 2 controls assume human-initiated actions with predictable blast radii. Autonomous AI systems violate that assumption. An agentic workflow can execute dozens of state mutations before a human is notified. The audit question has shifted from "do you have a governance policy?" to "show me the enforcement evidence."

Why SOC 2 Is Expanding Into AI

The AICPA Trust Services Criteria updates reflect a fundamental recognition: AI workloads introduce control risks that existing SOC 2 frameworks were not designed to address. Traditional controls govern human-operated systems where actions are discrete, reviewable, and reversible within known time windows. AI systems — particularly agentic architectures — operate continuously, make autonomous decisions, and compound errors across sessions rather than transactions.

Three factors drove the expansion. First, autonomous execution introduces non-deterministic behavior that static controls cannot constrain. Second, the blast radius of AI failures compounds before detection — a misconfigured tool invocation at 2 AM can propagate through downstream systems before any human reviews the output. Third, regulatory pressure from the EU AI Act and emerging US guidance has created audit expectations that SOC 2 frameworks must accommodate or risk irrelevance.

The result: auditors now assess AI-specific control areas that go beyond traditional IT general controls. Organizations running AI in production without these controls face audit findings, qualified opinions, and — in regulated industries — material compliance risk.

The Five AI-Specific Control Areas Auditors Assess

SOC 2 AI assessments evaluate five distinct control areas, each requiring evidence that extends beyond what traditional monitoring provides.

Risk assessment for AI-specific failure modes. Auditors verify that the organization has identified and classified AI-specific risks: hallucination, tool parameter misuse, policy citation mismatch, escalation loops, prompt injection, and behavioral drift. Each failure mode must map to a specific control with documented enforcement evidence.

Model monitoring and behavioral drift detection. Standard application performance monitoring (APM) tracks latency, throughput, and error rates. Auditors now require evidence of behavioral monitoring — evaluation pass-rate trends, guardrail trigger frequencies, response quality degradation, and tool invocation pattern deviation. The distinction matters: infrastructure metrics confirm the system is running; behavioral metrics confirm the system is governed.

Automated control testing with enforcement evidence. Manual control testing at quarterly intervals is insufficient for systems that execute thousands of actions daily. Auditors assess whether controls are tested continuously and whether test results produce enforcement actions — not just alerts. A control that detects drift but takes no automated enforcement action fails the evidence standard.

Governance enforcement documentation. This is where most organizations fail. Policy documents describe intent. Auditors require evidence of enforcement — proof that governance was evaluated before state mutation, that authority was verified, and that the evaluation result was recorded in a tamper-evident format. The gap between "we have a policy" and "we can prove enforcement" is where audit findings originate.

Incident response procedures for autonomous system failures. Traditional incident response assumes human-initiated actions. AI incident response must account for autonomous execution, compounding blast radius, and the need for containment before diagnosis. Auditors verify that severity classification exists (P1 through P4), that containment procedures are documented and tested, and that evidence is captured during containment — not after.

What Constitutes Audit-Ready Evidence

The distinction between logs and receipts is the most consequential technical detail in SOC 2 AI compliance.

Logs are telemetry. They record that something happened — a request was made, a response was generated, a tool was invoked. Log aggregation platforms collect, index, and query these records. Logs answer the question: "what occurred?"

Receipts are enforcement artifacts. They prove that governance was evaluated and authority was granted before a state change occurred. A receipt contains the identity of the requester, the policy evaluated, the authority determination, the timestamp, and a cryptographic attestation that the record has not been modified. Receipts answer the question: "who authorized this, under what policy, and can you prove it?"

Auditors require the second. Log aggregation — regardless of volume, indexing sophistication, or retention period — does not satisfy the evidence standard for governance enforcement. Logs can be modified, lack non-repudiation, and do not prove that governance was evaluated before execution. They prove observation. They do not prove enforcement.

Immutable, append-only receipt ledgers satisfy the evidence standard. Each state-mutating action produces a receipt before execution proceeds. The receipt ledger is tamper-evident. The attestation chain is verifiable. This is the audit evidence that withstands scrutiny — not because it is more data, but because it is structurally different data.

Mapping the Four-Layer Enforcement Stack to SOC 2 AI

The runtime enforcement stack maps directly to SOC 2 AI control requirements. Each layer addresses a specific audit evidence category.

Authority Gate maps to risk treatment controls. Every AI execution path is gated by authority evaluation. The default state is denial — fail-closed. Tool invocations are constrained by allow-lists. Pre-execution attestation ensures that no state-mutating action proceeds without verified authority. This layer produces the evidence auditors require for governance enforcement documentation. (See: AI Governance Consulting: A Practical Framework for Runtime Enforcement — /insights/ai-governance-consulting)

Immutable Receipts map to evidence capture. Every state change is cryptographically attested in an append-only ledger. The receipt satisfies non-repudiation requirements. Auditors can verify who authorized what, when, under what policy. This layer transforms telemetry into proof.

Drift Guard maps to monitoring and alerting controls. Evaluation pass rates, guardrail trigger frequencies, and escalation metrics are tracked over rolling windows. Enforcement thresholds trigger automated actions — deployment freeze, governance lead escalation, workflow quarantine. This layer provides the continuous monitoring evidence that quarterly manual testing cannot.

Gated Substrate maps to access and isolation controls. AI workloads are isolated at the infrastructure level. Tool boundaries are enforced by capability removal, not restriction. Network segmentation prevents unauthorized resource access. This layer ensures that even if an AI workflow attempts to exceed its authority, the substrate physically prevents execution.

Common Gaps That Trigger Audit Findings

Four gaps account for the majority of SOC 2 AI audit findings.

Missing behavioral drift detection. Organizations monitor infrastructure health but not behavioral health. Evaluation pass rates decline over weeks. Guardrail trigger frequencies increase gradually. Response quality degrades below thresholds. Without behavioral monitoring, these signals are invisible until a customer-facing incident forces detection. Auditors flag the absence of behavioral monitoring as a control gap.

Absent pre-execution attestation. AI workflows execute state-mutating actions without prior authority verification. The system acts, then logs the action. There is no evidence that governance was evaluated before mutation. This is the most common finding and the most consequential — it means the organization cannot prove that any specific AI action was authorized.

Ungoverned tool access. Agentic systems invoke external tools — APIs, databases, file systems, communication channels — without allow-list constraints. Any tool accessible to the runtime is accessible to the AI. Auditors assess whether tool boundaries are enforced by the governance layer, not merely documented in policy. The gap between documented tool restrictions and enforced tool restrictions is where findings concentrate.

Inadequate escalation documentation. When an AI system fails, the incident response is ad hoc. No severity classification. No documented containment procedure. No evidence capture during the incident. Auditors require that AI incident response be documented, tested, and classified by severity with specific response procedures for each level. (See: ISO 42001 vs NIST AI RMF — /insights/iso-42001-vs-nist-ai-rmf for framework-specific requirements.)

How to Prepare: 30/60/90-Day Compliance Hardening

SOC 2 AI compliance hardening follows a phased approach that mirrors the enforcement stack deployment.

Days 1–30: Inventory and mapping. Identify every AI workflow in production or pre-production. For each workflow, document: what data it accesses, what tools it invokes, what state mutations it performs, what authority model governs it (if any), and what evidence is captured. Map existing controls to the five SOC 2 AI control areas. The output is a control-plane gap map — a precise record of what is governed, what is monitored, and what is uncontrolled.

Days 31–60: Enforcement gate deployment. For the highest-risk workflows identified in Phase 1, deploy: authority gates with fail-closed defaults, receipt capture for all state-mutating actions, behavioral drift monitoring with enforcement thresholds, and tool boundary constraints. The goal is not full-stack deployment across every workflow — it is audit-ready enforcement evidence for the workflows with the largest blast radius.

Days 61–90: Mock audit and remediation. Run a mock audit cycle. Review all enforcement evidence against the five control areas. Identify remaining gaps. Close them. Validate that every control area produces artifacts that satisfy the evidence standard: receipts (not just logs), enforcement actions (not just alerts), and documentation that proves governance was evaluated before execution.

Organizations that complete this arc produce audit-ready evidence for SOC 2 AI controls. Organizations that skip Phase 2 — deploying governance enforcement — produce policy documents that auditors will flag as insufficient.

When to Run a Readiness Scan

A Readiness Scan is a 30-minute, artifact-backed assessment that baselines your SOC 2 AI governance posture before the auditor does. The Scan maps your current state against the five SOC 2 AI control areas, identifies enforcement gaps, and produces a prioritized remediation roadmap.

Deliverables: control-plane gap map, failure-mode heatmap, evidence checklist, and a 30/60/90 hardening plan. The Readiness Scan is the starting point for organizations that need audit-ready governance enforcement — not another governance document. (See: AI Governance Audit Checklist — /insights/ai-governance-audit-checklist for the full evidence requirements.)

Schedule a Readiness Scan at /readiness-scan — baseline your posture before the auditor does.

Need help designing or deploying this?

Ready to map your governance roadmap?