Pressure-TestedResearch-Verified
Enterprise AgenticControl Architecture
A full reference model for production agentic systems — mapping mature standards, custom engineering, and the gaps the industry has not yet solved.
| Tier | Areas | Standards In Force |
|---|---|---|
| Mature | Software supply chain, workload identity, policy enforcement, durable workflows, runtime isolation, rate limiting, logging and traces. | SLSA v1.2 · SPIFFE · OPA/Cedar · Temporal · gVisor/Firecracker · OpenTelemetry · Istio/Envoy |
| Custom | Decision-to-provenance binding, context provenance at decision time, multi-agent dependency graphs, architecture-level versioning, compound risk reclassification. | No formal standard; built from mature primitives. |
| Unsolved | Model-weight provenance, semantic validation of model outputs, non-deterministic replay, universal multi-agent lineage, cross-system context attestation. | Active work: NIST CAISI · OWASP ASI · OpenTelemetry GenAI SIG. No settled standard. |
| Framework | Status & Relevance |
|---|---|
| NIST AI RMF + Agentic Profile | AI Agent Standards Initiative launched Feb 2026 via CAISI. Three pillars: security, interoperability, governance. SP 800-53 overlays (COSAiS) in development. Cyber AI Profile (IR 8596) maps CSF 2.0 to AI risks. Full agentic overlays expected Q4 2026. |
| OWASP Top 10 for Agentic Applications (2026) | Released Dec 2025. ASI01–ASI10: Goal Hijack, Tool Misuse, Identity & Privilege Abuse, Supply Chain, Unexpected Code Execution, Memory & Context Poisoning, Inter-Agent Comms, Cascading Failures, Human-Agent Trust Exploitation, Rogue Agents. Introduces least agency. |
| Singapore MGF for Agentic AI | Released Jan 2026 (Davos). First global governance framework specifically for agentic AI. Four dimensions: risk bounding · human accountability · technical controls · end-user responsibility. Voluntary but influential. |
| EU AI Act | High-risk obligations enforceable 2 August 2026. Articles 9 (risk mgmt), 13 (transparency), 14 (human oversight), 16 (provider obligations). Penalties up to €35M or 7% global turnover. |
| OpenTelemetry GenAI Conventions | Still in Development status. Agent spans, tool execution spans, evaluation events defined but not yet stable. |
| SLSA v1.2 | Build track stable. Source track in development. Covers code & artifacts only. No equivalent exists for model weights. |
| D3 | External / regulated action — max controls. Writes external system · changes legal/financial state · regulated data across jurisdictions · grants privileged access · exceeds aggregate impact threshold. |
| D2 | Bounded internal action. Changes internal state or triggers bounded internal effects; cannot independently create external legal, financial, or customer-facing consequences. |
| D1 | Read-only recommendations. |
| D0 | Informational or observational operations. |
| Field | Requirement | Purpose |
|---|---|---|
| Source ID | Document or system identifier | Trace origin |
| Chunk hash | SHA-256 of retrieved content | Detect tampering since retrieval |
| Retrieval time | Timestamp | Reason about freshness |
| Trust class | Canonical · transformed · external · agent-generated · user-supplied | Let policy weight context quality |
| Transformation ref | ID if content was summarised or translated | Inspect derived context |
| Session scope | Workflow or case ID | Prevent cross-case contamination |
| OWASP Risk | Primary Control Planes | Residual Risk |
|---|---|---|
| ASI01 · Agent Goal Hijack | Context Integrity · Output Validation · Behaviour Bounding | Sophisticated indirect injection may evade plausibility checks |
| ASI02 · Tool Misuse | Tool Execution · Policy · Output Validation | Novel tool abuse patterns require updated detection rules |
| ASI03 · Identity & Privilege Abuse | Identity & Trust · Multi-Agent Delegation · Key Lifecycle | Insider threat with legitimate credentials |
| ASI04 · Supply Chain Vulnerabilities | Build Integrity · Model Integrity · Deployment Admission | No model-weight provenance standard |
| ASI05 · Unexpected Code Execution | Tool Execution (sandboxing) · Workflow (no ad hoc) | Zero-day escape from isolation boundary |
| ASI06 · Memory & Context Poisoning | Context Integrity · Context Absence Detection | Slow poisoning below detection threshold |
| ASI07 · Insecure Inter-Agent Comms | Identity (mTLS) · Multi-Agent Delegation (scoped tokens) | Token theft via side-channel |
| ASI08 · Cascading Failures | Behaviour Bounding · Governance Feedback · Circuit Breakers | Novel cascade pattern not covered by current thresholds |
| ASI09 · Human-Agent Trust Exploitation | Human Approval Architecture · Evidence presentation | Automation bias despite controls |
| ASI10 · Rogue Agents | Detection & Response · Quarantine · Kill Switches | Sophisticated evasion of behavioural monitoring |
| Gap | Current State & Mitigation |
|---|---|
| No standard for model-weight provenance | NIST CAISI developing guidance. Mitigation: record model path per decision, periodic eval benchmarks, contractual notification from providers. |
| No standard for decision-to-provenance binding | Custom implementation linking runtime decisions to build attestations via decision packages. No formal specification exists. |
| No standard for multi-agent decision lineage | Dependency graph tracking is custom engineering. NIST planning multi-agent overlays for SP 800-53 (late 2026). |
| No deterministic replay for LLMs | Reframed as decision context reconstruction plus policy re-evaluation. Exact model output reproduction is not claimed. |
| No universal context integrity attestation | Minimum viable provenance model implemented; full cross-system attestation standard does not exist. |
| Semantic validation is partially solvable | Layered plausibility checks reduce but do not eliminate risk. Backstop: human review for high-consequence decisions; anomaly detection. |
| OpenTelemetry GenAI not yet stable | Agent spans, tool execution spans, evaluation events defined but in Development status. Production should use opt-in stability flags. |