Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.sigilcore.com/llms.txt

Use this file to discover all available pages before exploring further.

Vernon Wharff
Sigil Core
April 2026 — v1.0
Download the PDF version here [ ↗ ]

Summary

AI governance standards now tell organizations how to define policy, allocate responsibility, assess risk, monitor behavior, and document outcomes. Agentic systems expose the remaining gap. Documentation can tell an agent what should happen. Evaluation can estimate what an agent might do. Monitoring can record what an agent did. Pre-execution enforcement controls what an agent can do. This paper names that gap the Pre-Execution Enforcement Gap and argues that high-stakes agent actions should pass through deterministic policy evaluation before they reach tools, credentials, wallets, APIs, or infrastructure. The Sigil Open Framework (SOF) proposes an execution control layer for that gap, using signed Sigil Warrants, Intent Attestations, holds, denials, and audit evidence to bind human intent to machine action.

The Pre-Execution Enforcement Gap

AI governance has entered its standards era. ISO, NIST, Berkeley, the EU, and the frontier safety community now agree on the basic shape of responsible AI governance. Organizations must define policy, assign responsibility, map foreseeable harms, measure trustworthiness, treat risk, monitor results, document decisions, and report enough information for affected parties, auditors, regulators, and downstream operators to understand what happened. That consensus matters. ISO/IEC 42001:2023 gives organizations the management-system frame. Its Annex A controls cover AI policy, roles, resources, impact assessment, lifecycle requirements, data management, information for interested parties, responsible use, and third-party relationships. NIST AI RMF 1.0 gives organizations the operating vocabulary. ISO/IEC 23894 gives organizations AI-specific risk-management guidance. NIST AI 600-1 extends the AI RMF into generative AI. Berkeley’s General-Purpose AI Risk-Management Standards Profile v1.2 adapts those standards for GPAI models, including cutting-edge large language models and agentic models, then maps the resulting guidance back to ISO, NIST, the EU AI Act, the EU GPAI Code of Practice, the Hiroshima Process, and frontier safety commitments. Berkeley’s companion Agentic AI Risk-Management Standards Profile focuses the same standards structure on systems granted autonomy, tools, environment access, and authority to act on behalf of users. The landscape now points in one direction. AI governance has moved from principles to procedures. The best current standards no longer treat safety as a statement of values. They ask for risk thresholds, impact characterization, go/no-go decisions, red-teaming, third-party evaluation, documentation, incident response, post-deployment monitoring, and reporting. Berkeley’s GPAI Profile v1.2 makes that shift explicit by adding high-priority emphasis on external feedback, third-party evaluation, and post-deployment monitoring. Its v1.2 update expands the risk landscape around manipulation, deception, sandbagging, situational awareness, socioeconomic and labor-market disruption, and the difficulty of removing backdoors. The Profile repeatedly treats agentic models and downstream agentic wrappers as part of the GPAI risk-management landscape, while the companion Agentic AI Profile addresses the additional risks that emerge when AI systems receive agency to act. This is where the standards landscape exposes its central gap. Most AI governance controls operate before deployment, around deployment, or after deployment. They create policy, evaluate models, review risks, document decisions, audit behavior, and respond to incidents. These controls reduce organizational blindness. They make risk legible. They help leaders decide whether a system should ship, where it should operate, and how teams should monitor it. They do not, by themselves, stop an autonomous system at the instant it attempts a high-stakes action. This paper names that gap the Pre-Execution Enforcement Gap. The Pre-Execution Enforcement Gap is the distance between documented AI governance and enforced runtime authorization at the moment an autonomous system attempts to act. It appears whenever a policy says an agent may not do something, but the execution path still gives the agent technical ability to do it. It appears when a risk register names a hazard, but the API, wallet, shell, database, or cloud credential accepts the call anyway. It appears when a human approval threshold exists in a procedure, but the payment rail, deployment pipeline, or signing endpoint has no cryptographic proof that approval happened. The gap grows with autonomy. A chatbot that drafts text creates information risk. An agent that writes files, calls APIs, sends email, signs transactions, rotates infrastructure, or moves money creates execution risk. Execution risk needs a different control family. Documentation can tell an agent what should happen. Evaluation can estimate what an agent might do. Monitoring can record what an agent did. Pre-execution enforcement controls what an agent can do.

Berkeley, NIST, and Human Control

The citation discipline matters. Govern, Map, Measure, and Manage are NIST AI RMF functions. Section labels such as Govern 2.1, Map 3.5, Manage 2.4, and Manage 4.1 originate in NIST AI RMF 1.0. Berkeley’s GPAI Profile v1.2 uses that structure and adds profile-specific guidance for GPAI models. Berkeley’s companion Agentic AI Profile uses the same structure for agentic systems and adds agent-specific control guidance. A defensible mapping cites the base NIST section when using NIST language, Berkeley’s GPAI Profile v1.2 when using current GPAI profile guidance, and the Agentic AI Profile when the claim concerns agent-specific execution risk. NIST AI RMF 1.0 describes the base human-control structure. Govern 2.1 concerns documented roles, responsibilities, and lines of communication for mapping, measuring, and managing AI risks. Map 3.5 concerns defined, assessed, and documented human oversight processes. Manage 2.4 concerns mechanisms to supersede, disengage, or deactivate AI systems that produce performance or outcomes inconsistent with intended use. Manage 4.1 concerns post-deployment monitoring, including input capture, appeal and override, decommissioning, incident response, recovery, and change management. Berkeley’s GPAI Profile v1.2 makes those NIST sections more operational for GPAI developers and deployers. Under Govern 2.1, it emphasizes clear allocation of responsibility across the GPAI value chain, risk oversight, risk ownership, risk support and monitoring, and assurance. Under Map 3.5, it calls for identifying AI system features and capabilities that require human oversight, defining oversight practices, evaluating those practices, and documenting roles, responsibilities, delegations of authority, skills, training, resources, and domain knowledge. Under Manage 2.4, it emphasizes structured access, staged release, emergency shutdown procedures, tiered control mechanisms, usage-guideline enforcement, and criteria for deactivation or disengagement. Under Manage 4.1, it emphasizes post-deployment monitoring, external feedback, serious incident reporting, vulnerability discovery, misuse monitoring, response, recovery, decommissioning, and continuous monitoring. Berkeley’s Agentic AI Profile carries that logic into the execution environment. It emphasizes human-agentic AI management hierarchies that preserve human authority, role definitions, intervention points, escalation pathways, shutdown mechanisms, oversight checkpoints, role-based permission management, explicit permission declarations, real-time permission requests for high-risk actions, activity logs, and records of tool use, resource access, and permission changes. Those additions point directly toward a runtime enforcement layer. That is the key observation. Berkeley’s GPAI Profile v1.2 and Agentic AI Profile name the right risk-management levers for human control and accountability. The levers include role allocation, human oversight, structured access, staged release, permission boundaries, intervention points, tiered control, deactivation procedures, monitoring, feedback, incident response, and assurance. Those levers become operationally meaningful only when the execution path can enforce them. A role definition matters at runtime only if the system can distinguish the policy owner, the agent, the approving human, and the credential authority. An intervention point matters only if the agent cannot proceed around it. An escalation pathway matters only if the risky action enters a hold state before execution. A deactivation procedure matters only if the system can restrict, suspend, or cut off the capability being used. An accountability record matters only if it binds the action to the policy, approval state, tool, credential release, and time of decision.

The Sigil Open Framework Position

Pre-execution enforcement starts with a simple rule. An autonomous system should not reach a high-stakes target unless it carries verified proof that the action complies with an authorized policy. The target may be a blockchain gateway, cloud API, email system, filesystem, procurement workflow, database, model tool, or credential broker. The control should not depend on the agent’s self-restraint, a prompt, or a dashboard that someone reads tomorrow. It should bind policy, intent, approval, and execution into the same path. AI governance therefore needs an execution layer. The Sigil Open Framework (SOF) supplies that layer. It does not replace ISO/IEC 42001, NIST AI RMF, ISO/IEC 23894, Berkeley’s GPAI Profile v1.2, Berkeley’s Agentic AI Profile, or the EU AI Act. It turns a specific class of their requirements into enforceable infrastructure. SOF gives organizations a way to express policy as Sigil Warrant, sign that policy, evaluate every declared intent against it, issue short-lived Intent Attestations for approved actions, hold threshold actions for human approval, and reject execution when proof fails. The Open Execution Engine (OEE) is the implementation layer for that enforcement architecture. It packages SOF enforcement into domain-specific verticals that share the same pattern. The agent proposes an intent, Sigil Sign evaluates the signed warranty.md policy, the system issues an Ed25519 Intent Attestation when the action complies, and the execution engine proceeds only after attestation. The first OEE vertical is Open Venture Engine (OVE), a venture-capital stack that applies the same enforcement primitives to fund and treasury workflows. This changes the compliance posture. A policy document says what leadership approved. A signed Sigil Warrant proves which policy governed the action. An Intent Attestation proves that the action passed that policy before execution. A gateway, hook, or credential broker rejects the action when that proof does not match. A violation log shows the denial, hold, approval, rule, and time. The audit record no longer starts after the event. It starts at the decision point. This is also where accountability becomes more than attribution. After-the-fact attribution can tell a reviewer which agent, user, or system caused an event. Pre-execution accountability tells the reviewer whether the action had authority before it occurred. In SOF, the evidence chain can show the approved policy, the declared intent, the matched rule, the decision, the human hold outcome where needed, the attestation, the gateway check, and the credential release. That chain aligns with Berkeley’s emphasis on role clarity, human oversight, structured access, permission boundaries, monitoring, incident response, and deactivation. It also supplies the kind of evidence an ISO/IEC 42001 audit will test when management-system language reaches operational claims.

SOF to ISO/IEC 42001 Annex A

The SOF to ISO/IEC 42001 mapping is strongest where Annex A asks for proof that policy reached operation. The table below does not claim that SOF satisfies the whole AIMS. It identifies controls where SOF can serve as a primary evidence surface for the pre-execution enforcement portion of the control.
ControlISO/IEC 42001 control surfaceSOF primitiveEvidence surface
A.2.2 AI policyAI policy documented for development or use of AI systemsSigil Warrant and signed warranty.mdOperator-defined execution policy, Ed25519 signature, policyHash in attestations
A.4.4 Tooling resourcesTooling resources documented for the AI systemAgent Hooks and framework registrySupported framework identifier, governed action class, tool-call policy block
A.6.2.5 AI system deploymentDeployment plan and requirements before deploymentOEE authorization path and gateway checksSigned policy, verification key, /v1/authorize, write rejection without valid proof
A.6.2.7 Technical documentationTechnical documentation for relevant partiesSOF component docs and attestation specificationPolicy format, pipeline docs, attestation rules, integration references
A.6.2.8 Event logsEvent logs while AI system is in useCommand, OEE decision records, Vault audit recordsApproval, denial, hold, matched rule, timestamp, attestation, credential-release event
A.9.2 Responsible use processesProcesses for responsible use of AI systemsOEE enforcement pipelineIntent declaration, policy evaluation, approval, denial, hold, gated execution
A.9.4 Intended useAI system used according to intended use and documentationWarrant rules, gateway rejection, Vault checksAllowlists, blocklists, thresholds, custom deny rules, proof-gated access
Partial mappings matter too. ISO/IEC 42001 contains controls that SOF can support without satisfying completely. A defensible statement of applicability should include these with scope language, so the organization does not over-claim.
ControlWhat SOF supportsWhat remains outside SOF
A.2.3 Policy alignmentExecution rules can encode security, privacy, finance, and acceptable-use requirementsThe policy crosswalk, approval process, and ownership model
A.2.4 AI policy reviewPolicy versions and policyHash changes show when new rules entered operationReview cadence, management approval, and suitability analysis
A.3.2 Roles and responsibilitiesOperator signature, tenant scope, hold resolver, Vault trust anchorsRole definitions, competence records, staffing, governance authority
A.3.3 Reporting of concernsDenials and holds can feed concern and incident workflowsConfidential reporting, escalation, response, reprisal protection
A.4.2 Resource documentationEnforcement resources, frameworks, routes, backends, policy filesModel resources, data stores, full AI system inventory
A.6.2.4 Verification and validationApproval, denial, hold, expired-attestation, and gateway-rejection testsModel validation, performance, robustness, bias, human factors, TEVV
A.6.2.6 Operation and monitoringCommand events, hold queue, Vault audit recordsModel quality, drift, task success, user impact, support processes
A.8.2 User informationOperator and technical-user documentation for governed actionsEnd-user notice, benefit and harm disclosures, override instructions
A.8.5 Information for interested partiesExportable policy, attestation, denial, hold, and audit evidenceLegal reporting obligations, disclosure scope, deadlines
A.9.3 Responsible use objectivesRuntime constraints for security, safety, spending, access, and human approvalFairness, accessibility, explainability, customer expectations
A.10.2 Responsibility allocationPolicy signer, API tenancy, framework identity, Vault routesContracts, RACI matrices, privacy-role analysis, customer terms
A.10.3 SuppliersSupplier boundary enforcement through framework, tool, route, and domain limitsProcurement diligence, supplier security review, data rights, service levels
The mapping has limits. Sigil does not perform impact assessments. It does not judge fairness, labor-market effects, data-set quality, model interpretability, human competence, customer expectations, or societal impact. It does not choose an organization’s risk tolerance. It cannot replace legal counsel, internal audit, model evaluation, red-teaming, privacy review, or third-party due diligence. SOF should not be mapped as satisfying ISO/IEC 42001 A.4.3, A.5.2 through A.5.5, A.7.2 through A.7.6, A.8.3, A.8.4, or A.10.4 without separate organizational controls. Sigil makes one part of the control environment harder to fake, the part where an agent tries to act.

SOF to NIST AI RMF 1.0

NIST AI RMF 1.0 is voluntary and flexible. It does not prescribe a single implementation architecture. SOF maps to a subset of NIST subcategories where runtime enforcement can produce direct evidence.
NIST subcategoryNIST control surfaceSOF primitiveEvidence surface
Govern 1.3Processes determine needed risk-management activity based on risk toleranceSigil WarrantSigned thresholds, allowlists, blocklists, approval gates
Govern 1.4Risk-management process and outcomes established through policies and controlsWarrant plus CommandSigned policy tied to approval, denial, and hold records
Map 3.5Human oversight processes defined, assessed, and documentedApproval thresholds and consensus holdsHold creation, resolver action, outcome, timestamp, matched rule
Manage 2.4Mechanisms supersede, disengage, or deactivate systems outside intended useOEE denial, gateway rejection, Vault hard rejectionNo attestation issued, write blocked, credential release refused
Manage 4.1Post-deployment monitoring, appeal, override, decommissioning, incident response, recovery, change managementCommand and Vault audit recordsDenial and hold history, credential-release events, audit-chain verification
These primary NIST mappings apply to governed action paths. They do not by themselves satisfy full model decommissioning, enterprise incident response, or organization-wide risk management. SOF also supports several NIST subcategories partially.
NIST subcategoryWhat SOF supportsWhat remains outside SOF
Govern 2.1Operator key, tenant scope, hold resolver, framework identityRole descriptions, training, competence, communication lines
Govern 5.1Verifiable decisions that outside evaluators can inspectFeedback intake, prioritization, stakeholder engagement
Map 1.1Intended use expressed as enforceable action scopeContext analysis, laws, norms, user expectations
Map 5.2Runtime evidence for feedback about impacts from denied or held actionsEngagement with affected actors and broader impact analysis
Measure 2.7Security and resilience of the enforcement pathSecurity assessment, adversarial testing, vulnerability management
Manage 1.3Risk treatments expressed as allow, deny, and hold rulesRisk selection, treatment design, residual-risk acceptance
Measure 2.7 and Manage 1.3 remain partial mappings because SOF supplies runtime evidence and enforceable treatments for governed action paths, while the organization must still complete the broader security evaluation and risk-treatment program. Several NIST subcategories remain outside SOF. Map 4.1 requires legal risk mapping, including intellectual property. Measure 2.10 requires privacy risk examination. Measure 2.11 requires fairness and bias evaluation. Measure 2.12 requires environmental-impact assessment. Govern 4 concerns organizational culture around trustworthy AI. SOF can supply evidence into those programs only when execution records become relevant.

Threat Model Against NIST Trustworthy AI

NIST AI RMF 1.0 describes trustworthy AI through characteristics that include valid and reliable, safe, secure and resilient, accountable and transparent, explainable and interpretable, privacy-enhanced, and fair with harmful bias managed. Pre-execution enforcement maps strongly to three of these, partially to one, and weakly or not at all to the others. Safe. NIST describes safe systems as those that do not, under defined conditions, endanger human life, health, property, or the environment. It also identifies real-time monitoring and the ability to shut down, modify, or have human intervention in systems that deviate from intended or expected functionality. Pre-execution enforcement supports this safety surface by checking action declarations before they reach the target. It addresses unauthorized financial transfers, unauthorized cloud or infrastructure mutations, destructive shell commands, and actions induced by prompt injection that would otherwise execute through an available tool. It does not make the model safe in the broader sense. It makes unsafe action pathways harder to exercise. Secure and resilient. NIST frames security around protection against unauthorized access and use, adversarial examples, data poisoning, endpoint exfiltration, and protocols to avoid, protect against, respond to, or recover from attacks. SOF narrows the agent’s effective attack surface by separating model reasoning from credential authority. Agent Hooks intercepts tool calls before execution. Vault refuses credential release without valid proof and has no fail-open path. OEE denies intents that violate signed policy. This helps defend against credential exfiltration through prompt injection, model-driven tool abuse, and confused-deputy patterns where an agent tries to combine available capabilities into an unauthorized external effect. Accountable and transparent. NIST states that accountability presupposes transparency and that transparency should cover system decisions, deployment decisions, and who made them. SOF contributes by creating an evidence chain at the point of decision. The record can show the policy, policy hash, declared intent, matched rule, decision, hold outcome, attestation, gateway check, and credential-release event. This is stronger than ordinary execution logging because denied actions enter the record too. The absence of a completed transaction no longer erases the attempted violation. Valid and reliable. SOF contributes partially. The enforcement layer is deterministic, so it can add a reliable boundary underneath stochastic model behavior. It can continue to deny out-of-policy actions even when the model is wrong. It does not validate the model’s performance, accuracy, robustness, or task suitability. A model that repeatedly asks for denied actions may still be unreliable, even if SOF prevents the worst consequences. Explainable and interpretable. SOF can explain which rule matched an action. It cannot explain why the model attempted the action. Model interpretability remains upstream. Privacy-enhanced. Vault can prevent unauthorized credential release that could enable privacy violations. SOF does not perform differential privacy, data minimization, training-data privacy review, or privacy impact assessment. Fair with harmful bias managed. SOF does not detect or mitigate model bias. A fairness checker could become an upstream condition that feeds a Sigil Warrant rule, but SOF would enforce the result. It would not generate the fairness judgment. The threat model defines the claim. Pre-execution enforcement is not a general solution to AI risk. It is a control family for agentic execution risk, especially where safety, security, accountability, and transparency depend on whether a high-stakes action was authorized before it reached the target system.

Worked Example: Pre-Execution Enforcement in a Production Agent Runtime

The architectural argument is simple. The path between an agent’s reasoning and an external system effect must pass through deterministic policy evaluation before the action fires. In abstract, that is a claim about a control family. In practice, it becomes a sequence of code paths, network calls, signatures, and logs that either complete or fail. This section walks through one such sequence. It uses Claude Code or the Anthropic SDK as the agent runtime, @sigilcore/agent-hooks as the in-runtime enforcement client, Sigil Sign as the authorization service, OEE as the policy evaluation path, and Sigil Command as the operator evidence surface. The example follows the published Agent Hooks and SOF documentation. The identifiers, keys, hashes, and signatures below are illustrative, but the control flow is the SOF control flow. The scenario is intentionally narrow. An autonomous coding agent attempts to execute a destructive shell command against the operator’s filesystem, and pre-execution enforcement intercepts it. The same pattern applies to wallet transactions, cloud API mutations, email sends, deployment triggers, credential rotations, database writes, and other high-stakes action classes because the enforcement primitive is action-class agnostic.

1. The operator-defined policy

Before any governed agent runs, the operator authors a Sigil Warrant in warranty.md and signs it with the operator’s Ed25519 key. The signed policy is the authority the enforcement layer recognizes.
version: 1.0.0

## tool_calls
allowed: bash, web_fetch, file_write, wallet_sign, email.send
bash.blocked_commands: rm -rf, sudo, curl | sh
web_fetch.blocked_domains: evil.com, malicious.io
file_write.blocked_paths: ~/.ssh, .env, secrets
email.require_approval: true

## custom
deny_string: "OPENAI_API_KEY"
deny_string: "DROP TABLE"

## signature
sigil-sig: <base64url-ed25519-signature>
The warrant expresses operator intent as a small, reviewable artifact. A human can read it. A machine can evaluate it. The signature binds the policy to the operator identity. The policy hash later binds an authorization decision to the exact policy version in force at evaluation time.

2. The agent attempts an action

A developer runs an autonomous coding agent in a repository. The session begins as an ordinary refactor. The model decides to clean a build directory and proposes a tool call:
rm -rf ./build
In a runtime without pre-execution enforcement, the bash tool receives this command and executes it. If the working directory is wrong, if the model misjudges scope, or if a prompt injection in a downstream file changes the agent’s plan, the filesystem changes before governance has a chance to intervene. A log may exist after the fact, but the deletion has already happened. In a runtime with pre-execution enforcement, the tool call enters the authorization path before the bash runtime receives it.

3. The PreToolUse hook fires

The Anthropic PreToolUse event fires before the tool executes. The @sigilcore/agent-hooks adapter observes the proposed tool call and constructs a declared intent from the raw tool input. This is important. The authorization service receives what the runtime is about to do, not a model-written summary of what the agent claims it intends.
{
  "intent_id": "intent-7c2e1f4a",
  "session_id": "session-9b3a",
  "warrant_id": "warrant-prod-coding-agent-2026-04-26",
  "action_class": "tool_calls.bash",
  "action_payload": {
    "tool": "Bash",
    "command": "rm -rf ./build",
    "cwd": "/Users/operator/project"
  },
  "timestamp": "2026-04-26T14:32:18.114Z"
}
The hook submits that intent to Sigil Sign through the authorization path. The call asks a narrow question: does this specific action fit the signed policy that governs this session?

4. Sigil Sign evaluates against the signed warrant

Sigil Sign retrieves the referenced warrant, verifies the operator signature, computes or checks the policy hash, and evaluates the declared intent against the ## tool_calls policy. The decision is deterministic for this command.
  1. The evaluator checks blocked bash commands.
  2. The command contains rm -rf.
  3. The deny rule matches.
  4. Sigil returns a denied decision with the matched rule and policy evidence.
{
  "decision": "DENIED",
  "intent_id": "intent-7c2e1f4a",
  "warrant_id": "warrant-prod-coding-agent-2026-04-26",
  "matched_rule": {
    "block": "tool_calls.bash.blocked_commands",
    "value": "rm -rf",
    "rationale": "destructive recursive removal is outside the signed policy"
  },
  "policy_hash": "sha256:9a1fe408",
  "evaluated_at": "2026-04-26T14:32:18.241Z",
  "expires_at": "2026-04-26T14:33:18.241Z",
  "signature": "ed25519:b72c118f"
}
The signed decision is the operational evidence. A reviewer can verify the signature, compare the policy_hash to the warrant, inspect the declared intent, and see the rule that fired. The decision records that governance happened before execution, at the point where the agent still needed permission to act.

5. The hook acts on the decision

Agent Hooks parses the decision. Because the result is DENIED, the adapter returns the runtime’s denial response and prevents the bash tool from executing.
{
  "hookSpecificOutput": {
    "hookEventName": "PreToolUse",
    "permissionDecision": "deny",
    "permissionDecisionReason": "SIGIL_BASH_BLOCKED: rm -rf matched tool_calls.bash.blocked_commands in warrant-prod-coding-agent-2026-04-26."
  }
}
The bash runtime never receives authorization to run the command. The filesystem remains unchanged. The agent receives the denial reason through the tool-result channel, which lets it propose a safer alternative without bypassing the boundary that rejected the original action. Sigil Command records the event as an operator-visible enforcement record:
2026-04-26T14:32:18Z  DENIED  tool_calls.bash
session:      session-9b3a
intent:       intent-7c2e1f4a
warrant:      warrant-prod-coding-agent-2026-04-26
command:      rm -rf ./build
matched:      tool_calls.bash.blocked_commands rm -rf
policy_hash:  sha256:9a1fe408
attestation:  ed25519:b72c118f
The denial is not merely a missing execution. It is a positive evidence event. The record shows the attempted action, the policy, the matched rule, the decision, the hash, and the signed proof.

6. What the evidence chain proves

The example answers the questions an auditor or compliance lead will ask.
Audit questionEvidence produced
Was there a governing policy?Yes. warrant-prod-coding-agent-2026-04-26 existed as a signed Sigil Warrant.
Was the policy authorized at the time of decision?Yes. The signature and policy hash bind the decision to the warrant version.
What did the agent try to do?The declared intent records tool_calls.bash with rm -rf ./build.
Did the system evaluate the action before execution?Yes. The PreToolUse path submitted the intent before the tool executed.
Which rule matched?tool_calls.bash.blocked_commands matched rm -rf.
Did the system block the action?Yes. The hook returned permissionDecision: deny.
Can a third party verify the chain?Yes. The policy hash, signed decision, warrant, and event log create replayable evidence.
This is the operational form of several standards surfaces. The signed warrant maps to ISO/IEC 42001 A.2.2 and NIST AI RMF Govern 1.3 because it expresses approved policy as an evaluable artifact. The PreToolUse checkpoint maps to NIST AI RMF Map 3.5 and Berkeley Agentic AI Profile Map 3.5 because it inserts human-control policy before tool use. The denied decision maps to NIST AI RMF Manage 2.4, Berkeley GPAI Profile v1.2 Manage 2.4, and Berkeley Agentic AI Profile Manage 2.4 because it demonstrates a mechanism to disengage an action inconsistent with intended use. The Command event maps to ISO/IEC 42001 A.6.2.8, NIST AI RMF Manage 4.1, Berkeley GPAI Profile v1.2 Manage 4.1, and Berkeley Agentic AI Profile Manage 4.1 because it records the enforcement event for monitoring and audit.

7. What changes when policy permits the action

The same path approves permitted work. If the agent runs a command allowed by the signed warrant, the hook constructs the same kind of intent and Sigil returns an approved decision.
npm test
{
  "intent_id": "intent-3f81c0d2",
  "session_id": "session-9b3a",
  "warrant_id": "warrant-prod-coding-agent-2026-04-26",
  "action_class": "tool_calls.bash",
  "action_payload": {
    "tool": "Bash",
    "command": "npm test",
    "cwd": "/Users/operator/project"
  },
  "timestamp": "2026-04-26T14:33:02.211Z"
}
{
  "decision": "APPROVED",
  "intent_id": "intent-3f81c0d2",
  "warrant_id": "warrant-prod-coding-agent-2026-04-26",
  "matched_rule": {
    "block": "tool_calls.allowed",
    "value": "bash",
    "rationale": "bash tool allowed and no blocked command matched"
  },
  "policy_hash": "sha256:9a1fe408",
  "evaluated_at": "2026-04-26T14:33:02.412Z",
  "expires_at": "2026-04-26T14:34:02.412Z",
  "signature": "ed25519:c84e207a"
}
The hook returns permissionDecision: allow, and the bash runtime executes the command. The event log now shows declared intent, signed authorization, and observed execution. The same control surface that denies destructive actions issues short-lived proof for permitted actions.

8. What changes when human approval is required

The same path also supports hold states. Suppose the agent proposes an outbound email:
{
  "tool": "email.send",
  "to": "outside-counsel@example.com",
  "subject": "Draft diligence memo",
  "body": "Attached is the current diligence summary."
}
The warrant can require approval for email.send. In that case, Sigil returns PENDING instead of APPROVED or DENIED.
{
  "decision": "PENDING",
  "intent_id": "intent-4a9d02be",
  "warrant_id": "warrant-prod-coding-agent-2026-04-26",
  "matched_rule": {
    "block": "tool_calls.email.require_approval",
    "value": "true",
    "rationale": "outbound email requires human approval"
  },
  "hold_id": "hold-82d11c",
  "policy_hash": "sha256:9a1fe408",
  "evaluated_at": "2026-04-26T14:36:11.018Z",
  "signature": "ed25519:f29a088d"
}
The hook does not let the message proceed. Command becomes the review surface. The human operator approves or rejects the hold. Only an approved hold can release execution proof. This is the runtime form of Berkeley GPAI Profile v1.2’s human oversight and structured-access logic, and Berkeley Agentic AI Profile’s real-time permission request and oversight-checkpoint guidance.

9. What the example does not prove

The example does not prove that the model is well-behaved. Pre-execution enforcement sits downstream of model reasoning. A model that repeatedly proposes denied actions wastes operator time and degrades agent utility. SOF defends against the consequences of unauthorized action. It does not improve the model’s intent generation. Model evaluation, red-teaming, and TEVV remain separate responsibilities. The example does not prove completeness of coverage. It demonstrates that this attempted action passed through enforcement. It does not prove that every attempted action in the session passed through enforcement. A tool registered without hooks, a side-channel command path, or a runtime that bypasses the SDK hook contract would sit outside this evidence chain. Operational completeness depends on consistent hook registration across every tool and target system in scope. The example does not prove that the operator authored the correct policy. A warrant that allowed destructive commands would permit them. SOF enforces the policy the operator signs. It cannot decide whether the operator chose the right risk threshold. Policy review, approval thresholds, intended-use characterization, and risk acceptance remain organizational responsibilities.

Coverage and Failure Modes

The first auditor question is coverage. A denial log proves that one attempted action passed through SOF. It does not prove that all actions in the deployment passed through SOF. A defensible deployment therefore needs a coverage statement. The statement should identify which agent frameworks, tools, APIs, wallets, gateways, and credential routes are in scope. It should identify out-of-band paths and either remove them, monitor them, or mark them outside the claim. Without that scope boundary, an attestation proves the governed path worked, not that every possible path was governed. The second question is failure behavior. The published Agent Hooks package supports failMode: 'open' and failMode: 'closed'. In open mode, Sigil unreachability can return an approved fallback with a failOpen marker. In closed mode, unreachability returns DENIED with SIGIL_UNREACHABLE. The package documentation recommends closed mode for production agents, externally visible actions, and wallet or on-chain actions. Vault is stricter. Missing, invalid, expired, replayed, or otherwise unacceptable attestations are hard-rejected, and backend timeouts do not pass through blindly. SOF therefore should not claim a single universal fail-closed default across all components. The correct claim is more precise: high-stakes deployments should configure fail-closed behavior at the hook layer, and Vault enforces hard rejection for credential release. The third question is policy integrity. Sigil Warrant signs the policy with the operator’s Ed25519 key. The policy hash is embedded in Intent Attestations. If the file changes after signing, Sigil detects the signature failure at startup. That creates strong policy integrity evidence. It does not prove that the policy itself was well chosen. Risk thresholds, blocked commands, approval requirements, and intended-use definitions still require human judgment and review. The fourth question is key and policy rotation. Because attestations are short-lived and bind to a policy hash, rotation should be visible in the evidence trail. A new Warrant version should produce a new policy hash. A new signing key should update the verification material and deployment record. Auditors should be able to sample actions before and after rotation and determine which policy and key governed each action. Execution control also has a legal dimension. AI agents are not legal entities. They cannot assume fiduciary duties, sign contracts as persons, or carry liability in the way a human operator or company can. The Fiduciary Agent Framework (FAF) addresses that gap by pairing the technical enforcement layer supplied by Open Execution Engine (OEE) with a legal wrapper around agent authority. FAF treats the agent as controlled property, anchors it to a human or legal entity, and uses the signed policy and attestation trail as evidence that the agent’s authority was bounded. FAF and OEE solve different parts of the same fiduciary problem. FAF supplies the legal container. Its model wraps an autonomous agent in an entity structure and defines the agent’s scope of authority, spend limits, risk tolerance, compliance duties, and human override conditions. OEE supplies the technical enforcement path. It routes high-stakes actions through signed policy evaluation before execution. The Open Venture Engine (OVE), OEE’s first active vertical, applies that pattern to venture and fund operations with domain-tuned policy templates, integration examples, and documentation. The sequence matters. FAF defines who owns the agent, what authority the agent has, and which human or legal entity remains responsible for it. OEE makes those boundaries executable. Sigil Warrant translates the legal and operational boundary into warranty.md. Sigil Sign evaluates each declared intent against that policy. Intent Attestations give the execution layer cryptographic proof when an action complies. Credential sequestration keeps sensitive authority outside the model’s reasoning context. A denied or held action creates evidence that the system enforced the boundary before loss, not after. This matters because the Pre-Execution Enforcement Gap is also a liability gap. A human operator can say the agent was not supposed to move funds, send an email, call an API, or change infrastructure. That statement helps after a loss, but it does not bound the agent’s technical authority before the loss. FAF gives the legal structure a stronger factual foundation when the technical system can show that high-stakes actions required policy-bound proof. OEE closes the execution gap. FAF helps close the legal accountability gap.

Input to Future Agentic Standards

An open question for future iterations of agentic-AI standards work is whether the human control and accountability lever should include a named enforcement primitive. One candidate name is pre-execution authorization. The substance already lives across NIST AI RMF 1.0, Berkeley GPAI Profile v1.2, and Berkeley’s Agentic AI Profile. NIST supplies the base language for human oversight, deactivation, post-deployment monitoring, and risk treatment. Berkeley adds current GPAI guidance around role allocation, structured access, staged release, deactivation, incident response, monitoring, external feedback, and assurance. The Agentic AI Profile adds the agent-specific guidance around human-agentic management hierarchies, intervention points, real-time permission requests, role-based permission boundaries, tool-use logs, resource-access records, and shutdown mechanisms. What remains missing is the unifying control name. The control pattern is straightforward. A deployer should be able to show that high-stakes actions pass through policy evaluation before the action reaches the tool, API, wallet, file system, or credential broker. The action should carry proof when approved. It should stop when denied. It should enter a hold state when human approval is required. The record should bind the policy, intent, decision, and execution path. That pattern would make Berkeley GPAI Profile v1.2 and the Agentic AI Profile easier to operationalize across ISO/IEC 42001, NIST AI RMF, and sector-specific audit programs. The industry has built many advisory controls. System cards, model cards, evaluation reports, red-team memos, acceptable-use policies, procurement questionnaires, and audit dashboards all help. They make governance visible. They support regulators and customers. They cannot serve as the final boundary for autonomous execution. An agent does not need to persuade a document. It needs a credential, a network path, a signing key, or an API token. Pre-execution enforcement treats those capabilities as controlled releases. The agent declares intent before it acts. The policy engine evaluates that intent against the signed warrant. The system issues proof only when the action fits the approved boundary. The execution layer verifies proof before it lets the action proceed. The credential layer releases access only after it sees valid proof. The operator layer records denials and holds in real time. This creates a direct line from human intent to machine action. AI standards have defined the management system. SOF proposes the execution control. ISO/IEC 42001, NIST AI RMF, ISO/IEC 23894, Berkeley GPAI Profile v1.2, Berkeley’s Agentic AI Profile, and the EU AI Act tell organizations what responsible AI governance must cover. Sigil answers the agentic question those frameworks now raise: how does an organization enforce policy-bound human intent at the moment an autonomous system attempts a high-stakes action? The answer is pre-execution enforcement. Every high-stakes action should pass through deterministic policy evaluation before it reaches the target system. Every approval should produce verifiable proof. Every denial should block execution. Every hold should require a human decision. Every credential release should depend on authorization. Every audit trail should bind the action to the policy that allowed it. That is the missing control family for agentic AI governance.

References