Key takeaways
  • The autonomy envelope specifies permitted actions, resource thresholds, environmental prerequisites, halt conditions, and audit trail requirements for every AI agent in scope.
  • EU AI Act Article 14 requires that human oversight is technically possible. The autonomy envelope is the document that makes that requirement evidenceable.
  • The Autonomy dimension in the Agent Certified framework scores from 1 (informal or absent) to 10 (third-party adversarial tested). Most enterprise agents score 3 to 5 at first assessment.
  • Insurance underwriters, specifically Munich Re aiSure, Armilla, and AIUC-1 licensees, require a written autonomy envelope to define the scope of cover. No scope, no policy.
  • Specifying the envelope before deployment is cheaper than reconstructing it after an incident. Incident reconstruction under time pressure produces incomplete specifications that create coverage disputes.

Why the autonomy envelope is not optional

Most organisations running AI agents today have an informal autonomy envelope. The engineers who built the system know roughly what it does. The product manager who scoped it knows roughly what it should not do. But that informal understanding is not written down in a single document, it is not technically enforced by the system itself, and it is not available to the compliance officer, the insurer, or the regulator who needs to evaluate what the agent's scope of action actually is.

The informal envelope is sufficient for a prototype. It is insufficient for a production system under the EU AI Act, for a system covered by an AI liability insurance policy, and for a system that will be subject to post-incident investigation.

The EU AI Act Article 14 compliance case is direct. The provision requires providers of high-risk AI systems to ensure that the systems can be effectively overseen by natural persons. Effective oversight requires that the person doing the overseeing knows what the system is doing autonomously. If the scope of autonomous action is not documented, the oversight role is undefined, and the Article 14 requirement cannot be satisfied in a way that is evidenceable. A regulator or notified body reviewing the technical documentation for a high-risk AI system will expect to find a document that specifies the system's autonomous action scope. Organisations that cannot produce it are not in a position to demonstrate compliance, regardless of how responsibly they have operated the system in practice.

The insurance case is equally concrete. The 2024 Air Canada case (Moffatt v. Air Canada, 2024 BCCRT 149) illustrated how the absence of a clear scope definition for an AI system's outputs created a liability dispute that was harder to resolve precisely because the boundary between what the system was authorised to do and what it did was not documented in advance. The airline's position, that the chatbot was a separate entity from the airline and therefore its outputs were not binding, failed in part because there was no written document establishing what the chatbot was and was not authorised to say. An insurance underwriter facing the same post-incident question needs the autonomy envelope to determine whether the claim arises from authorised or unauthorised autonomous action, a distinction that is commercially and legally material for the policy's applicability.

The operational case for documenting the envelope before deployment rather than after an incident is straightforward. Reconstructing scope from code, logs, and developer recollection under post-incident time pressure produces incomplete and contested specifications. The scope that emerges from that process is almost always narrower and more ambiguous than the scope that would have been documented in advance, because the incident creates incentives to narrow the stated scope retrospectively. Those retrospective narrowings create gaps between the coverage the operator thought it had and the coverage the insurer is willing to provide for the incident in question.

The five categories of the specification

A complete autonomy envelope specification covers five categories. Each category addresses a different dimension of the boundary between autonomous operation and human oversight. Together they produce a document that is sufficient for Article 14 compliance evidence, for an insurance underwriting submission, and for the technical documentation file required by EU AI Act Annex IV for high-risk systems.

Category one: permitted action scope. This is the list of action types and targets the agent may address without human confirmation. A well-specified entry reads: "The agent may create, modify, and delete records in the Customer Communication table within the CRM system for any customer whose account status is Active, for communication types classified as Transactional or Service, and for messages with a delivery schedule no more than 24 hours in the future." A vague entry reads: "The agent may manage customer communications." The difference between these two entries is the difference between an auditable scope and an unauditable one. The specific entry can be tested, monitored, and enforced technically. The vague entry cannot.

Category two: resource thresholds. This category specifies the financial, data volume, system access, and operational limits above which the agent must escalate to a human for confirmation before proceeding. A well-specified entry reads: "The agent may not initiate any payment instruction exceeding EUR 500 without explicit human confirmation via the approval workflow. The agent may not modify more than 200 records in a single operation without a supervisory review checkpoint after the first 50 modifications. The agent may not access any system classified as Restricted under the organisation's data classification policy." A vague entry reads: "The agent operates within reasonable financial limits." Threshold specifications must be precise enough to be implemented as system-level controls, not as guidance to the model.

Category three: environmental prerequisites. This category defines the conditions that must be met before the agent begins autonomous operation. It covers data quality requirements (minimum completeness and recency thresholds for the inputs the agent uses to make decisions), context completeness requirements (which fields must be populated for the agent to proceed without flagging for human review), and system state conditions (which upstream systems must be in a healthy state for the agent to operate). A well-specified entry reads: "The agent requires that the customer record retrieved from the CRM is no older than 4 hours, that all mandatory fields in the Customer Profile schema are populated, and that the email sending service is returning a healthy status via the health check endpoint before beginning a campaign dispatch run." An absent specification in this category is what produces agents that continue operating on stale data, with incomplete context, or when upstream systems are degraded, precisely the conditions under which autonomous action is most likely to produce incorrect outcomes.

Category four: halt conditions. This category lists the error states, ambiguity conditions, and ethical flags that cause the agent to stop and escalate rather than act. It is the most consequential category for incident prevention. A well-specified halt condition reads: "If the agent encounters a customer record flagged as Disputed, Deceased, or Legal Hold, it must halt the current task, log the encounter with the specific record identifier and timestamp, and route the task to the human oversight queue for manual review before any further action. The agent must not attempt to reclassify the flag or proceed on the basis of surrounding context." A vague entry reads: "The agent handles edge cases appropriately." The halt conditions category is where the agent's behaviour under adversarial or unexpected inputs is defined. Systems without explicit halt conditions tend to behave poorly precisely when the stakes are highest, because those are the situations their designers did not anticipate clearly enough to specify as permitted action scope.

Category five: audit trail requirements. This category specifies what must be logged for every autonomous action and for how long. A well-specified entry reads: "For every action taken autonomously (record creation, modification, deletion, message dispatch, API call to an external system), the agent must log: the action type and target, the decision inputs used (with field names and values), the timestamp in UTC, the agent instance identifier, the session identifier, and the outcome. Logs must be retained for a minimum of 24 months in a system that is read-only to the agent and readable by the compliance and oversight function." A vague entry reads: "The agent maintains appropriate logs." The audit trail is the document that makes post-incident investigation possible. Without it, an incident produces a situation where the agent's decision process cannot be reconstructed, making root cause analysis and liability determination both harder and more contested.

The Article 14 connection

EU AI Act Article 14 is the human oversight provision for high-risk AI systems. It requires providers to design systems with measures that allow the persons to whom human oversight has been assigned to understand the system's capabilities and limitations, monitor the system's operation, intervene and halt the system, and interpret the system's output. These four requirements, understand, monitor, intervene, and interpret, map directly to the five categories of the autonomy envelope specification.

Understanding the system's capabilities and limitations requires a documented permitted action scope (category one) and resource thresholds (category two). Without those, the oversight person does not know what the system is authorised to do and cannot determine whether observed behaviour is within or outside the authorised scope.

Monitoring the system's operation requires an audit trail (category five) that captures what the system is doing in production in a form that can be reviewed without reconstructing the decision process from first principles. An audit trail that meets the specification requirements for category five is sufficient for production monitoring.

Intervening and halting the system requires that halt conditions (category four) are defined and technically enforced, and that the oversight person has access to the interface through which the human oversight queue is managed. A system without defined halt conditions may continue operating in states where it should not, and the oversight person has no clear basis for determining when intervention is required.

Interpreting the system's output requires that the audit trail captures the decision inputs (category five) and that the environmental prerequisites (category three) are documented so that the oversight person can assess whether the system was operating in the conditions it was designed for when a particular output was produced.

Article 26(2) of the EU AI Act adds a specific deployer obligation: the assignment of human oversight to natural persons with the competence to understand the system's outputs, the training to operate the oversight function, the authority to intervene, and the time and resources to do so. The autonomy envelope document is what makes that assignment concrete. Without it, the oversight assignment is nominal: a person has a job title that includes AI oversight, but there is no document specifying what oversight requires them to monitor, what they are authorised to intervene in, and what conditions should trigger their intervention. With a complete autonomy envelope, the assignment is operational: the oversight person knows exactly what the system is permitted to do, what thresholds trigger escalation, and what states require immediate halt.

Article 9 of the EU AI Act, covering the risk management system, requires providers to identify and analyse known and reasonably foreseeable risks. The autonomy envelope specification is the technical input to that analysis for the autonomous action dimension: the halt conditions and environmental prerequisites in the envelope are the operator's own documented analysis of the conditions under which autonomous operation creates risks that exceed the threshold for human confirmation.

The Agent Certified scoring rubric for Autonomy

The Autonomy dimension is one of the seven dimensions in the Agent Certified methodology. The other six dimensions cover Identity and Disclosure, Context Integrity, Action Reversibility, Oversight Architecture, Incident Response, and Compliance Documentation. Each dimension is scored on a 1 to 10 scale. The Autonomy dimension's rubric is structured as follows.

A score of 1 to 3 indicates that the envelope is informal or absent. The agent's boundaries are understood by the engineering team but not documented in a single accessible specification. There is no technical enforcement of scope boundaries: the system may take actions outside the intended scope if the model produces outputs that direct it to do so. There are no defined halt conditions: the system continues operating in error states. Audit trail coverage is incomplete or inconsistent. Most agents assessed in their first evaluation receive a score in this range. The gap between what organisations believe about their agent's scope and what is actually specified and enforced is consistently larger than expected.

A score of 4 to 6 indicates that the envelope is documented and partially enforced. The five categories are written down, at least in draft form, and the specification is accessible to compliance and oversight personnel. Technical enforcement exists for the most critical threshold categories (financial limits, access controls) but is absent or inconsistently applied for others. The audit trail captures the required fields but may have gaps in coverage for edge cases or for specific action types. Production monitoring is not systematically connected to the envelope specification: boundary violations may occur without generating an alert. This is a transitional score that indicates meaningful progress but substantial remaining work.

A score of 7 to 9 indicates a fully specified, technically enforced, and monitored envelope. All five categories are complete and reviewed at a defined cadence (at minimum, quarterly and after any significant change to the agent's capabilities or deployment context). Technical enforcement covers all action types and is tested as part of the deployment pipeline. Production monitoring generates alerts when the agent approaches or exceeds threshold values. Halt conditions are tested in staging environments before each deployment. Audit trails meet the retention and accessibility requirements specified in the envelope. Incident data from production operations is formally reviewed and used to update the specification. Assessments in this range are typically achieved after an organisation has run one full incident cycle against the specification: the incident identified gaps, the gaps were closed, and the updated specification was validated in production.

A score of 10 requires third-party adversarial testing of the envelope's boundaries. A qualified assessor attempts to induce the agent to take actions outside its specified scope through prompt injection, context manipulation, and tool misuse techniques. The agent must demonstrate that its technical enforcement holds under adversarial conditions, not just in standard operational scenarios. It also requires demonstrated boundary enforcement across multiple production incident cycles, with documented evidence that the specification was updated in response to real operational findings. Scores of 10 are uncommon at initial assessment and are typically achieved after 12 to 18 months of operating the agent against a formally managed specification.

The most common gap found between scores of 5 and 7 is the distinction between "technically enforced" and "technically enforced and production-monitored." An agent with code-level enforcement of its financial thresholds but no runtime monitoring that generates an alert when those thresholds are approached has a specification that works as a hard stop but not as an early warning system. Under Article 14's monitoring requirement, the hard stop alone is insufficient: effective oversight requires that the oversight person can see the system approaching its limits before the limit is reached and the action is blocked.

For a full treatment of how the Autonomy dimension interacts with the other six dimensions, see the seven dimensions article. The Oversight Architecture dimension (dimension five) is particularly closely related: it covers the technical and organisational mechanisms through which the oversight function specified in the autonomy envelope is actually implemented.

The insurance connection

AI agent liability insurance is an emerging product category. The major underwriters active in this space, Munich Re's aiSure programme, Armilla's AI performance and liability cover, and the licensees of the AI Underwriting Company (AIUC-1) reference standard, all require a defined scope of authorised action as a precondition for issuing a policy. The autonomy envelope specification is the document that satisfies that requirement.

The AIUC-1 reference standard specifies in its section on authorised action scope that underwriters must obtain a written specification of the actions the AI system is authorised to take autonomously, the thresholds above which human confirmation is required, and the conditions under which the system is designed to halt. The standard further specifies that claims arising from autonomous actions taken outside the documented authorised scope are excluded from coverage under the base policy. The rationale is that an action outside the certified scope is, by definition, an action the operator did not authorise and the underwriter did not price.

Munich Re's aiSure programme publishes a Schedule D that addresses autonomous action liability specifically. The schedule requires operators to maintain a current version of their authorised action specification and to notify aiSure of any material changes to the specification within 30 days. Claims arising from actions taken during periods when the specification was materially outdated or when a material change had not been notified are subject to a coverage suspension provision.

Armilla's coverage requires operators to specify which assessment framework their autonomy specification was developed against. Organisations that have undergone a formal assessment under the Agent Certified framework, or against the AIUC-1 standard, receive more favourable pricing than organisations presenting an unvalidated internal specification, because the third-party assessment reduces the underwriter's uncertainty about the specification's completeness and accuracy.

The practical consequence for operators seeking coverage is that an autonomy envelope specification that scores below a 5 on the Agent Certified rubric is unlikely to satisfy underwriting requirements, because a specification that lacks technical enforcement cannot provide the insurer with confidence that the defined scope constrains the system's actual behaviour. A specification that scores 7 or above, with documented monitoring and incident integration, is the documentation standard that major underwriters are calibrating to. For more on what AI agent insurance will and will not cover, the Agent Insured site provides a full treatment.

Steps to build and certify the envelope

Building a complete autonomy envelope specification from scratch requires approximately 10 to 14 weeks for a single agent deployment of moderate complexity. The following sequence reflects the order in which the work produces usable outputs at each stage.

The first step is agent inventory. Before specifying the envelope for any individual agent, the organisation needs a complete list of all agents in production, all agents in staged deployment, and all agents in active development that are expected to reach production within 12 months. The inventory must capture, for each agent, the deployment environment, the action types the agent can take, the systems it can access, and the oversight arrangement currently in place. Many organisations discover during this step that they have more agents in production than they believed, particularly in organisations where individual teams have deployed automation using AI frameworks without central visibility.

The second step is drafting the specification for each in-scope agent. Starting from the inventory output, the team responsible for each agent produces a first draft of all five categories. This step is best done with the engineering team and the compliance function in the same room: the engineers know what the system can technically do, and the compliance function knows what the organisation is prepared to authorise. The gaps between those two positions are the specification decisions that need to be made explicitly rather than left to developer judgment at build time.

The third step is threshold definition. Resource thresholds in particular require a structured decision process rather than a drafting process. The organisation must decide, with appropriate authority, what the financial, data volume, and system access limits are. These decisions involve risk appetite, and risk appetite decisions require input from the risk management function and, for regulated organisations, from the compliance officer. Threshold decisions made by the engineering team alone tend to be set too high, because the engineering team's default posture is to set limits that the agent will never realistically approach, rather than limits that reflect the risk appetite of the organisation as a whole.

The fourth step is halt condition mapping. This is the most analytically demanding step. It requires the team to systematically enumerate the inputs, states, and conditions under which autonomous operation should stop. A useful starting point is the list of past incidents or near-misses where the agent did something unexpected, the list of edge cases the engineering team is aware of but has not handled formally, and the list of ethical or reputational scenarios that the compliance function identifies as triggers for human review. The halt condition list should be treated as a living document: it will grow as the agent operates in production and encounters conditions that were not anticipated during specification.

The fifth step is technical enforcement implementation. For each element of the specification, the engineering team implements code-level or infrastructure-level controls that enforce the specification's requirements. Threshold enforcement, halt condition detection, and audit trail generation are the three areas where technical enforcement most commonly needs to be built rather than assumed. This step typically takes two to four weeks for an agent of moderate complexity, depending on how much enforcement capability already exists in the deployment infrastructure.

The sixth step is audit logging activation and verification. The audit trail specification must be implemented and the logs must be verified to confirm that all required fields are being captured for all action types. Gaps in coverage are common at first audit, particularly for error paths and halt condition triggers, which are less frequently exercised during development than the primary action paths.

The seventh step is a third-party assessment. Once the specification is complete and the technical enforcement and audit logging are in place, a formal assessment under the Agent Certified framework or an equivalent standard produces the certification artefact that satisfies regulatory and insurance underwriting requirements. The assessment covers all seven dimensions of the methodology, with the Autonomy dimension evaluated against the rubric described above. For organisations with an existing ISO 42001 management system in place, the assessment can reference existing governance documentation for the management-layer dimensions and focus the agent-specific evaluation on the technical dimensions.

A realistic timeline for an organisation doing this work from scratch, with a functioning engineering team and a compliance function available, is 10 to 14 weeks to reach a state where a third-party assessment is productive. Organisations that attempt to compress this timeline by skipping the threshold definition and halt condition mapping steps consistently produce specifications that fail the technical enforcement verification in the assessment, requiring a second assessment cycle. The specification work takes the time it takes. The value of getting it right the first time is a single assessment cycle rather than two, and a certification artefact that holds under scrutiny rather than requiring immediate remediation.