From reviewing governance documentation across organizations deploying agentic AI in 2026, the shift from general AI policies to agent-specific charters mirrors how DevOps matured from broad security policies to per-service access controls. The insurance industry sits at the same inflection. Carriers with AI in production grew from 37% to 61% in a single year, according to a Datos Insights survey of 36 senior carrier technology leaders presented at the April 2026 Insurance Leadership and Technology Forum. But enterprise AI governance policies, the kind that fill binders and satisfy board-level inquiries, remain disconnected from the operational reality of an AI agent processing submissions, comparing policies, or triaging claims in real time. The Agent Charter is the document designed to close that gap.
The concept gained visibility in a May 12, 2026 IA Magazine feature by Steve Forte of Patra Corp, who outlined a structured framework for defining AI decision authority at the individual agent level. The timing is not accidental. The NAIC’s Spring 2026 panel on agentic AI flagged autonomous systems as the industry’s next governance gap. Grant Thornton’s 2026 AI Impact Survey found that only 24% of insurance executives were confident they could pass an independent AI governance review in 90 days. And the Federal Reserve’s new SR 26-2 guidance, which superseded the SR 11-7 model risk management framework on April 17, 2026, explicitly excluded generative and agentic AI from its scope, leaving insurers without a federal benchmark for the systems they are deploying fastest.
The Agent Charter addresses a specific problem: the gap between enterprise-level AI policy ("we govern AI responsibly") and production-level AI operations ("this agent can approve policy comparisons with less than 5% variance without human review"). That specificity is what makes it useful, and what distinguishes it from the broader governance frameworks the industry has been building over the past two years.
What an Agent Charter Actually Is
An Agent Charter is an internal governance document that defines the operational authority of a specific AI agent or agentic workflow. Where an enterprise AI governance policy articulates principles, risk appetite, and organizational accountability at the company level, the Agent Charter operates one layer below: it specifies, for a given AI agent in a given business context, exactly what decisions the agent can make autonomously, what triggers require human review, and what performance metrics determine whether the agent continues operating within its defined boundaries.
The framework distinguishes three categories of AI decision-making, each mapped to a different level of human involvement:
- Deterministic decisions: Data normalization, classification, and formatting tasks where the AI operates autonomously. These are operations with well-defined inputs and outputs, minimal ambiguity, and negligible regulatory or coverage consequences if the output contains an error. Document intake parsing, field extraction, and carrier code lookup fall into this category.
- Probabilistic recommendations: Risk scoring, market matching, and coverage comparison tasks where the AI produces a recommendation that a human reviews before action. The AI adds value by processing volume and identifying patterns, but the output represents a judgment call that benefits from human contextual understanding. Submission triage, renewal pricing recommendations, and cross-policy comparison sit here.
- Human-required decisions: Coverage interpretation, binding authority, policy language analysis, and client advisory functions where licensed professionals retain full decision-making authority. The AI may prepare materials, surface relevant precedents, or flag anomalies, but the decision itself remains human. No amount of AI performance data should move these decisions into the autonomous category without explicit regulatory and organizational sign-off.
This three-tier structure draws an implicit analogy to banking’s lending authority matrices, where individual loan officers have defined approval limits and anything above those limits requires escalation to a committee or senior officer. The insurance version adds a regulatory dimension that banking authority matrices do not carry: many of the decisions in the “human-required” tier involve licensed activities (binding coverage, interpreting policy language, advising clients) where state insurance law, not just company policy, mandates human involvement.
Measurable Approval Thresholds
The operational core of the Agent Charter framework lies in its approval thresholds: specific, measurable conditions that determine when an AI agent’s output passes through automatically versus when it triggers human review. Without these thresholds, the three-tier classification is just taxonomy. With them, governance becomes an executable rule set that can be audited, tested, and adjusted based on performance data.
The IA Magazine feature provided a concrete example from policy checking workflows. When an AI agent compares a newly issued policy against the quoted terms:
- Policy field variances under 5% pass automatically. The agent marks the comparison as complete, logs the variance, and moves the policy into the “verified” queue.
- Variances exceeding 5% trigger a manual review workflow. The agent flags the specific fields, quantifies the deviation, and routes the case to the appropriate reviewer.
- Critical field differences in limits, endorsements, or exclusions mandate immediate escalation to a licensed professional, regardless of the magnitude of variance. A $1 difference in a general aggregate limit is functionally different from a $1 difference in a premium calculation.
This threshold approach converts governance from what the framework describes as "a philosophical concept into an operational mechanism." The 5% threshold is not arbitrary; it reflects a calibrated judgment about where the cost of human review exceeds the risk of automated processing error. Carriers implementing Agent Charters will need to calibrate these thresholds to their own risk tolerances, and critically, document the rationale for the threshold levels they choose. When a regulator under the NAIC’s 12-state AI evaluation pilot asks why a particular agent operates without human review below a given threshold, the carrier needs an answer grounded in performance data, not convenience.
Early adopters report 60% to 80% time reductions in policy checking workflows when AI decision boundaries are clearly documented and enforced. That efficiency gain is significant, but it also creates its own governance challenge: the better the AI performs, the stronger the organizational incentive to widen its autonomous decision authority. The Agent Charter framework builds in a structural counterweight by requiring explicit, documented justification for any threshold change.
Authority Creep: The Governance Risk Agent Charters Address
The most underappreciated risk in agentic AI deployment is not dramatic failure. It is what the Agent Charter framework calls authority creep: the gradual, undocumented expansion of an AI agent’s effective decision authority as staff members grow comfortable with its outputs and reduce their review intensity.
The pattern follows a predictable arc. An AI agent launches with rigorous human oversight. Staff review every output, challenge edge cases, and escalate uncertainties. Over weeks, the agent produces consistently acceptable results. Review becomes cursory. Validation shrinks to spot-checking. Eventually, staff accept AI outputs as correct by default, and the human-in-the-loop requirement becomes a rubber stamp rather than a genuine review. Stanford’s Human-Centered AI Institute documented this dynamic in a 2023 study that found users were significantly more likely to accept incorrect AI outputs when the system had established a track record of accuracy, a finding with direct relevance to insurance workflows where a single uncaught error in policy language or coverage limits can generate material claims exposure.
Agent Charters counter authority creep through three mechanisms. First, they document decision boundaries in writing, creating an auditable record of what the agent is authorized to do. If staff are accepting outputs that fall outside the agent’s documented authority, the gap is visible to compliance reviewers. Second, they require periodic threshold review, preventing the informal expansion of agent authority without formal governance approval. Third, they mandate performance monitoring tied to specific metrics, not just pass rates, so that a decline in accuracy that would otherwise go unnoticed triggers an automatic governance review.
For actuaries, authority creep has a direct professional liability dimension. If an agent operating in a pricing or reserving workflow has informally expanded beyond its documented boundaries, and the actuary signing the rate filing or reserve opinion was unaware of that expansion, the actuary’s exposure is not reduced by the lack of documentation. ASOP No. 56 requires understanding the models on which actuarial work relies. An agent that has drifted from its charter is, by definition, a model whose behavior the actuary does not fully understand.
Where Agent Charters Sit in the Regulatory Stack
The Agent Charter framework occupies a specific niche in the emerging AI governance architecture: the operational layer between enterprise policy and regulatory compliance. Understanding how it relates to, and differs from, existing governance instruments is essential for carriers trying to build a coherent governance program rather than a collection of overlapping documents.
Versus the NAIC Model Bulletin. The December 2023 Model Bulletin, adopted by 24 states and the District of Columbia as of March 2025, requires insurers to maintain written AI governance programs covering risk management, documentation, third-party oversight, and consumer protection. It is a compliance mandate. The Agent Charter is not a replacement for Model Bulletin compliance; it is a method of implementing the Bulletin’s governance requirements at the individual agent level. A carrier with a Model Bulletin-compliant governance program but no agent-level operational documentation has satisfied the regulatory requirement but may lack the operational infrastructure to enforce it. Conversely, a carrier with detailed Agent Charters but no enterprise governance program has operational controls but no compliance framework.
Versus model risk management (MRM). Traditional MRM, codified in the banking sector through the Federal Reserve’s SR 11-7 guidance (April 2011) and voluntarily adopted by many insurers, focuses on validating individual models: testing inputs, outputs, assumptions, and performance against holdout data. Agent Charters operate at a different unit of analysis. Rather than validating a model’s statistical properties, they define the operational boundaries within which a model’s outputs can be acted upon without human review. The two are complementary: MRM validates that the model works correctly, while the Agent Charter governs how much authority the model’s outputs carry in production workflows.
Versus the NAIC AI Evaluation Tool. The 12-state pilot running through September 2026 uses four exhibits to assess carrier AI governance. Exhibit C, which gathers detailed information on high-risk AI systems, comes closest to the Agent Charter’s scope but is structured around individual model documentation, not per-agent operational authority. Carriers that have implemented Agent Charters will find Exhibit C significantly easier to complete because the operational data (decision boundaries, threshold justifications, escalation rates, performance metrics) already exists in structured form.
Versus enterprise AI policy. Most carriers with formal AI governance programs have an enterprise policy that articulates principles, risk appetite, organizational roles, and high-level controls. These policies are necessary for board governance and regulatory communication. They are insufficient for production operations because they do not specify, at the agent level, what the AI can and cannot do. The Agent Charter bridges the gap between "our company governs AI responsibly" and "this specific agent in the policy checking workflow has these specific decision authorities with these specific escalation triggers."
The SR 26-2 Gap and Why It Matters
On April 17, 2026, the Federal Reserve issued SR 26-2, superseding the 15-year-old SR 11-7 framework that had served as the banking sector’s definitive model risk management guidance. Many insurers had adopted SR 11-7 voluntarily as a best-practice benchmark, so the update has cross-sector significance.
SR 26-2 modernizes MRM in several ways: it tailors requirements by institution size and complexity, narrows the definition of “model” to exclude simple calculators and deterministic tools, and shifts validation timing to a risk-based approach driven by materiality rather than default annual cycles. The philosophy, as summarized by regulatory analysts at Sia Partners, emphasizes “precision over volume, risk alignment over uniformity, governance effectiveness over procedural rigidity.”
But SR 26-2 contains a critical exclusion: it explicitly carves out generative and agentic AI from its scope, indicating that these systems will “likely” receive future regulatory treatment elsewhere. For the growing number of insurers who had been using SR 11-7 as their MRM framework for AI systems, this creates an immediate gap. The old guidance applied broadly enough that carriers could argue their agentic systems fell within its scope. The new guidance says, in writing, that they do not.
A GARP analysis by Krishan Sharma of Citigroup identified three specific ways SR 11-7 was already insufficient for agentic systems before SR 26-2 made the exclusion explicit. First, the framework assumes models “whose structure and behavior remain stable between review cycles,” while agentic systems can recalibrate autonomously. Second, the concentration of insurance AI stacks on a small number of foundation model providers (OpenAI sits in 90% of carrier AI stacks) creates correlated risks that firm-specific SR 11-7 controls cannot address. Third, the framework offers “limited guidance on what constitutes sufficient explainability for complex models,” a gap that deepens with every LLM-powered agent deployment.
Agent Charters do not replace model risk management. But in the gap between SR 26-2’s explicit exclusion and whatever future regulatory guidance covers agentic AI, they provide a structured operational framework that carriers can implement now. The combination of MRM (validating that the model works) and Agent Charter (defining what the model is allowed to do) gives actuaries and compliance teams a defensible governance position while the regulatory landscape develops.
High-Impact Use Cases
The Agent Charter framework identifies three insurance workflows where the combination of volume, complexity, and time pressure makes agent-level governance particularly valuable:
Policy checking. Comparing issued policies against quoted terms is a high-volume, detail-intensive task with clear right-and-wrong outcomes for most fields but material judgment calls for coverage language. The 60–80% time reduction reported by early adopters reflects the efficiency of automating the deterministic comparisons (premium amounts, effective dates, named insureds) while escalating the judgment calls (endorsement language, exclusion scope, sublimit structures). The 5% variance threshold provides a calibrated boundary that agents can process consistently at volumes no human team could match.
Submission intake. Processing incoming insurance submissions requires data extraction from inconsistent formats, classification into lines of business, and initial risk assessment. Agent Charters define which fields the AI extracts autonomously, which classifications it can assign without review, and which risk indicators trigger underwriter involvement. AIG’s Lexington Insurance unit processed over 370,000 E&S submissions in 2025 with a target of 500,000 by 2030; at that scale, the difference between well-governed and ungoverned AI intake is the difference between scalable underwriting and a compliance liability.
Renewal intelligence. Aggregating loss history, market conditions, account changes, and competitive pricing for renewal recommendations involves both data synthesis (deterministic) and competitive positioning judgment (human-required). The Agent Charter delineates between synthesizing data from multiple sources, which the AI handles, and recommending retention strategies or competitive concessions, which require human judgment informed by client relationships the AI cannot observe.
Measuring Success: From Activity Metrics to Outcome Metrics
One of the more operationally useful elements of the Agent Charter framework is its explicit shift in success measurement from activity metrics to outcome metrics. Traditional workflow reporting counts policies processed, submissions reviewed, or renewals completed. These metrics measure throughput but reveal nothing about whether the AI agent’s involvement actually improved business outcomes.
The framework points to revenue per employee as a more meaningful benchmark. Top-performing insurance agencies reached a record $228,321 in revenue per employee in 2025, according to the Big “I” and Reagan Consulting 2025 Best Practices Study, which analyzed 349 agencies qualifying from over 1,100 nominees. Those agencies also posted record 10.7% organic growth rates and 25.1 Rule of 20 composite scores. The study does not attribute these gains directly to AI, but the timing aligns with the first wave of production AI deployments in agency workflows.
For carriers evaluating AI agent performance under a charter framework, the relevant metrics extend beyond per-employee productivity to include escalation accuracy (how often human-escalated cases actually required human judgment), threshold calibration (whether the 5% variance boundary, or whatever threshold a given agent uses, correctly separates cases that need review from those that do not), and downstream error rates (whether policies that passed through automated comparison without human review subsequently generated coverage disputes, E&O claims, or regulatory findings).
The Capgemini World P&C Insurance Report 2026 found that 42% of insurers track no AI metrics at all, and 55% report unclear ROI from AI initiatives. Agent Charters provide a structural remedy for this measurement gap by requiring metric definition as a condition of agent deployment. An agent without defined success metrics does not receive a charter, and without a charter it does not deploy to production.
Cross-Survey Context: The Deployment Landscape
The Agent Charter framework emerges against a backdrop of rapidly accelerating but unevenly governed AI deployment across the insurance industry. Multiple concurrent surveys paint a consistent picture of an industry that has moved decisively toward AI adoption but has not kept governance pace.
| Survey | Sample | Key Finding |
|---|---|---|
| Datos Insights ILTF (April 2026) | 36 carrier tech leaders | AI in production: 37% to 61% in one year |
| Grant Thornton AI Impact (2026) | 100 insurance executives | Only 24% confident in passing AI governance audit |
| Capgemini World P&C (May 2026) | 344 executives, 809 employees | Only 10% successfully scaling AI; 42% track no metrics |
| AM Best (April 2026) | 150+ rated insurers/MGAs | 41% actively use AI; ~20% at advanced implementation |
| Conning AI & Insurance Tech (2025) | Industry-wide survey | 90% evaluating GenAI; 55% in early or full adoption |
| NAIC Survey (2025) | Carrier reporting | 88% of auto, 70% of home insurers adopting AI |
The pattern across these surveys is consistent: adoption is high but maturity is low, investment is substantial but governance lags, and the carriers that report both strong AI performance and strong governance are a small minority. Capgemini’s “Intelligence Trailblazers,” the top 10% of surveyed P&C insurers, showed 21% higher revenue growth and 51% greater share price gains over three years. These trailblazers were four times more likely to invest in change management beyond basic training and three times more likely to have explainable AI infrastructure. They were also, notably, two times more likely to embed AI responsibilities into job descriptions, which is a granular governance practice that parallels the per-agent specificity of the Agent Charter approach.
McKinsey estimates $50 to $70 billion in annual value from generative AI in insurance. But that value requires operational governance infrastructure to realize. The Agent Charter provides one component of that infrastructure: the per-agent operational document that turns high-level AI strategy into auditable, enforceable, measurable production controls.
Why This Matters for Actuaries
Agent Charters intersect with actuarial practice at multiple points, and the profession’s exposure grows as agentic AI embeds deeper into pricing, reserving, and underwriting workflows.
Model validation scope expansion. ASOP No. 56 requires actuaries to understand the models on which their work relies, including limitations and known weaknesses. For an agentic AI system operating under a charter, the validation scope now includes not just the model’s statistical properties but the operational boundaries the charter defines. If the charter specifies a 5% variance threshold for autonomous processing, the validating actuary needs to assess whether that threshold is appropriate given the downstream impact on pricing accuracy, reserve estimates, or coverage consistency. This is a different kind of validation from testing a GLM’s lift curves or a gradient-boosted model’s feature importance, and it requires actuaries to develop competency in operational governance assessment alongside statistical model evaluation.
Escalation calibration as an actuarial question. The effectiveness of human-in-the-loop controls depends entirely on escalation trigger calibration. Setting thresholds too low generates noise that overwhelms reviewers. Setting them too high allows material errors to pass through. This is fundamentally an actuarial problem: calibrating the tradeoff between Type I errors (unnecessary escalations) and Type II errors (missed escalations) based on the economic cost of each, precisely the kind of analysis actuaries perform when setting credibility standards, loss development factors, or risk margins. Carriers building Agent Charters should involve actuaries in threshold calibration, not just in subsequent model validation.
Audit trail for regulatory filings. As the NAIC’s 12-state evaluation pilot matures, carriers will face increasing scrutiny of how AI outputs influence rate filings, reserve opinions, and market conduct. An Agent Charter provides the documentation backbone for this scrutiny: it shows what the AI was authorized to do, how its outputs were constrained, what performance thresholds it met, and where human judgment intervened. Without this documentation, an actuary defending a rate filing that incorporated AI-generated risk scores has no structured way to demonstrate governance over the AI’s role in the filing.
The Hartford precedent. Hartford’s voluntary publication of an Algorithmic Impact Assessment in February 2026, covering bias audits for ZIP code, age, and property type, demonstrates that governance documentation has begun to function as a competitive differentiator. Carriers with comprehensive Agent Charters can point to documented, auditable governance when regulators, reinsurers, or large commercial clients ask how AI is governed in production. Carriers without them face the governance version of the Grant Thornton proof gap: they may be governing well in practice, but they cannot prove it on demand.
Professional liability protection. For the appointed actuary, signing actuary, or consulting actuary whose work product depends on AI agent outputs, the Agent Charter provides a documented governance layer that strengthens the professional’s position if outcomes are later questioned. The alternative, relying on AI outputs from agents with no documented operational boundaries, leaves the actuary unable to demonstrate the due diligence that ASOP No. 56 and ASOP No. 41 (Actuarial Communications) require when reporting results that depend on models.
Further Reading
- NAIC Flags Agentic AI as Insurance’s Next Governance Gap
- Grant Thornton Survey Exposes the Insurance AI Proof Gap
- Hartford’s Algorithmic Impact Assessment Sets the Carrier Transparency Bar
- The AI Governance Gap in Actuarial Practice
- Carrier AI Projects Fail at the Audit Layer, Not the Tech
- 30-Hour AI Agents Push the Limits of Carrier Oversight
- Insurer AI Adoption Hits 82% But Only 7% Reach Full Scale
- Insurance AI Pivots From Claims Efficiency to Underwriting: ILTF 2026
Sources
- IA Magazine: Agent Charter: Creating an AI Governance Framework to Ensure Operational Reliance (May 12, 2026)
- Datos Insights: ILTF 2026: Defining the New Insurance Carrier Operating Model for AI
- Grant Thornton: 2026 AI Impact Survey Report
- Capgemini: World P&C Insurance Report 2026
- Big “I” & Reagan Consulting: 2025 Best Practices Study
- GARP: SR 11-7 in the Age of Agentic AI (February 2026)
- Sia Partners: SR 11-7 vs. SR 26-2: Model Risk Management Modernization
- NAIC: Model Bulletin on the Use of Artificial Intelligence Systems by Insurers (December 2023)
- Plante Moran: How the NAIC AI Model Bulletin Is Evolving (March 2026)
- Insurance Thought Leadership / Roots Automation: 2026 Is the Year AI Goes Operational in Insurance
- Carrier Management / Finys: Plug-and-Play Operating Model for Insurance AI (April 2026)