By Steve Raju

For Insurance

Cognitive Sovereignty Checklist for Insurance Underwriting and Claims

About 20 minutes Last reviewed March 2026

Your underwriters and claims assessors built their instinct over years of case work. AI tools like SAS, Guidewire, and Watson now make decisions at machine speed, often without explanation. The risk is real: systemic bias gets baked into the model, then applied to every policy, while your organisation loses the human discretion that once caught unfair but actuarially correct decisions.

Tool names in this checklist are examples. If you use different software, the same principle applies. Check what is relevant to your workflow, mark what is not applicable, and ignore the rest.
Cognitive sovereignty insight for Insurance: a typographic card from Steve Raju

These are suggestions. Take what fits, leave the rest.

Download printable PDF
0 / 21 applicable

Tap once to check, again to mark N/A, again to reset.

Audit AI Underwriting Decisions Against Your Actual Claims Data

Pull 100 recent underwriting decisions made by your AI tool alongside the actual claims outcomesbeginner
Compare what the model recommended against what actually happened in claims. This shows whether the model's logic matches real risk. If the model rejected applicants for reasons that never materialised as claims, your AI may be rejecting good business.
Segment your audit results by protected characteristics (age, postcode, gender, ethnicity where you hold the data)intermediate
Look for patterns where the AI approves or declines one group at a different rate than another. A model can be statistically accurate overall yet systematically unfair to a subset. Your regulator will ask for this analysis.
Ask your underwriters to manually assess 20 cases the AI flagged as high-risk declinesbeginner
Your underwriters have context the model lacks. They spot applicants with legitimate explanations for red flags. If underwriters overturn many AI declines, the model is missing judgement, not replacing it safely.
Document when your AI tool cannot explain why it rejected or approved a specific applicationintermediate
Black-box models create regulatory risk. If the FCA or your ombudsman asks why you declined someone, saying the AI decided it is not a valid answer. Track these unexplainable decisions separately so you know where your compliance exposure sits.
Compare approval rates between AI underwriting and your experienced underwriters on the same test setintermediate
Run 50 applications past both your best underwriter and your AI model. Note where they agree and disagree. Large gaps suggest the model is not capturing the contextual factors your underwriter uses.
Test your AI model on historical applications from 5 years ago to see if it would have made the same decisionsadvanced
If the model recommends different underwriting for identical older cases, it may have learned temporal bias rather than lasting risk patterns. This flags overfitting to recent data.
Create a hold-out group of high-value or complex applications that never go to the AI modeladvanced
Use your best underwriters only on these cases. After one year, compare claims performance against the AI-processed applications. This gives you a real control group to measure whether human judgement adds value.

Protect Claims Assessment Discretion Where Ambiguity Exists

Flag all claims where the AI recommends denial but multiple policy clauses could be interpreted either waybeginner
Ambiguous claims language is where human discretion matters most. AI tools like Guidewire may score a claim as likely fraud or non-covered based on pattern matching. Your assessor must review these before denial, because the policyholder deserves a fair interpretation.
Require a human sign-off on any AI-recommended claim denial for customers over 10 years tenurebeginner
Long-standing policyholders have built trust with your organisation. Denying their claim based on an algorithm creates reputation risk and regulatory scrutiny. A human assessor can weigh the relationship value against the technical denial.
Track how often your claims staff override AI recommendations and log the reasonsintermediate
If assessors overturn the AI 15 percent of the time, the AI is not ready for full automation. If overrides drop below 2 percent, you may be losing the human checks that catch errors. Regular review of override patterns shows whether your team still has meaningful discretion.
Do not use fraud detection AI scores as sole evidence to deny a claimbeginner
AI fraud tools like IBM Watson flag patterns but cannot prove intent. Combine the AI score with evidence of deliberate deception. A pattern match is not proof. Your team must explain the actual fraud mechanism to the customer.
Audit 50 claims that the AI recommended approval on to check for missed red flagsintermediate
The opposite of over-denial is under-payment. If your AI is too permissive, you are bleeding money. Have experienced claims staff review approved claims for quality. This protects your reserves and shows whether the model is being manipulated by applicant data quality.
Create a quarterly report showing the AI claims decision accuracy broken down by claim type and amountintermediate
High-value claims (over GBP 50,000) may need different thresholds than small claims. An AI model trained on all claims equally may make poor choices on outliers. Segment your accuracy review so you catch where the model performs worst.
Require narrative explanation from assessors before closing any AI-recommended claim denialadvanced
Force your team to articulate the actual reason for denial in plain language. If an assessor cannot write a clear explanation, the denial is not solid enough to withstand complaint or regulatory review.

Maintain Regulatory Compliance and Fair Decision Explanation

Map every input variable your underwriting AI uses back to the actuarial principle it is meant to testintermediate
If your model uses postcode, it must map to a measurable risk factor (claims frequency, severity). If postcode is just a proxy for wealth or ethnicity, your regulator will reject it. Document the actuarial reasoning for every input.
Write a plain-language summary of how your AI underwriting model works for your compliance and FCA filesbeginner
Your model may be complex, but regulators and complaint handlers need to understand it. Simplify the logic for someone who is not a data scientist. This summary becomes your legal defence if a customer challenges a decision.
Test whether your AI model treats similar applicants the same way regardless of how their data is enteredadvanced
If an applicant's age is entered as date of birth versus years old, does the model score them differently? Inconsistency suggests the model is brittle and may be vulnerable to discrimination claims.
Run a fairness assessment at least twice per year comparing AI decisions across demographic groupsintermediate
Use a tool like AI Fairness 360 or your vendor's built-in fairness check. Document the results and any corrective actions. This is now standard regulatory expectation. Documented fairness reviews protect you in complaint investigations.
Do not allow your underwriting AI to use health data, criminal history, or other sensitive attributes without explicit legal reviewbeginner
Some data is permissible for insurance but only under strict conditions. Using it without a clear actuarial justification creates discrimination risk. Get your legal team to sign off on every sensitive variable before it goes into the model.
Establish a process for customers to request the reasoning behind an underwriting declineintermediate
You must be able to explain why you declined someone. If your AI cannot give a coherent reason, you have a problem. Design your AI outputs to be customer-facing from the start, not a retrofit.
Conduct annual retraining of your AI models and document what changed between versionsadvanced
Models drift over time as claims data accumulates. Retraining can introduce new biases or degrade performance on underrepresented groups. Keep a version history and test each new model against the last for fairness and accuracy regressions.

Five things worth remembering

Related reads


Common questions

Should insurances pull 100 recent underwriting decisions made by your ai tool alongside the actual claims outcomes?

Compare what the model recommended against what actually happened in claims. This shows whether the model's logic matches real risk. If the model rejected applicants for reasons that never materialised as claims, your AI may be rejecting good business.

Should insurances segment your audit results by protected characteristics (age, postcode, gender, ethnicity where you hold the data)?

Look for patterns where the AI approves or declines one group at a different rate than another. A model can be statistically accurate overall yet systematically unfair to a subset. Your regulator will ask for this analysis.

Should insurances ask your underwriters to manually assess 20 cases the ai flagged as high-risk declines?

Your underwriters have context the model lacks. They spot applicants with legitimate explanations for red flags. If underwriters overturn many AI declines, the model is missing judgement, not replacing it safely.

The Book — Out Now

Cognitive Sovereignty: How To Think For Yourself When AI Thinks For You

Read the first chapter free.

No spam. Unsubscribe anytime.