Download

The Long View on Insurance's Transformation

To understand where insurance is heading, look at the history of computing — from batch processing to today's instant-answer capabilities. 

Image
Futuristic sky

I often tell people I've been watching the same movie for decades — it will be 40 years this fall since I started covering IBM as a young pup of a reporter at the Wall Street Journal. I've watched the disruption that hit IBM spread to the rest of the computer industry, then to commerce in general, thanks to the personal computer, internet, search engines, smartphones and now AI. 

Having watched the movie so often, I have a pretty good sense of how today's story lines will play out.

Today, I'll start even earlier than 1986 and offer a quick history of computing because I think the long view provides useful perspective on where insurance is — and where it's going. Some insurance processes are firmly stuck in the 1950s and 1960s, when batch processing was the only game in town. Others have made it to the 1980s and 1990s, with their PCs and networking. Still others are becoming fully modern, as they take advantage of mobile devices and generative AI.

On the theory that every industry is becoming a technology industry, insurers will eventually catch up on all fronts. Understanding where we lag the most and imagining a world where insurance can operate at the speed of Amazon will, I hope, provide a road map that will help us get to that future faster.

So, yes, I've set myself a rather ambitious goal this week.

To understand the starting point for computing (and insurance), think of my college roommate Mike. He was a computer science major, so he was wedded to the campus mainframe. He'd type out a program on a stack of punch cards, hand them in at the window in the computer center... and wait. When his turn finally came on the mainframe, he'd get a printout with the results. Given the complexity of what he was doing, and that even a typo would derail things, he inevitably had errors. So he'd debug the program, type out some more punch cards, turn them in at the window... and wait some more. 

Because turnaround times were shorter at night, after most students had gone back to their rooms, Mike typically stayed out into the wee hours of the morning, napping on a table while waiting for his latest printout. (The way our habits meshed led to a comical relationship, where we sometimes didn't see other while both were awake for weeks at a time. I'd leave in the morning while he was asleep and, after working a job, not get back until he'd left for the computer center in the evening. He went home on weekends to see his girlfriend, so I'd sometimes find myself asking mutual friends, "Hey, how's Mike? I haven't talked to him in ages. Tell him I said hi.")

Mike's travails were a holdover from the era of batch processing, when a computer could do only one thing at a time. Big efforts, such as processing payroll or reconciling accounting records, were done in a single batch at a time reserved on the mainframe. Mike's programs obviously weren't on anything like accounting's scale, but he still had to run a program in a single batch of cards and wait his turn. 

Even though computing technology has improved by orders of magnitude since Mike and I were in college, a lot of business still operates at the speed of batch processing. You have a meeting on some issue, and a question comes up. Someone is assigned to do some analysis and comes back a week or two or three later with an answer. The issue is discussed again, and another question arises. More analysis over more weeks ensues. The batch processing influence is even stronger in insurance than in most industries because there is so very much data to analyze.

Computer scientists saw early how much better interactive computing would be and spent decades getting us there. By the '60s and '70s, time-sharing became possible. The setup was awkward: You had a keyboard and printer but had to type out a program on special tape that you fed into the machine, and turnaround times were painfully slow because you were queueing up behind all the programs running on a distant mainframe or minicomputer. But time-sharing spread the power of computing far beyond the walls of the data center. (Bill Gates got his career started on a time-sharing terminal at his high school. I, too, had access to a terminal in high school but somehow didn't do as much with it as he did. Alas)

By the late 1970s and into the 1980s, Xerox PARC had worked its magic, and the Apple II and then the IBM PC were putting real power on individuals' desktops. The computers delivered big benefits to business because of the electronic spreadsheet but otherwise proved to be rather limited when used in isolation. Fortunately, Xerox took care of that issue, too, with the Ethernet networking standard that let businesses link their in-house computers. Then the internet took networking into the stratosphere thanks to the World Wide Web's debut in 1989 and the Mosaic browser in 1993. By the late 1990s, search engines were doing a good job of fulfilling Google's goal "to organize the world's information and make it universally accessible and useful." Then smartphones, led by the iPhone debut in 2007, put all the computing power and information in our hands. Generative AI is now letting us gather, process and use far more of the world's data than we humans could ever do on our own.

Big tech has taken advantage of the remarkable progression of technology to gather all sorts of signals about individuals (many of which I wish they didn't have) and target us with ads, with memes that keep us engaged, with dynamic pricing that maximizes their clients' revenue. Progress in other spheres is more uneven, but you can look at big retailers like Amazon and Walmart and see how they sense demand and respond to it in real time.

I'd say insurance has done a so-so job of taking advantage — acknowledging that our situation is complicated by heavy regulation and by the confusion of state-by-state oversight in the U.S. A lot of insurance work is still in a sort of batch mode — the analysis of loss runs, actuarial tables, and so on. While insurers have taken advantage of all the power on the desktop that PCs provide, I'm not sure we've done the best job of internal networking — why, for instance, isn't claims data always fed in real time to underwriters to inform future decisions? Insurers certainly haven't been great about taking advantage of all the information that's out there beyond their four walls; they're starting to figure out what data to trust and how to absorb it, but they've been slow. Insurers are also still figuring out what to do about smartphones. Yes, every company has an app these days, but my impression is that customers still want to be able to do a lot more self-service via phones than is possible today.

I'll withhold judgment on how insurance is doing on gen AI. We're headed in some good directions by gathering and doing initial processing for those in claims, underwriting and agencies, but we clearly haven't figured gen AI out — yet nobody has, so we're in good company. 

The nice thing is that, whatever our inadequacies to this point,  our version of the technology movie can have a happy ending for two reasons. One is that any new computer technology builds on everything that's come before in an exponential way. We're not just adding a gen AI capability alongside an information or networking capability. The capability increases by some exponent what was ushered in by smartphones, which raised what came before to some power, after it did the same to everything that came before that. The second reason is that we don't have to build the capability. The tech giants have done that over the past 75 years; we just have to take advantage. They're not done yet, either: The latest figure I saw is that the five biggest AI companies are investing $700 billion on infrastructure in this year alone

To me, the happy ending will come in a decade or so, when insurance can fully switch from batch processing to what I think of as conversational computing. You don't have a question in a meeting and send someone off to study the issue for weeks. You ask a question, and your AI uses all the internal and external information available to provide an answer. Loss runs and actuarial tables don't require massive studies. You converse with your computer and get the answers you need.

You can see glimmers of this sort of conversational future in some things going on today. Continuous underwriting is one great example. Why wait for an annual review of a policy when aerial imaging can tell you that a homeowner has added a pool, when an AI monitoring the internet can tell you that a restaurant has added a drinks menu or delivery options, etc.? Why not take advantage of the ability to sense what's going on among clients and prospects and respond? 

Embedded insurance is another example. Why should selling an insurance policy always be a formal project? Why not just use the ability to sense when a customer might want coverage and respond?

Technology never stops moving. Moore's law made sure of that for decades, with what became a sort of mandate for semiconductor makers to double the power of a chip every year and a half to two years at no increase in cost, and other forces, such as AI, are now amplifying those gains in capability by orders of magnitude. I figure I've gone through six tech revolutions since I debuted on the computer beat in 1986, and we could be in the middle of the next one, with agentic AI.

For insurers, I hope a look at the history of computing identifies some spots where we can and should improve. But I mostly hope the history shows us that we're headed toward a conversational future, where we ask questions and get answers in real time — and hope insurers will construct road maps toward that future so every incremental decision on IT can keep us moving in the right direction. Just imagine what insurance could look like at the speed of Amazon.

Cheers,

Paul

 

Traditional Insurers Can Still Win AI Race

Incumbents have operational context advantages AI-native startups can't replicate, but the window to leverage them is closing.

Side profile of a human's brain and face outline against a blue background with gears

Recently, there's been talk from AI-native insurance startups telling incumbents they'll never catch up. The argument goes like this: The barrier isn't technology; it's organizational DNA. Boards resist. Agent networks resist. Incentive structures resist. Even superintelligent AI can't rewrite a captive distribution network or a CEO's risk tolerance.

We built one of those AI-native insurers. We've spent nearly a decade learning where AI actually works in insurance - and where it doesn't. So we'll say what most people in our position won't: 

The critics are only half right.

The organizational immune system is real

We've watched it operate from the inside.

AI threatens more than processes. It threatens people, hierarchies, and decades of institutional knowledge that leaders built their careers on. The more powerful the technology gets, the more threatening the disruption feels, and the harder the organization pushes back.

The execution gap is genuine, too. Deloitte surveyed 3,200 enterprise leaders this year and found that executives feel strategically ready for AI but not operationally ready. Every insurance business we talk to confirms this. The board said yes. The pilot worked. But not much actually changed. They tripped in the last mile.

If you're reading those blog posts and feeling uneasy, trust your instincts. Standing still is falling behind.

Where the thesis breaks

The "incumbents are dead" argument assumes the only way to win with AI is to have been born with it. That organizational barriers are permanent. That traditional insurance businesses are evolutionary dead ends waiting for the asteroid.

This confuses two problems.

The first is building AI technology. AI-native startups have a real advantage here. Clean architectures, ML engineers who learned to work alongside actuaries, feedback loops from day one.

The second is having the operational context that makes AI actually work in insurance. Here, traditional businesses have an advantage no startup can replicate.

A startup can build a great claims model. But it doesn't know that your Florida team handles litigation differently than your Texas team because of venue-specific judicial considerations. It doesn't know that your underwriting knowledge base says one thing but your senior underwriters do another - and the deviation is actually producing better results. It doesn't know which of your 50 state regulatory constraints are real compliance requirements and which are institutional habits nobody has revisited in a decade.

That operational context - the messy, human, state-by-state reality of how insurance actually works - is the raw material AI needs to generate value. Technology is the engine. Context is the fuel. Insurance businesses have been accumulating this fuel for decades.

The startup pitch is: "We have the engine, and we'll figure out the fuel." The honest answer is that the fuel is harder to build than the engine.

The real question is speed

Can you close the execution gap before it shows up in your results?

The gap closes by connecting AI to the operational reality of how your business actually runs - across claims, underwriting, distribution, and compliance - in ways that compound over time.

Every month of operational AI data makes the system smarter. Every feedback loop accelerates the next one. This is an exponential curve, not a linear one. The businesses that start building now aren't just catching up. They're beginning a compounding process that gets harder to replicate with every cycle.

We spent nearly a decade building these feedback loops inside our own company. That experience made one thing clear: The distance between an AI demo that works and an AI system that changes how you operate is almost entirely about understanding the insurance underneath.

What I'm telling insurance executives right now

Your data is an asset that will appreciate with use. Your operational context can be youradvantage. The AI-native startups telling you it's over are talking their own book.

Some businesses already know this. The ones investing seriously in operational AI - not pilots, but production systems touching real policyholders - are proving the thesis wrong in real time.

We're seeing this from carriers, MGAs, and specialty businesses alike.

But the window is real. AI feedback loops compound. The businesses that start building them in the next 12 to 18 months will pull away from those that don't. You'll see it first in expense ratios, then in loss ratios, and then in competitive position.

The businesses that win won't become AI companies. They'll stay insurance companies that figured out how to make AI compound inside their operations before the window closed.


Kyle Nakatsuji

Profile picture for user KyleNakatsuji

Kyle Nakatsuji

Kyle Nakatsuji is the founder and CEO of Clearcover, an AI-native auto insurance carrier, and Dearborn Labs, which helps P&C carriers and MGAs operationalize artificial intelligence. 

Before founding Clearcover, he was a venture investor at American Family Insurance, where he led insurtech investments. He speaks regularly on AI strategy in insurance.

Smoother Insurance Agency Succession Planning

Most agents delay succession planning. The smoothest agency transitions start with technology-enabled operations built from day one.

Abstract Pattern on a Wall

For independent agents, the to-do list never gets shorter. New clients to win, policies to place, and revenue to grow. But there's one conversation that doesn't always make it onto the planning agenda, and it might be the most important one of all. What happens when it's time to hand things off?

Succession planning has long carried a reputation as something to worry about later. A conversation for agents nearing the end of their career, not those in the thick of building their business. But that thinking can be costly. The agencies that make the transition most smoothly aren't the ones that started planning at the last minute. They're the ones that built transferable tech operations from day one.

Here's the good news: if you're already using an agency management system to run your daily operations, you're likely closer to succession-ready than you think. The tools that help you manage client account data, track performance metrics, and stay on top of renewals can do double duty. Used consistently, they build the kind of organized and documented operation that makes handing things off far less daunting.

Performance Metrics Tell Your Agency's Story

When it comes time to demonstrate value, data speaks louder than anything else. Potential successors and buyers will want a clear picture of your agency's performance, including which lines of business are driving the most revenue, which producers are performing, and where coverage gaps exist across the book. Those answers need to be readily accessible.

A robust agency management system gives you this performance visibility in real time. Dashboards and reporting tools surface the metrics that matter most, from total annualized premium and active policies per customer to a detailed breakdown of your book of business by transaction type, often presented in intuitive visual layouts.

You can customize these reports too, filtering and drilling down into the data points that matter most. Some systems even let you benchmark your performance against peer agencies, giving you a clearer sense of where you stand. That kind of insight doesn't just serve a future transition. It sharpens your decision-making today, helping you spot growth opportunities and course correct before small issues become bigger ones.

Over time, these reports build a compelling picture of your agency's health and trajectory, one that tells a clear story to a successor and makes you a stronger agency today.

AI Keeps Client Knowledge Transferable

Serving clients without missing a beat is one of the first challenges any incoming leader faces. That means being able to find policy details quickly, understand coverage history, and get up to speed on the relationship without having to track down the person who used to handle it. When information is scattered across inboxes, desktop folders, and spreadsheets, that handoff becomes harder and more costly than it needs to be.

AI-powered agency management tools change that, and not just when a transition is on the horizon. Picture this: a newly onboarded staff member pulls up a long-standing client account in their first week. Rather than digging through months of email threads and agent logs, they get an instant summary of the relationship, enabling more knowledgeable client interactions and a much faster path to getting up to speed.

Clients expect continuity. They don't want to repeat themselves or re-explain their history with their agency. They expect whoever picks up the phone to already know them. AI makes that possible whether you're onboarding a new hire, navigating a leadership change, or simply trying to deliver a better client experience every day.

Renewal Tracking Protects What You've Built

Retention is the metric that tells the clearest story about an agency's health. A consistent renewal process signals that clients are being taken care of and that the book of business is stable.

A good management system gives you and any future leader a single view of every upcoming renewal, what has changed between the current policy and the renewal offer, and which clients are most at risk of shopping around. Predictive analytics flag at-risk policies before they become problems. Automated remarketing workflows retrieve updated rates and surface them alongside renewal details, so whoever is managing the book can act quickly and make informed recommendations.

For a buyer or successor, a clean and consistent renewal process is one of the most compelling things they can walk into.

No Matter Where You Are in Your Career — Start Now

Whether you're just launching your agency, in the middle of a growth run, or beginning to think seriously about the future, the time to invest in technology-enabled workflows is now.

The efficiencies you gain today will compound over time, and when the moment comes to pass the pen and the policies, you'll be glad you started early.


Rob Bourne

Profile picture for user RobBourne

Rob Bourne

Rob Bourne is the senior vice president and general manager of EZLynx

He previously served as SVP at Applied Systems, overseeing inside sales, account management, business development, and alliance partnerships. Before that, he held senior roles at Athelas and Podium. 

He has an MBA from Cornell University.

Why the Customer Experience Still Fails

When advisors and agents spend time hunting for information, switching screens, or reentering data, customers feel the friction immediately.

Robot and Human overlook hologram projection

Walk into any bank branch, insurer's office, or contact hub, and you'll see the same challenge playing out: frontline teams juggling conversations, systems, and expectations in an environment where customers expect answers now -- and on the channel of their choice.

Consumers today benchmark banking, financial services and insurance (BFSI) service not against competitors but against the best digital experiences they see anywhere. The organizations that succeed in 2026 won't simply add more tools; they will blend human advice with smart, invisible technology, so their advisors, agents, relationship managers, and service reps can engage faster, personalize better, and build trust at scale.

What Customers Want (And Where Organizations Still Fall Short)

Customers want digital convenience, but they also want expert human guidance for financial decisions that carry risk:

  • Choosing an insurance plan
  • Applying for a loan
  • Completing know-your-customer (KYC) requirements
  • Making investment decisions
  • Resolving service or dispute issues

McKinsey's customer-experience research across financial services shows why this matters:

CX leaders achieve stronger revenue growth, lower costs, and higher employee engagement than their peers - and generate meaningfully higher shareholder returns.

But many journeys across BFSI remain fragmented.

Customers begin online, switch to chat, walk into a branch or connect with an advisor, only to repeat themselves at every step. Follow-ups may be inconsistent, and context often gets lost between channels.

Yet the human element remains critical: across financial services; trusted advisors and relationship managers consistently rank as the most valued touchpoint, so much so that switching behavior rises when advisors leave or service becomes inconsistent.

The message is universal across BFSI:

Customers don't want to choose between people and technology. They want both, seamlessly connected.

Why Frontline Productivity Is the New Customer Experience

Customer experience and employee experience are inseparable.

When advisors and agents spend time hunting for information, switching screens, or reentering data, customers feel the friction immediately.

McKinsey finds that organizations that engage frontline teams as co-creators in CX transformation see:

  • Higher customer satisfaction
  • 20% improvement in employee satisfaction
  • Faster turnaround time
  • Stronger cross-sell and retention

In BFSI, this link is especially clear:

  • A banker who has instant access to a customer's full financial picture can advise better.
  • A field officer with mobile access reduces turnaround times.
  • An insurance advisor with contextual prompts can personalize without delay.
  • A contact center agent with unified history avoids escalations.

Technology doesn't replace expertise - it amplifies it.

Accenture's latest research reinforces the urgency: customers increasingly report difficulty reaching a human when needed, or navigating service journeys. In a landscape where attrition is high and expectations are rising, BFSI organizations must remove friction, not add new layers of complexity.

Where AI Fits: Practical, Not Hype

AI's role is not to automate away financial conversations - it is to shorten the distance between a customer question and a confident, compliant answer.

IBM's Institute for Business Value says that:

75% of financial services executives believe AI will improve personalization and CX but warn that fragmented systems and weak data foundations remain barriers.

The promise of AI becomes real when paired with clean data, integrated workflows, and empowered frontline teams.

In BFSI, this looks like:

1. Advisor / agent assist

Real-time summaries, next-best-actions, explanations of product rules, contextual prompts - helping frontline teams spend more time advising and less time searching.

2. Smarter routing

AI triages inbound requests to the right channel:

  • Self-service for simple tasks
  • Human support for complex or high-value conversations

This reduces drop-offs and improves first-contact resolution.

3. Engagement

Detecting repayment cycles, renewal windows, eligibility changes, life events, and portfolio gaps - then nudging advisors with timely reasons to reach out.

4. Faster, cleaner onboarding

AI-assisted KYC checks, document classification, real-time validation, multilingual support, especially important for banks, insurers, and non-banking financial centers (NBFCs).

Designing the "Human + Tech" Operating Model Across BFSI

To make this transformation real, BFSI leaders are anchoring around a few core principles:

Human-led, tech-accelerated journeys

Map the moments where customers need reassurance (loan approvals, investments, claims, disputes). Then use tech to handle the rest - context-gathering, data prep, routing, and follow-ups.

Modernize the service core

Accenture's "Service on the Brink" findings highlight a striking gap: customers don't feel technology has meaningfully improved service quality yet.

That gap closes only when:

  • Data is unified
  • Systems are connected
  • AI is grounded in real rules
  • Handoffs are seamless

Measure what matters

Beyond cost KPIs, leading BFSI organizations track:

  • Advisor time saved
  • First-contact resolution
  • Customer effort scores
  • Quality and consistency of advice
  • Employee adoption and satisfaction

AI and automation must serve people, not distract them.

The New Standard of Service in BFSI

Picture a loan customer messaging a query.

A virtual assistant validates identity, surfaces repayment options, and flags eligibility for a top-up offer.

The customer asks a nuanced question; within seconds a relationship manager joins, armed with a clean summary of past interactions, risk profile, and suggested talking points.

The conversation finishes quickly, the customer feels understood, and the advisor moves confidently to their next engagement.

This is where BFSI is heading:

Frontline expertise supported, not overshadowed, by technology.

When AI, automation, and connected systems fade into the background, advisors and agents can lead every interaction with clarity, empathy, and insight.

That's the real frontier.

References

● McKinsey & Company. “Elevating customer experience: A win–win for insurers and customers,” 2023.
● IBM Institute for Business Value. “Insurance in the AI era,” 2025.
● Accenture. “Customer Service on the Brink,” 2025.
● McKinsey & Company. “The Human Advantage in Banking Customer Experience,” 2023.


Neeraj Malhotra

Profile picture for user NeerajMalhotra

Neeraj Malhotra

Neeraj Malhotra is CEO of AccelTree.

AccelTree focuses on enabling insurers to modernize distribution, improve agent and customer experience, and operationalize compliance, so insurance can shift from a product-centric industry into a responsive ecosystem led by experience.

Insurance Is Learning a Legal Lesson

For decades, insurance professionals could lean on muscle memory. But the environment has changed. Decisions must now be documented, explainable, and consistent over time.

View of Street from a Glass Window

In the legal profession, the work is only as strong as its support. A good argument isn't just persuasive, it's backed by citations. You can point to the contract clause, the case, the exhibit, and the chain of reasoning that got you to the conclusion. That's not academic formality. It's how legal work survives scrutiny in a court of law. 

Insurance is moving in the same direction.

For decades, insurance operations could lean on experience and muscle memory. A tenured underwriter knew what "this form usually covers." A claims leader knew the standard response posture. A broker knew which markets were flexible. That knowledge still matters, but the environment has changed. Regulators, legal teams, and procurement groups now expect decisions to be documented, explainable, and consistent over time.

"Trust me" is no longer a reasonable operating model.

Why legal workflows look the way they do

Legal work is predicated on the ultimate potential that it will end up in front of a judge. This fear shapes all the work lawyers do. 

A motion, an opinion letter, or a contract position might get scrutinized months or years later by a judge, with millions of dollars at stake. The only way to truly prepare for that situation is if the work product is structured to be audited. That's why legal workflows emphasize three things:

Citation. Show exactly where the claim comes from.

Reasoning. Make it possible to retrace the reasoning steps from source to conclusion.

Conclusion. Make it easy for another expert to validate or challenge the conclusion by having a very clear and articulate conclusion.

These practices aren't about slowing work down. They're how the legal industry moves quickly while staying defensible when the stakes rise.

Insurance is discovering the same truth, especially in claims and coverage interpretation.

Insurance is already under similar scrutiny

Insurance has always been regulated, but scrutiny is broader now and comes from more directions, including clients. Decisions can trigger omissions, bad-faith allegations, and liabilities that far exceed a coverage dispute.

State-by-state variability adds another layer. A defensible decision in one jurisdiction may be incomplete or risky in another.

At the same time, the work is deeply document-driven. Policies, endorsements, submissions, claim files, and correspondence are still stored in PDFs, scans, and formats that were created for manual review.

That means insurance decisions are often anchored in unstructured language that must be read carefully, compared across documents, and defended later.

In short, insurance faces legal-like constraints whether it realizes it or not.

The AI factor raises the bar, not just the speed

AI is often discussed as a productivity lever, but in insurance, the real challenge is credibility. When an AI-supported decision gets scrutinized, you need to show the basis for it. If the answer is a black box, you've created a new type of risk.

That's why the industry is increasingly prioritizing accuracy, explainability, and consistency over speed alone.

It's also why "model drift" matters. If a tool's behavior changes over time, it undermines consistency and auditability in regulated workflows.

This is one place where legal has a head start. Many legal technology workflows were designed around precedent and review. The focus is less about generating text and more about interpreting documents with citations and a clear path from source to conclusion.

Insurance now needs the same.

The future of insurance work

This shift isn't theoretical. It changes how teams should define quality.

In an insurance workflow built to withstand scrutiny, a good outcome isn't only correct—it's defensible. That means:

Your conclusions should point back to the policy language. Coverage positions and claim decisions need to be anchored in the actual text, not just institutional memory. Insurance has long depended on expert judgment in how professionals read policies, interpret exclusions, and apply precedent. The next step is making that judgment visible and reproducible.

Your reasoning should be transparent enough for peer review. If a colleague can't follow how you got there, an auditor or regulator won't either. Transparent reasoning isn't a luxury in high-stakes decisions, it's a requirement.

Your process should be consistent across teams and time. Insurance is full of niches and specialized expertise, but inconsistency is costly. As experienced practitioners retire, decision quality can decline if judgment stays trapped in individuals. Understanding the policy is the hardest problem in insurance. You can't solve it with expertise that only exists in people's heads.

Your documentation should be structured for "future you." Claims files, coverage analyses, and underwriting notes should read like work that's expected to be examined later. That's the legal mindset, and it's becoming the insurance mindset.

This is also why many leaders are talking less about flashy AI and more about repeatable operating models. The most valuable AI in insurance will look like consistency, documentation, and fewer surprises.

The practical takeaway for insurance leaders

If you're leading claims, legal, compliance, or brokerage operations, the question isn't whether your teams will use AI and automation. They already are. 

The real question is whether you're building systems that will hold up when scrutiny arrives. That means setting standards for citation, traceability, and reviewability in the work itself, not as an extra step at the end.

It also means resisting tools and processes that optimize for speed while sacrificing transparency. In insurance, "close enough" is often where the risk begins. 

Insurance is learning a lesson the legal industry learned long ago: when the stakes are high, the work must show its sources.


Dan Schuleman

Profile picture for user DanSchuleman

Dan Schuleman

Dan Schuleman is the co-founder and CEO of Qumis, a lawyer-built, AI-powered insurtech helping insurance professionals read and interpret policies. 

Before founding Qumis, he was associate general counsel at Kin Insurance. He previously practiced insurance coverage law at Am Law 200 firms.

He holds a J.D. from the University of Illinois College of Law and a B.A. with honors from Northwestern University.

How Property Carriers Can Scale AI

Property carriers face a critical gap between AI vision and execution as they work to scale automation across claims workflows.

An artist's illustration of AI

The AI market in the insurance industry is set to hit $80 billion by 2032, yet nearly two-thirds of carriers have a gap between their AI vision and reality. In essence, carriers understand what they want from AI and see the significant value it can drive but are struggling to actually put this into practice, especially in the property sector.

Marrying vision and reality is critical, however, as carriers look to scale automation, re-orchestrate and transform workflows, and fully realize the potential of AI across the lifecycle of a claim. There is a substantial downstream impact on claimants and clients, as well. 

There are a number of steps a carrier needs to take to find success with AI in 2026 and ensure they are future-proofed and ready for the next era of property claims.

The AI Journey for Property Carriers

First and foremost, carriers need to understand where they are on their AI journey. This is important in identifying what the next steps are, what's feasible in the short term, and eventually in the long term, and aligning company communications, operations, workflow, training, and more.

At the moment, anywhere from 58-82% of carriers are leveraging AI tools in their operations, but only 12% claim to have fully mature capabilities, and only 7% have achieved scalable AI success. This means that 93% of carriers are still in the part of their AI journey in which they are identifying how to scale AI to a point at which it is driving real, measurable outcomes. What we've seen so far is that adoption of AI has been popular in areas such as intake, triage, and documentation, but fully integrated technology and end-to-end AI workflows are still far away for most carriers. This, in turn, results in a fragmented technology experience, rife with different tools, vendors, and solutions. This limits AI's impact. It keeps value confined to each step in the lifecycle of a claim, can lead to inconsistent data or silos across systems, and weakens output.

Reliance on pilot programs or point solutions is the first step in an AI journey, but it certainly can't be the last. Most importantly, this technology is rapidly advancing, and the longer it takes carriers to find value and scalability, the further they'll fall behind the competition.

The Challenges Facing Carriers

There are three main challenges facing carriers. First is integration of AI into legacy technology. The majority of claims systems weren't built with API connectivity in mind, which introduces difficulties immediately into scaling this technology across workflows. Before integration even begins, carriers need to ensure that their claims systems can support orchestration.

Second is training a disconnected property workforce. An often-overlooked aspect of AI in the property space is preparing for the challenges that can arise when adjusters are managing heavy caseloads and working in the field. Support systems are critical to success in this area, and AI and any other new tools cannot feel like a burden to them. Training and communication in best-use cases are important in presenting these tools as benefits. This can be streamlined through rollout plans that align with day-to-day workflows, prioritize flexibility, and implement continuing training opportunities.

Finally, expecting AI tools to drive perfection is a key challenge. This technology won't deliver perfect outcomes from day one, but through gradual improvements can drive real change in processes. If too much focus is placed on perfection, then widespread implementation can be delayed. Instead, carriers should prioritize progress first and perfection second when measuring AI against real-world baselines, with the goal of refining capabilities over time.

What Real Impact Can Look Like

When challenges are addressed and overcome, and carriers understand how to progress from point A to point B in their AI journey, real impact can be achieved. This will be seen in a number of ways across operations.

Main points of impact will be evident in faster claims reviews in which AI is helping adjusters summarize claims, extract data, and capture notes more efficiently, as well as in improved program outcomes, smoother workflows across internal and external systems, and smarter claim routing. AI tools can evaluate loss severity, complexity, and fraud risk at intake, assisting in routing claims to the right recovery resource sooner.

In addition, adjusters will see stronger field operations through enhanced drones, sensors, and tablets, which enable faster mitigation, better assessments, and quicker resource mobilization.

The data backs up this impact. Intake automation has reduced average claim processing from 10 days to 36 hours, AI photo analysis boosts claim handling efficiency by 54%, and much more.

Prioritizing Human Expertise

A key consideration in the integration of AI is the increasingly important role that humans play, and will continue to play, in this process. AI should be seen as a tool to augment and support workforce expertise, rather than replace it.

This technology is powerful, but cannot be used to replace human judgment, empathy, or real-world experience in the claims process. Losses can be ambiguous, emotionally sensitive, or require nuanced, complex coverage decisions that AI cannot handle, and require human professionals to consider context, communicate clearly, and advocate for policyholders throughout each stage.

In essence, AI is a catalyst, not a cure-all, and carriers must aim to apply AI selectively while keeping people at the center of their claims decisions. Striking this balance will be the difference in staying ahead and remaining competitive in a rapidly changing technological and regulatory landscape.

For Sedgwick's full report on "Future-Ready Property Claims," click here.

Group Benefits Enters Decisive Phase

Platform consolidation among carriers usually promises modernization, but group benefits relies on "frankenstacks," so merging may deliver rigidity when adaptability matters most.

Blue and Purple Design

Consolidation is often presented as progress. In group benefits today, it may prove to be the opposite.

As consolidation continues to ramp up across the insurance technology landscape, mergers and acquisitions are being framed as a way to deliver broader capability, stronger platforms, and more complete ecosystems. For carriers under pressure to modernize quickly, the logic is appealing. Consolidation promises a larger vendor with a fuller suite of functionality that should be better equipped to support long-term transformation.

But for group benefits insurers, in particular, this assumption deserves scrutiny. Behind many of today's platform mergers lies a perfect storm. One that risks locking carriers into greater rigidity at precisely the moment they need to become more adaptable.

Building on a Fragile Foundation

To understand why platform consolidation can be problematic, it helps to start with the technology foundations many group benefits platforms already rest on.

Unlike some other insurance lines, group benefits technology did not evolve over decades on stable, purpose-built architectures. Much of it emerged more rapidly, often adapted from adjacent markets such as life, pensions, or individual products. Vendors re-engineered existing platforms to meet growing demand for employer-sponsored benefits, then layered on new functionality as customer expectations evolved.

Over time, these platforms became highly customized to individual carriers and employer requirements. New features were added to meet immediate needs. Integrations were built to support emerging distribution and service models. Documentation rarely kept pace with delivery pressure. What began as pragmatic adaptation gradually accumulated into significant technical complexity.

In other words, many group benefits platforms entered the current consolidation wave already carrying structural fragility. At the same time, market expectations have accelerated. Employers and employees increasingly expect flexible benefit design, digital enrollment, ecosystem integration, and personalized experiences. Delivering on these expectations requires technology that can be configured and extended quickly. Not simply maintained.

This is the environment into which consolidation has arrived.

The Consolidation Illusion

When technology vendors merge, the narrative is straightforward. Customers of the acquired platform are told they are becoming part of a larger, more advanced organization with greater investment capacity. Customers of the acquiring platform are told they will gain new functionality and broader capabilities. Both sides expect improvement.

In reality, consolidation is often driven first by market share and coverage, and only second by technological unification. This is not the result of poor intent. Vendors pursue acquisitions because they believe it is the fastest and most cost-effective way to fill capability gaps and respond to market opportunity. Building new functionality from scratch is expensive and time-consuming. Acquiring it appears faster and less risky. Larger scale also reassures risk-averse insurers, who often prefer established vendors with financial strength and broad offerings.

On paper, the logic holds. In practice, the technical challenge of integrating two heavily customized, architecturally distinct platforms is frequently underestimated. Particularly when decisions are driven primarily by commercial leadership rather than engineering reality.

From Safety Blankets to Patchwork Quilts

Every mature insurance platform reflects years of client-specific configuration, integration, and adaptation. No two are the same. Data models differ. Product logic differs. Workflow structures differ. Some critical functionality may exist in legacy code written decades earlier and never fully documented. Each system has evolved around the needs of its existing customers.

When two such platforms are combined, true unification requires deep re-engineering: rationalizing data structures, redesigning core services, and often rebuilding significant functionality. This is expensive, disruptive, and difficult to justify commercially. As a result, most merged platforms evolve through accommodation rather than transformation. New layers are added. Interfaces are built. Functionality is duplicated rather than consolidated.

The result is what many in the industry privately recognize: a frankenstack. Or, more accurately, the merging of two frankenstacks.

Over time, more and more IT investment is directed toward supporting this complexity rather than advancing capability. Roadmaps slow. Innovation competes with maintenance. What was intended to be a safety blanket for customers becomes a patchwork quilt that grows heavier and harder to adapt.

The industry has seen versions of this story before. Large technology estates built primarily through acquisition can become extraordinarily difficult to modernize, leaving both vendors and their customers managing accumulated complexity for years afterwards. Consolidation promises acceleration. Too often, it results in gradual technological stagnation.

Why This Matters in Group Benefits

Group benefits carriers are particularly exposed to this dynamic because their business demands constant configuration and change.

Benefit structures vary by employer. Employee expectations continue to evolve. New partnerships and services must be integrated rapidly. Distribution and engagement models are shifting toward more digital, personalized experiences. Technology must support continuous adaptation rather than periodic transformation.

Yet consolidation often redirects vendor focus inward. Engineering effort moves toward integration of acquired platforms and preservation of existing revenue streams. Transformation initiatives slow while complexity is stabilized. For customers, the experience can be subtle but significant: fewer meaningful enhancements, slower responsiveness to new requirements, and growing difficulty introducing new products or partnerships.

The risk is not immediate disruption. Most consolidated platforms continue to function adequately. The real danger is long-term loss of adaptability and a slow erosion of the ability to respond as the market evolves.

AI Raises the Stakes Even Further

If adaptability was already becoming the defining factor in group benefits, the rapid evolution of AI has raised the stakes significantly.

There is growing consensus across the industry that AI will transform underwriting, claims, service, and product design. But meaningful AI adoption is not achieved by layering point solutions onto legacy cores. Injecting AI "at the edges" of rigid platforms may create isolated efficiencies, but it does not fundamentally change how the business operates.

To unlock AI's full potential, insurers require something far more foundational: open, data-fluid architectures where operational and analytical data are unified; governance and controls are embedded by design; and experimentation can occur safely within defined guardrails. AI needs to sit at the core of the platform, not be welded onto the perimeter.

In consolidated environments built from multiple legacy estates, this becomes extraordinarily difficult. Data models remain fragmented. Core logic is tightly coupled. Every meaningful change requires an IT project, often with significant coordination across integrated systems. Instead of enabling experimentation, the architecture restricts it.

The consequence is subtle but powerful. Rather than allowing business teams to test new processes, deploy new journeys, or refine models quickly, innovation becomes dependent on complex technical programs. What should be controlled experimentation turns into multi-quarter initiatives. Governance becomes reactive rather than embedded. AI becomes a feature to manage, not a capability to leverage.

As group benefits carriers look to modernize, the question is no longer simply whether a platform can support today's products. It is whether it can support continuous experimentation and governed AI-driven evolution. In this context, architectural rigidity is not just a technical limitation, it is a strategic constraint.

Challenging the Plumbing Assumption

Part of the issue lies in how technology decisions are evaluated. Platform choices are sometimes treated as infrastructure decisions, where scale and vendor stability appear more important than architectural flexibility. There can be an implicit assumption that technology is interchangeable. That one platform can be merged into another without fundamentally altering its capacity to evolve.

But insurance technology is not generic plumbing. Every system reflects years of bespoke configuration and embedded business logic. Integrating two such environments is not a simple exercise in connection; it is a complex process of reconciliation that shapes what can and cannot be changed in the future.

As group benefits enters a period of accelerated transformation, that distinction becomes critical.

A Unique Opportunity to Reassess

For insurers whose technology partners are entering a merger or acquisition phase, this is not necessarily a cause for concern. Consolidation can deliver benefits when approached with architectural care and sustained investment. Consolidation is also a natural moment to reassess.

Carriers should seek clarity on how platforms will actually be integrated, where investment will be directed, and how innovation roadmaps may change during the process. Critical questions include: Will resources be focused primarily on maintaining and connecting existing systems, or on enabling new capabilities? How will duplicated functionality be rationalized? What will this mean for the speed of change over the next three to five years? Most importantly, will the combined platform become easier or harder to evolve?

The group benefits market is entering a decisive phase. As existing technologies approach end of life and expectations continue to rise, adaptability will define competitive advantage. Encouragingly, the technology now exists to support more flexible, staged transformation, allowing carriers to modernize incrementally rather than through high-risk, "big bang" replacement.

Consolidation may expand capability on paper. But in a market defined by constant change, it is adaptability, not scale, that will ultimately determine who wins.

Independent Agencies' Top Priorities for 2026

As carrier appetites shift and underwriting tightens, independent agencies turn to AI automation to streamline workflows and boost operational efficiency.

Abstract Black and White Shot

Independent agencies are always looking ahead, but they're still focused on how much work it takes to place and service a policy today. Underwriting remains tight, carrier appetites change, and remarketing eats up time and resources. At the same time, clients expect clear, timely communication and fewer surprises.

In a recent Vertafore survey, independent agencies told us what will shape their operations in 2026. Three trends emerged:

  • Using AI automation to reduce manual work
  • Preparing for uncertainty in carrier appetite and placements across admitted and E&S markets
  • Prioritizing proactive, consistent client communication as a defining trait of high-performing agencies.

What connects these trends? Operational efficiency supported by the right technology.

High-performing agencies aren't trying to predict every market shift, and they're not just adding headcount to keep up with more work. Instead, they're focused on how the work gets done. The agencies that win in 2026 are using technology designed for their workflows to reduce manual effort, adapt quickly to appetite shifts, and communicate consistently with clients.

Using AI to reduce manual work in insurance agency workflows

Carriers always ask for more information, pose follow-up questions, and make adjustments as requirements change. When agencies run that work through manual steps and disconnected systems, it quickly turns into duplicate entry, rework, and resubmitting applications.

To manage this workload, agencies are turning to more AI automation because they can't scale this work by hiring alone. In the survey, nearly 30% said they expect AI-driven process improvements to deliver the strongest return on investment in 2026, and more than one-third said the greatest value will come from AI embedded into the solutions they already use.

Today, agencies are successfully using AI-powered solutions to pull data from ACORD forms and carrier documents so the information can be reviewed instead of re-entered. AI supports everyday service work, validating information before it reaches service or sales, and even determining whether it makes sense to move forward before spending time investing in a full rating.

Over time, these improvements will expand further. AI takes on more of the manual work—redundant clicks, document routing, and non-licensed processes that consume time. This automation works in the background so teams can focus on licensed work and client conversations.

Managing carrier appetite shifts and E&S placements efficiently

Carrier appetite changes are expensive when agencies aren't structured to adapt. Every shift means rebuilding submissions, re-entering information, and spending more time redoing work that was already completed.

In the survey, nearly half of agencies said they expect to place about the same amount of business in the E&S market this year, and 40% expect to place more than they did in 2025. That tells us E&S isn't going away. Agencies are already seeing movement in both directions—some accounts shifting back to admitted markets as appetite expands, and others moving into E&S as underwriting tightens.

When an agency uses systems that are not integrated, a carrier decline can mean significant manual work—re-entering data, rebuilding documents, and recreating submissions from scratch. In the E&S market, where supplemental applications often appear late in the process, this friction adds up fast.

To manage these changing processes, agencies are using submission and application management tools to collect and organize risk information once by integrating with their agency management system. They pair these tools with benchmarking solutions to understand which carriers or MGAs will be most likely to write the risk, then carry the information out to the market without someone having to re-enter it. Solutions that talk to each other keep the work moving and teams don't have to rebuild the same submission over and over.

Why consistent client communication separates high-performing agencies

After years of pricing pressure, underwriting changes, and continuing remarketing, clients rely on access to their agents more than ever. They want to talk about coverage shifts and rate increases before they happen. When they have to call and ask why something's changed, it's often too late—they're already shopping.

This emphasis on consistent communication showed up in the survey. More than half of respondents agreed that providing proactive, timely communication will set high-performing agencies apart in 2026.

But for busy agencies, it's difficult to provide a higher level of outreach. Juggling renewals, remarketing risk, and doing the daily work that keeps the doors open keeps teams from picking up the phone or sending an email. When delivery depends on individual effort, important messages become reactive and clients get frustrated.

That's why using a marketing automation tool for communications is so important. These solutions can either be integrated within your existing management system or built into your CRM or AMS tool. Agencies that use these tools turn routine messaging like renewal updates, document readiness, and general market info over to technology, and use the time savings to have deeper conversations about more important issues. Automating second-tier communication means clients stay informed and agency professionals aren't spending time manually writing and sending emails.

Using this technology also means clients receive intentional touchpoints that show a higher level of attention and care. Agencies can automatically send focused reminders to homeowners about fire season, educational materials about important topics, or alerts about upcoming trends. Templated messages go out on time and people step in where judgment and empathy are required.

It may feel counterintuitive, but automating routine communication often improves satisfaction. Clients feel informed and have clearer expectations. Teams aren't overwhelmed trying to reach everyone at once. With the right solutions carrying some of the load, agencies can elevate their clients' experience and reduce workloads.

Operational efficiency matters for independent agencies in 2026

Across the three themes—AI-driven workflows, E&S readiness, and automating client communication—the common thread is operational efficiency.

The agencies that will perform best in 2026 won't predict every market move. They will invest in how work gets done—making sure their systems talk to each other, their processes are disciplined, and their teams aren't buried in manual tasks.

It's not really a technology story. It's about how agencies manage their business day to day. And the agencies that focus on that now are going to be in a much better position in 2026 and beyond.

Improving Understanding of Risk Appetite

AI-driven appetite scoring can filter submissions, delivering efficiency gains in underwriting that exceed 30% across P&C lines.

Abstract Geometric Architectural Design in Black and White

Many insurance companies struggle to articulate and operationalize a precise appetite for risk. When underwriting guidelines lack clarity, producers and agents lack context for submission decisions. As a result, misaligned risks crowd pipelines and slow quoting timelines, reducing overall productivity. With competitive pressures rising across property and casualty (P&C) lines, improving the precision of risk intake is becoming essential.

Insurers embedding artificial intelligence (AI) into core functions, like underwriting and intake, can realize efficiency gains of more than 30%, primarily through reduced manual workload and better decision flows, according to Boston Consulting Group (BCG) research.

Why Traditional Appetite Communication Falls Short

Like it or not, communicating appetite through static documents, such as PDFs, spreadsheets, or email blasts, is still the norm. These formats are often misplaced, degrade quickly, and offer little real-time clarity. Agents often submit risks with incomplete information or insight into what aligns with underwriting goals, and underwriting teams then spend valuable time reviewing misaligned leads.

Further, while 88% of insurers use AI in at least one business function, few have scaled predictive decision-making tools enterprise-wide, according to a 2025 McKinsey & Company survey. This gap between experimentation and enterprise adoption presents an opportunity for first movers to gain a strategic edge.

The Rise of Predictive Appetite Scoring

Predictive appetite models bring intelligence to the earliest point of engagement by scoring submissions at pipeline entry. These models evaluate internal guidelines, performance metrics, and third-party data to calculate appetite fit and route each lead accordingly.

Instead of relying on static underwriting rules, predictive appetite scores interpret real-time risk context to determine which submissions align with portfolio goals. High-fit leads move straight to quoting or prioritized review while others are flagged for additional enrichment, redirected, or held back from underwriter queues altogether. Industry best practices followed by many insurers dictate the use of clean, normalized third-party data on-demand for pre-fill. Typically, third-party data is sourced for risk classifications (like NAICS), revenue, headcount, years in business, and more.

This shift enhances decision accuracy while minimizing effort on low-potential submissions.

Embedding Appetite into Distribution Workflows

When appetite scores appear directly in agency portals or APIs, agents are guided toward an insurer's desired result during the quoting process. In that scenario, submissions become more aligned, quote quality improves, and back-and-forth clarification drops. For the existing or potential policyholder, the process becomes faster and more transparent.

This embedded logic supports producers while reinforcing underwriting discipline without changing the core infrastructure.

The Underwriting Efficiency Advantage

Predictive appetite enables strategic triage. Underwriters spend more time on viable opportunities and less on sorting through noise. When intake is clean and appetite is operationalized, quote-to-bind ratios increase and loss ratios stabilize.

It's not theoretical. BCG reports that predictive models reduce acquisition waste and improve conversion rates across insurers actively using them in distribution and underwriting.

The Future of Appetite as Intelligence

Predictive appetite is the foundation for broader transformation. Insurers that build scoring into submission pipelines today will soon use those models to steer portfolios, enhance pricing, and select distribution partners more effectively.

Over the next 12 to 24 months, expect a widening gap between insurers using appetite scoring to optimize workflows and those still relying on static documentation. Appetite scoring will become central to broader underwriting transformation, informing everything from risk selection to strategic planning.

A Strategic Imperative

Insurers will increasingly turn to predictive appetite modeling to make smarter decisions earlier in the pipeline, reducing manual work and improving distribution alignment without needing to replace core systems.

For insurance companies looking to operationalize predictive appetite, there are some critical elements to add to the roadmap, including:

  • Appointing a senior champion: A CIO, CTO, or chief underwriting officer (someone with actual decision-making power) should sponsor the initiative.
  • Assembling a cross-functional team: Give underwriting, data science, product, and distribution stakeholders equal input when it is time to define appetite criteria and success metrics.
  • Starting modular: Use existing submission workflows and enrich them with third-party data and scoring logic.
  • Measuring and iterating: Use real-time feedback loops to refine models and improve appetite alignment over time.

Done well, predictive appetite modeling strengthens broker relationships, protects margins, and unlocks smarter growth. At the end of the day, it's about quoting faster (and more accurately) by building intelligent underwriting engines that learn and adapt over time.


Jay Bourland

Profile picture for user JayBourland

Jay Bourland

Jay Bourland is the chief technology officer at Fenris Digital.

Previously, he was SVP of engineering at Alteryx. He has also held leadership roles at Pitney Bowes, including VP/GM and group technology officer.

He holds five patents in geocoding and a Ph.D. in applied mathematics from Southern Methodist University. He has held academic positions at Stanford, the University of Washington, and Colorado State University.

The Municipal Catastrophe Insurance Crisis

Carriers must innovate in products for natural catastrophes, in distribution and in capital—or municipalities will build alternatives themselves.

Beige Low-angle Photo High-rise Building

The catastrophe insurance crisis facing American municipalities isn't a crisis of risk — it's a crisis of imagination.

When the insurance industry tells municipal leaders to "invest in resilience" without developing insurance products that reward those investments with affordable coverage, they're asking communities to subsidize the industry's own innovation deficit. When carriers collect premiums from communities while simultaneously withdrawing coverage and providing no recourse to close protection gaps, they leave municipalities to bear the burden of adaptation alone. And when carriers pull away from vulnerable communities while posting record profits, their risk decisions become a way of abdicating their core function of pooling and transferring risk.

This mismatch is unsustainable and dangerous, and the consequences cascade far beyond equity concerns. Low-income residents can't rebuild after disaster. Without business interruption coverage, small businesses close, devastating local employment and sales tax revenue. Affordable housing providers, facing uninsurable properties, stop developing in climate-vulnerable areas, exacerbating housing shortages and displacement. These "equity issues" then become systemic economic failures that threaten the tax base carriers rely upon for commercial lines business.

If carriers don't evolve, they won't just be disrupted by climate change; they'll be made irrelevant by it. To meet the needs of climate-challenged jurisdictions and address these cascading failures, carriers must dramatically increase their R&D investment across product, distribution, and capital structure.

Product Innovation

The catastrophe insurance products available to municipalities today are fundamentally inadequate for current climate realities, let alone future projections. Carriers must invest in genuine product R&D across several critical areas.

Microinsurance for Vulnerable Populations: Carriers should develop lower-limit, lower-premium products specifically designed for low-income households and small businesses. Communities with functioning insurance markets maintain economic activity after disasters while communities without insurance face cascading failures that destroy the commercial premium base.

Advanced Modeling That Provides Incentives for Novel Adaptation: Current catastrophe models treat risk as static or worsening, with no meaningful premium reduction for innovative resilience projects. Carriers must bolster their modeling capabilities, so they can actually quantify the risk reduction from novel adaptation investments. This requires fundamental R&D investment in coupled physical-financial modeling, not incremental improvements to existing cat models.

Parametric Insurance for Rapid Response: Traditional indemnity insurance is ill-suited for disaster response, but parametric products that trigger immediate payouts based on objective measurements (flood depth, wind speed, earthquake magnitude) enable rapid recovery while reducing time and costs. The question is whether carriers will invest in developing parametric products tailored to municipal needs or whether municipalities will turn to specialized parametric providers who will.

Distribution Innovation

Even when adequate insurance products exist, traditional distribution channels systemically fail to reach those who need coverage most. Insurance carriers must invest in new distribution models, particularly embedded insurance mechanisms that integrate coverage directly into existing community touchpoints.

Affordable Housing: Catastrophe coverage can be embedded directly into affordable housing and stay with the property, not the individual tenant or owner. This closes protection gaps for vulnerable residents while creating stable, pooled risk for carriers.

Utility Bill: Municipalities can embed catastrophe coverage directly into customers' utility bills, creating universal protection while drastically reducing acquisition costs and adverse selection.

Employee Benefit: Municipalities and anchor institutions (hospitals, universities) employ thousands of workers, many of whom lack adequate catastrophe coverage. Embedding basic catastrophe protection in employee benefit packages closes protection gaps while creating group purchasing efficiency.

Small Business District: Local business districts and chambers of commerce can serve as aggregators for embedded small business catastrophe coverage. Small businesses often lack business interruption and property coverage. District-level embedded insurance solves the distribution problem while enabling risk pooling across merchants.

Capital Structure

Perhaps the most fundamental innovation required is at the capital structure level. Traditional reinsurance capital is designed to maximize returns for institutional investors and reduce exposure to complex, small-scale risks that don't fit standardized cat bond structures.

Community Development Reinsurance Institutions (CDRIs) — mission-driven nonprofit reinsurers structured similarly to community development banks — offer an alternative capital model specifically designed to support municipal resilience and insurance market function.

Traditional carriers should see CDRIs not as competitors but as catalysts for market development. By providing reinsurance for products serving underserved markets, CDRIs enable primary carriers to write business they otherwise couldn't while maintaining risk tolerance within board-approved limits. This expands the overall insurance market rather than displacing existing business.

Yet most carriers remain unaware of or unengaged with the emerging CDRI sector. If carriers don't invest in understanding, partnering with, and leveraging mission-driven reinsurance capacity, they'll soon discover that CDRIs have enabled an entirely new ecosystem of insurance providers serving municipalities—providers who didn't need traditional carrier participation to succeed.

The Choice Before Carriers: Lead, Follow, or Become Obsolete

The insurance industry faces a stark choice: invest in the R&D necessary to develop products, distribution channels, and capital structures adequate to municipal climate realities or continue business as usual and risk becoming obsolete.

The communities that carriers are abandoning won't simply accept uninsurability. They'll build alternatives — self-insurance pools, parametric coverage through specialized providers, embedded insurance through MGAs, risk financing through CDRIs and innovative bonds that will chip away at carriers' market relevance. Eventually, the "alternative" insurance ecosystem serving municipalities will become the mainstream, and traditional carriers will lose out on large parts of the market.

This isn't speculation; it's already happening. The only question is whether the insurance industry will rise to meet the demands and challenges in the municipal markets.


Charlie Sidoti

Profile picture for user CharlieSidoti

Charlie Sidoti

Charlie Sidoti is the founder and executive director of Innsure, a nonprofit with a mission to foster innovation in insurance and a focus on catalyzing insurance industry response to climate change.

He has 25 years in the insurance industry, all with commercial P&C carriers in a variety of risk management leadership roles. He served on the board of the Insurance Institute for Business and Home Safety. Sidoti has also spent 10 years working on insurance-adjacent startups.  

Sidoti is a visiting lecturer and adviser to Northeastern University on the new Insurance Analytics and Management master's program.