AI Agents for insurance: How to Automate document extraction (multi-agent with LlamaIndex)
Insurance document intake is still too manual in most carriers and brokers. Claims packets, ACORD forms, loss runs, policy schedules, medical bills, and underwriting submissions arrive in mixed formats, and teams burn hours rekeying data into core systems.
Multi-agent extraction with LlamaIndex gives you a way to split that work into specialized steps: classify the document, extract the right fields, validate against policy rules, and route exceptions to humans. For a CTO or VP of Engineering, the value is simple: lower handling cost, faster cycle times, and fewer downstream errors in claims and underwriting.
The Business Case
- •
Claims intake time drops from 20-40 minutes per file to 2-5 minutes
- •For FNOL packets and supporting documents, an extraction agent can pre-fill claim systems and cut manual review by 70-85%.
- •In a mid-sized P&C carrier processing 10,000 claims documents/month, that’s roughly 3,000-6,000 labor hours saved annually.
- •
Operational cost falls by 25-40% on document-heavy workflows
- •Manual indexing, transcription, and validation are expensive because they require trained ops staff.
- •If your document operations team costs $60k-$90k per FTE loaded, automating even 5-8 FTEs in intake and triage can save $300k-$700k/year.
- •
Field-level error rates drop from 3-8% to under 1%
- •Common failures include wrong policy number, missed deductible values, bad date parsing, and misread ICD/CPT codes on medical attachments.
- •Lower error rates reduce rework in claims adjudication and underwriting referrals.
- •
Cycle time improves enough to move business metrics
- •Faster first notice of loss processing improves customer satisfaction and reduces leakage from delayed routing.
- •In commercial lines underwriting, cutting submission prep from 2 days to same-day can materially improve broker response times.
Architecture
A production setup should not be a single prompt calling OCR. It should be a controlled multi-agent pipeline with clear responsibilities.
- •
Ingestion + OCR layer
- •Use Azure Form Recognizer, AWS Textract, or Google Document AI for scanned PDFs and images.
- •Normalize outputs into a canonical JSON schema before any LLM touches the data.
- •Store raw documents in encrypted object storage with retention controls aligned to your data policy.
- •
LlamaIndex orchestration layer
- •Use LlamaIndex for document parsing, chunking, retrieval over policy manuals, coverage guidelines, claims playbooks, and underwriting rules.
- •Add a classifier agent to detect document type: ACORD application, declaration page, invoice, medical bill, proof of loss, police report.
- •Add an extractor agent per document family so prompts stay narrow and testable.
- •
Validation and routing layer
- •Use LangGraph to coordinate multi-step workflows:
- •classify
- •extract
- •validate
- •resolve conflicts
- •escalate exceptions
- •Store embeddings in pgvector if you need semantic retrieval over prior submissions or internal procedures.
- •Add deterministic checks for policy number format, date ranges, coverage limits, deductible thresholds, and named insured matching.
- •Use LangGraph to coordinate multi-step workflows:
- •
Audit and integration layer
- •Push validated output into claims platforms like Guidewire or Duck Creek through APIs or message queues.
- •Log every extraction decision with source span references for auditability.
- •Keep an immutable trail for compliance reviews under SOC 2, privacy obligations under GDPR, and healthcare-related workflows under HIPAA when medical documents are involved.
A practical workflow
flowchart LR
A[Incoming PDF/Image] --> B[OCR / Document AI]
B --> C[Doc Classifier Agent]
C --> D[Extractor Agent]
D --> E[Rules Validator]
E --> F{Confidence >= threshold?}
F -- Yes --> G[Claims / UW System]
F -- No --> H[Human Review Queue]
H --> G
This is the pattern that works:
- •keep extraction narrow,
- •make validation deterministic,
- •route uncertainty to people,
- •store evidence for every field.
What Can Go Wrong
| Risk | Where it shows up | Mitigation |
|---|---|---|
| Regulatory exposure | Medical claim docs may contain PHI; EU policyholder data may fall under GDPR; model logs can leak sensitive data | Encrypt at rest/in transit, redact before logging, enforce role-based access control, define retention policies, run DPIAs for GDPR-covered flows |
| Reputation damage | Wrong reserve inputs or incorrect coverage interpretation can trigger customer complaints or bad-faith allegations | Require human approval above confidence thresholds; show source text for every extracted field; start with low-risk documents like certificates of insurance or invoices |
| Operational failure | OCR noise, poor scans, handwritten notes, or edge-case endorsements break extraction quality | Build exception queues; use fallback OCR providers; maintain a golden test set of real documents; monitor precision/recall by doc type weekly |
A common mistake is treating the LLM as the system of record. It is not. The system of record is your validated JSON plus traceable evidence from the source document.
For regulated environments like insurance groups touching banking-adjacent risk data or reinsurance reporting tied to capital processes under frameworks influenced by Basel III, you want strict controls around lineage and human sign-off. If you cannot explain where a field came from in an audit trail, it does not belong in production automation.
Getting Started
- •
Pick one narrow workflow
- •Start with a single high-volume use case: ACORD submission intake for commercial property/casualty, loss run extraction for underwriting renewal prep, or invoice extraction for claims payment support.
- •Avoid “all documents” pilots. That turns into a platform project before you have evidence.
- •
Build a pilot team of 4-6 people
- •One product owner from operations or claims
- •One backend engineer
- •One ML/AI engineer
- •One data engineer
- •One QA/UAT analyst
- •Optional compliance partner part-time
- •
Run a 6-8 week pilot
- •Week 1-2: collect sample docs and define schema
- •Week 3-4: build OCR + LlamaIndex extraction + validation rules
- •Week 5: integrate human review queue
- •Week 6-8: measure accuracy on at least 500-1,000 real documents
- •
Set hard success metrics before rollout
- •Target at least:
- •95%+ field accuracy on critical fields like policy number and insured name
- •80%+ straight-through processing on clean documents
- •50% reduction in manual handling time
- •zero unresolved compliance issues from security review
- •Target at least:
If those numbers hold on real traffic—not lab samples—you have something worth scaling. If they do not hold, tighten scope until the workflow is deterministic enough for production insurance operations.
The right way to deploy this is not as a chatbot. It is as an audited document pipeline with agents doing bounded work inside a controlled architecture. That is how you get automation past risk committees without creating another shadow IT experiment.
Keep learning
- •The complete AI Agents Roadmap — my full 8-step breakdown
- •Free: The AI Agent Starter Kit — PDF checklist + starter code
- •Work with me — I build AI for banks and insurance companies
By Cyprian Aarons, AI Consultant at Topiax.
Want the complete 8-step roadmap?
Grab the free AI Agent Starter Kit — architecture templates, compliance checklists, and a 7-email deep-dive course.
Get the Starter Kit