How InsureAI Works
A production-grade RAG + LLM architecture that gives you genuinely personalised insurance recommendations — not just keyword matches.
System Architecture
1. Data Ingestion
CSV datasets (AutoInsurance, claims data, premium tables) are parsed and converted into rich policy documents with pricing factors.
Python · Pandas · NumPy2. Embedding & Indexing
Policy documents are embedded using sentence-transformers (all-MiniLM-L6-v2) and stored in a Qdrant vector database for semantic search.
Sentence-Transformers · Qdrant3. RAG Retrieval
User input is converted to a query vector. Top-20 semantically similar policies are retrieved, filtered for eligibility and budget.
Qdrant cosine search · Eligibility filters4. LLM Analysis
Retrieved policies + user profile are sent to Claude (Sonnet) which ranks policies, calculates personalised premiums, writes pros/cons and an AI summary.
Claude Sonnet 4.6 · Anthropic APITech Stack
Frontend
Next.js 15 · React 19 · Tailwind CSS
Backend
FastAPI · Python 3.11+
Vector DB
Qdrant (local file-based or cloud)
Embeddings
all-MiniLM-L6-v2 (sentence-transformers)
LLM
Claude Sonnet 4.6 (Anthropic API)
Data
Kaggle car insurance datasets (3 sources)