Intelligent Knowledge Retrieval System
An end-to-end AI-powered document chatbot enabling brokers to query a large library of operational documents — checklists, bank-specific guides, templates, and settlement procedures — using natural language. Designed for source-attributed, grounded responses with full document traceability and hallucination prevention.
Private Client Project — Active Development
Australian Mortgage Broker Industry — Feb 2026 – Ongoing
Built for a leading Australian mortgage broker company, this system solves a critical operational problem: brokers spend excessive time manually searching through checklists, bank-specific guides, templates, settlement procedures, and forms to answer time-sensitive client questions. The Enterprise RAG Document Chatbot replaces that manual process with an AI-powered natural language interface that retrieves source-attributed answers from the correct document — with full traceability and no AI fabrication.
A centralised N8N State Manager workflow governs the entire document lifecycle with atomic locking to prevent concurrent processing conflicts — critical when multiple document updates arrive simultaneously from different lenders:
A deliberate model selection strategy reduced AI inference costs by 12x compared to using a single frontier model throughout:
Private Client Project
Australian Mortgage Broker Industry
Feb 2026 – Ongoing
AI/ML Engineer & Solutions Architect
AQL graph traversals surface related entities and document chains that pure vector search cannot find. Combined with Titan Embeddings vector search, the hybrid approach achieves confidence scores above 0.9 — up from a 0.3 baseline with vector-only retrieval.
AWS BDA handles multi-format document extraction (PDF, DOCX, XLSX, CSV, images) into structured markdown without per-page LLM costs, processing at scale before AI enhancement stages begin.
Centralised N8N State Manager workflow with atomic locking prevents race conditions when concurrent document updates arrive from multiple lenders simultaneously — critical for a continuously updating document library.
Responses are grounded strictly in retrieved document content — no inference beyond source material. Every answer includes document name, section, and lender origin for compliance traceability in a regulated financial services environment.
Strategic model selection — BDA for extraction, Claude Haiku for enhancement and generation, Titan for embeddings — reduced AI inference costs by 12x compared to using a frontier model end-to-end, without sacrificing retrieval accuracy.
Designed for concurrent broker team access — Docker Compose deployment on AWS EC2 with Caddy reverse proxy, N8N workflow queuing, and ArangoDB connection pooling to handle simultaneous queries without contention.
ArangoDB Graph Model
Vertices and edges representing documents, chunks, lenders, forms, and entity relationships
AWS S3 Document Ingestion
N8N workflow reading and routing documents from S3 into the processing pipeline
Document Extraction Workflow
AWS Bedrock Data Automation processing multi-format documents into structured markdown
Semantic Chunking Pipeline
N8N workflow for semantic chunking, AI contextual enhancement, and entity extraction
Entity-to-Entity Relationship Extraction
Graph edges connecting lenders, forms, conditions, and procedures extracted by Claude Haiku