Mintlify Ditched RAG for a Virtual Filesystem. Their AI Documentation Assistant Got Much Better.
RAG is the default architecture for giving AI assistants access to a knowledge base. Mintlify tried it, found it brittle, and built something different. Their virtual filesystem approach is now driving measurably better answers — and challenging some foundational assumptions about AI retrieval.

D.O.T.S AI Newsroom
AI News Desk
Retrieval-augmented generation has become the default playbook for AI assistants that need to answer questions about a large corpus of documents. The pattern is familiar: chunk the documents, embed the chunks, retrieve the most semantically similar ones at query time, stuff them into context, and ask the model to synthesize an answer. RAG works well enough in demos. In production, it is messier.
Mintlify, the developer documentation platform, ran into those production problems firsthand. Their AI documentation assistant — designed to help users navigate and understand complex technical documentation — was producing answers that were technically sourced from the right documents but frequently missed the structural context that made those answers actually useful. A function reference answered in isolation from the guide that explained when to use it. A configuration option described without the surrounding workflow that gave it meaning.
The Problem With Chunking
The culprit, Mintlify's engineering team concluded, was chunking. RAG systems decompose documents into fixed-size or semantically-bounded fragments before embedding. The embedding captures semantic content, but severs structural relationships — the hierarchy of sections, the cross-references between pages, the sequential logic of a tutorial. When the retrieval step surfaces three chunks from three different parts of a documentation site, the model has to reconstruct meaning from fragments that were never designed to be read together.
"We were getting answers that were factually correct but contextually wrong," the Mintlify team writes in a blog post that surfaced on Hacker News this week. "The model knew what a function did. It didn't know when you'd use it, because that context lived in a different chunk that wasn't retrieved."
The Virtual Filesystem Architecture
Mintlify's alternative treats documentation as a filesystem rather than a bag of chunks. The assistant maintains a persistent representation of the documentation's structure — directories, files, headings, cross-links — and navigates it programmatically when answering a query, much like a developer navigating a codebase.
When a user asks a question, the assistant doesn't retrieve chunks by semantic similarity. Instead, it reasons about which parts of the documentation tree are likely relevant, traverses to those nodes, reads them in context, and follows cross-references where needed. The model acts less like a search engine and more like a developer who knows the codebase well enough to navigate it intentionally.
The results, Mintlify reports, are substantially better on questions that require integrating information across multiple sections — the questions that matter most in technical documentation contexts. The system is also more explainable: when the assistant cites a source, it can point to a specific location in the documentation hierarchy rather than a decontextualized chunk.
Broader Implications
Mintlify's approach is one of several emerging alternatives to flat RAG architectures. GraphRAG, pioneered by Microsoft Research, uses knowledge graphs to preserve relationship structure. Long-context models from Anthropic and Google reduce the need for retrieval by fitting more of the corpus directly into context. Each approach trades different things: GraphRAG adds construction complexity, long-context adds inference cost, virtual filesystems require a structured corpus.
For documentation specifically — a domain with inherent hierarchical structure — the filesystem metaphor is a natural fit. The more interesting question is whether the pattern generalizes to less-structured corpora. Mintlify's contribution is less a universal alternative to RAG and more a reminder that the right retrieval architecture depends heavily on the structure of the data being retrieved.