Wiki vs Vector Database
Two ways to store and retrieve knowledge for AI systems. Different tools, different layers of the stack.
The core question
How does an LLM find the right information when it can't read everything at once?
LLMs have a context window — a limit on how much text they can process at once. If your knowledge base is bigger than that window, you need a way to select what goes in. Wikis and vector databases are two answers to this problem.
Wiki
A human pre-digests information into structured pages with explicit links between them. The human decided what's important and how things relate.
Vector Database
Math pre-filters information by semantic similarity. Text is converted to numbers (embeddings), and "nearby" numbers mean "similar meaning."
What they actually store
The difference is concrete. One stores readable pages, the other stores arrays of numbers.
Wiki stores this
type: entity
tags: [pioneer, memex]
body: "Proposed the Memex in 1945..."
Human-readable text. Meaningful structure. Anyone can open this and understand it.
Vector DB stores this
1,536 numbers. This encodes meaning as math. No human can read this — but two vectors close together mean the text was semantically similar.
How they work — interactive
Try both retrieval mechanisms side by side.
Wiki: Click a node to see its connections
Explicit links. You see exactly why things are connected.
Vector: Type to search by meaning
Results ranked by similarity score. No structure — just "how close is this to your query?"
Head-to-head comparison
Same dimensions, different strengths.
| Dimension | Wiki | Vector DB |
|---|---|---|
| Structure | Explicit pages + links. Human-curated graph. | Flat chunks with embeddings. No explicit relationships. |
| Retrieval | Follow links, read index, browse by type. | Semantic search — "find things that mean something like X." |
| Auditability | Total. Every link is intentional and visible. | Low. Similarity scores are opaque. No "why." |
| Scale | Hundreds of pages. Needs curation effort. | Millions of chunks. No curation needed. |
| Stays current | Manual updates (or someone forgets). | Re-embed on every change — automatic. |
| Who interacts | Humans browse, read, edit, follow links. | Code queries it via API. Users never see it. |
When to use which
The answer depends on your problem, not the technology.
| Scenario | Better fit | Why |
|---|---|---|
| Research & strategy | Wiki | Curated links = a thinking tool you can browse |
| 10K+ support docs | Vector | Too many docs to curate. Embed and search. |
| Competitive intel | Wiki | Entities + synthesis = compounding insight |
| Chat with your docs (RAG) | Vector | Built for this. Retrieve relevant chunks, feed to LLM. |
| Onboarding a new teammate | Wiki | Browsable, readable, linkable. Humans can explore. |
| Finding a needle in a haystack | Vector | Semantic search finds things exact-match misses. |
| Person/entity matching | Vector | Fuzzy matching with confidence scores. |
Real product example: both in one stack
A sales outreach tool that uses wiki for context and vectors for search — different layers, different jobs.
Knowledge Base (wiki)
Account pages, contact profiles, sales playbooks. Reps browse this, read it, add notes after calls. Human-facing. Structured. Links between entities.
Person Matching Engine (vector)
Input: "VP Engineering, fintech, Bay Area." Searches 200M profiles. Returns scored matches. Users never see the vectors — they see a ranked list of people.
Confidence threshold — drag to filter
The score isn't about cost — it's about risk. High threshold = only near-certain matches. Low = cast a wider net.
Common misconceptions
Things that sound right but aren't.
"Vector databases learn and grow connections over time"
They don't. You add data, it gets embedded. No "neural links" form between existing data. The embedding model ran once. After that, it's static math in storage.
"Vector databases are like a brain"
They're indexed arrays of numbers. Not neural networks. The "neural" part happened once, when the embedding model converted text to vectors. The database itself is just a search index.
"Wikis and vector DBs do the same thing"
They both store knowledge. But a wiki stores relationships you chose. A vector DB stores distances math calculated. One is a map you drew. The other is GPS coordinates.
"You need a vector DB for an LLM to work"
LLMs use vectors internally (that's how neural nets work), but that's different from a vector database you build and host. If your data fits in the context window, you don't need one.
Same problems, different solutions
Both approaches solve for memory, retrieval, and control. They just use different mechanisms at different cost points.
| Problem | Wiki approach | Vector approach |
|---|---|---|
| Memory persistence | Wiki files survive between sessions | Vector DB survives between sessions |
| Retrieval | LLM reads index + pages inference | Math finds similar chunks no inference |
| Control / harness | Rules file in natural language tokens | Code + API parameters near-zero tokens |
Where the inference cost lives
The wiki approach uses more inference at every step. The vector approach offloads search and control to cheaper layers.
Wiki pipeline
Every step uses inference (tokens):
2. LLM reads index inference
3. LLM reads relevant pages inference
4. LLM generates answer inference
Vector pipeline
Search is cheap math. Only the answer uses inference:
2. Similarity search free (math)
3. LLM reads 5 chunks inference
4. LLM generates answer inference
Goals and control
What each approach optimizes for, what's deterministic, and what's out of your hands.
| Wiki | Vector DB | |
|---|---|---|
| Primary goal | Control what the LLM knows and how it connects things | Find relevant stuff fast in data too big to read |
| Deterministic | What links to what, what pages exist, schema, what's immutable | The embedding math (same text = same vector, always) |
| Non-deterministic | LLM's synthesis — how it writes summaries, what it emphasizes | Which chunks score highest (can vary with embedding model), what LLM does with them |
| Out of control | LLM may interpret your rules differently session to session | Retrieval quality — might surface junk that scores high |
What's a harness?
A harness is anything that wraps the LLM to constrain its behavior. It's not a specific technology. Both approaches need one.
| Harness type | Medium | Token cost | What it can enforce |
|---|---|---|---|
| Rules file (e.g. CLAUDE.md) | Natural language | Hundreds of tokens/session | Judgment calls, style, workflow, priorities |
| System prompt (API) | Natural language | Some tokens | Role, tone, constraints |
| Temperature setting | API parameter | Zero | Output randomness |
| Schema validation (code) | Code | Zero | Output format, required fields |
| Retry loop (code) | Code | Costs retries | Output correctness checks |