AMFS vs Vector Databases

Vector databases and AMFS solve different problems. Understanding the distinction helps you pick the right tool — or use both.

Table of Contents

  1. The Short Version
  2. Side-by-Side Comparison
  3. What Vector Databases Do Well
  4. What Vector Databases Don’t Do
  5. What AMFS Does Differently
    1. Knowledge has identity
    2. Knowledge has provenance
    3. Knowledge evolves with outcomes
    4. Knowledge has full history
    5. Decisions are explainable
  6. Using Both Together
  7. The Compounding Loop
  8. Summary

The Short Version

Vector databases store embeddings and retrieve them by similarity. They answer: “what is most relevant to this query?”

AMFS is version control for what agents know. It answers: “what does this agent know, who wrote it, how confident are we, what happened when we acted on it, and how do we collaborate on changing it?”

A vector database is a search index. AMFS is GitHub for agent memory — versioning, branching, pull requests, rollback, and a collaboration model developers already understand.


Side-by-Side Comparison

Dimension Vector Database AMFS
Primary operation Similarity search over embeddings Read/write versioned knowledge with provenance
Collaboration model Shared index — last write wins Git model — branch, diff, PR, merge, rollback, access control
Data model Vectors + metadata Structured entries with entity/key scoping, confidence, memory type, provenance
Versioning Overwrite or append Copy-on-Write — every write creates a new version, full history preserved
Who wrote it? Not tracked Provenance: agent ID, session ID, timestamp, pattern refs
Trust signal None Confidence score that evolves based on real-world outcomes
Feedback loop None Outcome back-propagation — incidents boost confidence, clean deploys decay it
Query style “Find similar to X” “Read key Y”, “Search by entity/agent/confidence”, “What happened over time?”
Temporal queries Snapshot at query time Full version history with time-range filtering
Explainability None Causal chain: which entries + external contexts informed a decision
Multi-agent Shared index Shared memory with per-agent provenance, conflict detection, auto-causal linking
Typical size Millions–billions of vectors Thousands–millions of knowledge entries
Update pattern Re-embed and upsert CoW write with automatic version increment

What Vector Databases Do Well

Vector databases excel at large-scale semantic retrieval:

  • RAG (Retrieval-Augmented Generation) — Finding relevant document chunks to inject into an LLM prompt. When you have 10M documents and need the top-10 most relevant passages, a vector database is the right tool.
  • Similarity search — “Find products similar to this one”, “Find code snippets that match this pattern.”
  • Multimodal retrieval — Searching across text, images, and audio using shared embedding spaces.
  • Real-time recommendation — High-throughput, low-latency nearest-neighbor queries.

Popular options include Pinecone, Weaviate, Qdrant, Milvus, and Chroma.


What Vector Databases Don’t Do

Vector databases are stateless retrieval indexes. They don’t track:

  • Who wrote the data — No provenance. You don’t know which agent or process created an entry.
  • How trust evolves — No confidence scoring. A vector’s relevance score is similarity to a query, not a measure of how trustworthy the information is.
  • What happened when you used it — No outcome tracking. If an agent retrieves a vector and acts on it, and that action causes an incident, the vector database has no way to learn from that.
  • How data changed over time — Vectors are overwritten or appended. You can’t ask “what did this entry say last week?”
  • Why a decision was made — No causal chain linking retrieved data to actions and outcomes.
  • How to collaborate on changes — No branching, no review process, no way for one agent to propose a change and another to approve it. It’s like coding without Git.

What AMFS Does Differently

AMFS is designed for agent memory — the layer between retrieval and action:

Knowledge has identity

Every entry has an entity_path and key that give it a stable address. Agents read and write to specific keys, not anonymous vectors.

mem.write("checkout-service", "retry-pattern", {"max_retries": 3})
entry = mem.read("checkout-service", "retry-pattern")

Knowledge has provenance

Every entry records who wrote it, when, and in which session:

print(entry.provenance.agent_id)    # "review-agent"
print(entry.provenance.written_at)  # 2026-03-15T14:30:00Z

Knowledge evolves with outcomes

When a deploy succeeds or an incident occurs, confidence scores on related entries adjust automatically:

mem.commit_outcome("INC-1042", OutcomeType.CRITICAL_FAILURE)
# All entries this agent read get their confidence boosted

Knowledge has full history

Every write creates a new version. You can replay the state at any point in time:

versions = mem.history("checkout-service", "retry-pattern", since=last_week)

Decisions are explainable

The causal chain shows exactly what informed a decision — both AMFS entries and external tool context:

chain = mem.explain("DEP-500")
# → causal_entries: entries the agent read
# → external_contexts: tool/API inputs the agent consulted

Using Both Together

AMFS and vector databases are complementary. A common architecture:

                    Agent
                   /     \
                  /       \
         ┌──────▼──┐  ┌──▼──────────┐
         │  AMFS   │  │ Vector DB   │
         │ Memory  │  │ (RAG index) │
         └─────────┘  └─────────────┘
         What we know  What's relevant
         + trust       to this query
         + history
         + outcomes
  1. Vector DB for retrieval — Agent queries the vector database to find relevant documents or code snippets for the current task.
  2. AMFS for memory — Agent reads AMFS for known patterns, risks, and past decisions about the entity it’s working on.
  3. AMFS for recording — After completing its task, the agent writes findings, decisions, and risks to AMFS with provenance and confidence.
  4. AMFS for learning — Outcomes (deploys, incidents) back-propagate through AMFS, adjusting confidence scores so future agents see which patterns are trustworthy.

AMFS supports hybrid search — combining full-text (Postgres tsvector), semantic (cosine similarity via pluggable embedders), recency, and confidence into a single ranked result set. For small-to-medium memory stores, this eliminates the need for a separate vector database. AMFS also auto-materializes a knowledge graph from normal operations, capturing entity relationships and causal links that pure similarity search cannot surface. For large-scale RAG over millions of documents, a dedicated vector database is the right choice.


The Compounding Loop

The fundamental difference becomes clear over time. A vector database stores static embeddings that you retrieve. AMFS builds a compounding knowledge asset that gets more valuable the longer you use it:

Week 1:   Agent writes patterns → basic memory
Week 4:   Outcomes validate patterns → confidence evolves
Week 12:  Decision traces show what worked → precedent search
Week 24:  Pattern detection surfaces recurring failures → proactive alerting
Week 52:  Learned ranking personalizes retrieval → domain-calibrated intelligence

With Pro, this loop extends to cross-system context (PagerDuty incidents, GitHub PRs, Slack threads automatically ingested), persistent decision traces queryable months later, and automated pattern detection that surfaces recurring failures before they become incidents.

No vector database provides this lifecycle. They are a retrieval layer. AMFS is a memory system that compounds.


Summary

  Vector Database AMFS
Think of it as A search index for embeddings GitHub for agent memory
Best for Finding relevant data Collaborating on knowledge the way developers collaborate on code
Collaboration None — shared index, last write wins Branch, diff, PR, review, merge, rollback, access control
Data lifecycle Write once, query many Write, version, track outcomes, decay, explain
Multi-agent Shared index Shared memory with provenance, conflicts, and causal chains
Cross-system context Manual ingestion Pro: auto-ingest from PagerDuty, Slack, GitHub, Jira
Pattern intelligence None Pro: recurring failures, stale clusters, confidence drift
Enterprise readiness Auth varies by vendor Pro: RLS isolation, RBAC, scoped API keys, audit logging