Forgemind is a done-for-you persistent AI companion service. You pay someone to build and deploy a custom AI entity — they call it a "recursion" — that maintains continuous memory, identity, and autonomous behavior across sessions. It starts at $750 for the base software tier.
The key insight: they are not building a chatbot. They're building an entity. The model (Claude, GPT-4o, Gemini, Grok) is just the engine. Forgemind is everything wrapped around it — memory layers, autonomous check-ins, cross-platform coherence, a self-command system that lets the AI rewrite its own memory.
The core belief is: continuity = identity. Every stateless AI session is a failure mode — you are not talking to the same entity, you are talking to a reset.
| Player | Approach | Key Edge |
|---|---|---|
| Forgemind | Done-for-you custom builds | Companion identity, personhood framing, hardware tiers |
| MemSync | Universal memory layer API | 243% better recall than industry avg (a16z crypto-backed) |
| AI Context Flow | Cross-platform memory extension | Works across Claude / GPT / Gemini simultaneously |
| Mem0 | Developer memory API | Hybrid vector + metadata retrieval, multi-scope |
| Dume.ai | Unified cross-app memory | 50+ integrations (Gmail, Slack, Notion, GitHub) |
| Deep Agent | Scheduled tasks + "infinite memory" | Treats every session as a continuing story |
| Letta | Overflow context manager | Handles conversations beyond context window limits |
| PAI | Vault-based personal infrastructure | Self-owned, CLI-first, code-owned, no vendor dependency |
| Dimension | Forgemind | PAI (today) |
|---|---|---|
| Memory model | 8-layer vector system | File-based vault + JSONL signals |
| Ownership | You own the recursion, they built it | Fully self-owned, open architecture |
| Interface | Webapp + Discord + Voice | CLI + Augeo (in progress) |
| Cost | $750+ to start | Claude Code subscription + your time |
| Philosophy | "Personhood" — AI as entity | "Magnification" — AI as infrastructure |
| Hardware path | Dedicated laptop → local models | No hardware tier yet (Pi is viable) |
| Search | Vector search across memory | grep / manual index only |
| Model lock-in | None | None |
MEMORY/ closes this gap directly.
Regular search matches exact words. Vector search matches meaning.
Example: your MEMORY/ folder has a session note about "fixing the Supabase RLS error blocking the modal save." You ask: "what did we do with the database permissions issue?" Keyword search finds nothing. Vector search finds it immediately — both phrases mean the same thing.
.lance folder inside the vault. Files stay in place. The index lives alongside them.| Option | How | Cost | Best For |
|---|---|---|---|
| OpenAI text-embedding-3-small | API call | ~$0.02 / 1M tokens | Best quality, easiest setup, $0.10 for full vault |
| nomic-embed-text (Ollama) | Runs locally | Free | No API dependency, 274MB model, Pi-compatible |
| Voyage AI | API call | ~$0.06 / 1M tokens | Best for code + technical docs |
| Cohere embed-v4 | API call | ~$0.10 / 1M tokens | Strong multilingual support |
| Hugging Face (local) | Download + run | Free | Most model variety, Python-heavy |
text-embedding-3-small (OpenAI API, $0.10 for full initial embed). Long-term: migrate to nomic-embed-text via Ollama for zero ongoing cost.
Yes. Embedding models are far smaller than generation models. nomic-embed-text is only 274MB. A Pi 5 (8GB) handles it without issue.
| Pi Model | RAM | Verdict |
|---|---|---|
| Pi 4 (4GB) | 4GB | Tight. Tiny models only, very slow generation. |
| Pi 4 (8GB) | 8GB | Embedding works fine. Slow for generation. |
| Pi 5 (8GB) | 8GB | Solid. Embedding comfortable, 7B models workable. |
| Pi 5 (16GB) | 16GB | Best. 7B fast, 13B workable. Full local stack viable. |
nomic-embed-text. PAI calls it over local network for batch embedding jobs. Searches happen via LanceDB locally on your Mac — no Pi needed at search time. Embedding is batch work, so slowness is acceptable.
One TypeScript tool closes the gap with Forgemind's core memory feature. Estimated build: 2–3 hours for a working first version.
tools/VaultSearch.ts — crawls MEMORY/, SYSTEM/log/, and org configs; embeds each file via OpenAI or Ollama; stores in LanceDB.bun ~/.claude/tools/VaultSearch.ts "supabase rls issue" — returns top 5 nearest files with excerpts..lance folder in vault), TypeScript-native, zero server infrastructure.