Install one command, answer a few questions, and you have a knowledge vault that gets smarter every day you use it. NeuroStack scaffolds your vault, indexes everything by meaning, surfaces what needs attention, and gives any AI tool long-term memory—all running locally.
One command installs everything—no Python, git, or curl needed. The interactive setup walks you through vault location, model selection, and optional profession packs.
npm install -g neurostack
neurostack initUse as an MCP server with Claude Code, Codex, Gemini CLI, Cursor, Windsurf—or pipe CLI output into any LLM workflow.
neurostack serveHot notes surface what matters. Drift detection flags what's stale. Community detection reveals hidden themes. Your vault becomes a living knowledge system that improves the more you use it.
neurostack search "query"Combines FTS5 full-text with semantic embeddings for meaning-based retrieval. Keyword, semantic, or hybrid modes.
McClelland et al. 1995 — Complementary learning systemsTriples (~15 tok) → Summaries (~75 tok) → Full content. Auto-escalates based on result quality. Token-efficient by design.
CLS theory — hippocampal rapid bindingActive notes attract preferential connections — inspired by CREB-mediated neuronal excitability windows in memory consolidation.
Han et al. 2007 — Science 316(5823)Flags notes retrieved with high cosine distance from query intent. Surfaces outdated content and miscategorisation.
Sinclair & Bhatt 2022 — PNAS 119(31)Attractor basin community detection with hierarchical levels. Inspired by Hopfield network energy landscapes. Coarse thematic clusters and fine sub-themes with LLM-generated summaries.
Ramsauer et al. 2020 — Hopfield Networks is All You NeedAnswer thematic questions across the vault using map-reduce over community summaries. Global reasoning, not just retrieval.
Engram connectivity — Josselyn & Tonegawa 2020Auto-generated ~500-token context combining vault changes, git commits, top connected notes, and time context.
Contextual reinstatement theoryStructured Subject-Predicate-Object facts extracted from notes. Efficient factual lookup at ~10-20 tokens per triple.
Semantic memory organisation21-tool Model Context Protocol server. Works with Claude Code, Codex, Gemini CLI, Cursor, Windsurf - any MCP-compatible client. Provider-agnostic by design.
Standard transport: stdio, SSE, or HTTPAI agents write observations, decisions, conventions, learnings, context, and bugs into the vault via vault_remember / vault_forget / vault_memories. Memories have optional TTL, workspace scoping, and surface automatically in search results.
Episodic memory encoding — Tulving 1972Scope search and memory to subdirectories with --workspace flag or NEUROSTACK_WORKSPACE env var. Keeps project context isolated without separate databases.
Contextual binding - Howard & Kahana 2002Track AI coding sessions with start/end lifecycle hooks. Sessions group memories, enable summaries, and auto-clean orphaned sessions after crashes.
Episodic boundary detection - Zacks et al. 2007Two-tier insight extraction from session transcripts. Broad regex pre-filter selects candidates, then local LLM classifies and summarizes. Supports Claude Code, VS Code Chat, Codex CLI, Aider, and Gemini CLI session transcripts. Falls back to regex-only if Ollama is unavailable.
Consolidation during offline periods - Diekelmann & Born 2010Every core feature maps to established memory neuroscience research. This is not metaphor — the algorithms are direct computational analogues of biological memory mechanisms.
| Feature | Neuroscience Concept | Mechanism | Key Paper |
|---|---|---|---|
| Hot Notes | CREB-mediated excitability windows | Active notes attract preferential connections, mirroring how elevated CREB levels bias neuronal recruitment into engrams | Han et al. 2007, Science 316(5823), 457-460 |
| Drift Detection | Prediction error-driven reconsolidation | Notes surfacing in unexpected contexts trigger review, analogous to prediction errors destabilising consolidated memories | Sinclair & Bhatt 2022, PNAS 119(31) |
| Knowledge Graph | Engram connectivity networks | Wiki-links + PageRank model the interconnected neural ensembles that constitute distributed memory traces | Josselyn & Tonegawa 2020, Science 367(6473) |
| Communities | Hopfield attractor basins | Attractor basin community detection mirrors how memories settle into energy minima in Hopfield networks | Ramsauer et al. 2020, ICLR |
| Tiered Retrieval | Complementary learning systems | Progressive depth escalation models the hippocampal-neocortical memory consolidation pathway | McClelland et al. 1995, Psych. Review 102(3) |
Active notes attract preferential connections, mirroring how elevated CREB levels bias neuronal recruitment into engrams
Han et al. 2007, Science 316(5823), 457-460Notes surfacing in unexpected contexts trigger review, analogous to prediction errors destabilising consolidated memories
Sinclair & Bhatt 2022, PNAS 119(31)Wiki-links + PageRank model the interconnected neural ensembles that constitute distributed memory traces
Josselyn & Tonegawa 2020, Science 367(6473)Attractor basin community detection mirrors how memories settle into energy minima in Hopfield networks
Ramsauer et al. 2020, ICLRProgressive depth escalation models the hippocampal-neocortical memory consolidation pathway
McClelland et al. 1995, Psych. Review 102(3)
NeuroStack exposes a Model Context Protocol server with 21 tools. Works with Claude Code, Codex, Gemini CLI, Cursor, Windsurf, and any MCP-compatible client.
Primary retrieval tool — hybrid FTS5 + semantic search with tiered depth escalation. Start with depth="auto" and let it escalate based on coverage.
Get a pre-computed 2-3 sentence summary of a note (~100-200 tokens vs ~500-2000 for full file). Quick overview before committing to a full read.
Wiki-link neighbourhood with PageRank scoring. Discover related notes without manually following links.
Search structured Subject-Predicate-Object facts at ~10-20 tokens each. The most token-efficient retrieval method.
GraphRAG global queries over Leiden community summaries. Uses map-reduce synthesis for thematic questions that span the whole vault.
Index health metrics — coverage statistics for notes, chunks, embeddings, summaries, graph edges, triples, and communities.
Track note access for hotness scoring. Notes that get used more often rank higher in future searches.
Surface notes with retrieval anomalies — drift detection. Flags outdated content or notes appearing in unexpected contexts.
Write an observation, decision, convention, learning, context note, or bug into the vault. Memories surface automatically in future search results.
Remove a previously stored memory by ID. Use when a memory is outdated, incorrect, or no longer relevant.
List stored agent memories, optionally filtered by category or workspace. Memories are returned with their IDs for management.
Start a new memory session. Sessions group memories and enable lifecycle tracking with auto-harvest on end.
End a memory session. Optionally generates an LLM summary of session memories and auto-harvests insights from the transcript.
Compact ~500-token session context combining recent vault changes, git commits, external memories, top connected notes, and time context.
RAG Q&A with inline [[citations]]. Ask natural language questions, get answers grounded in vault content.
Find semantically similar notes by embedding distance. Discover notes you didn't know were connected.
Task-scoped context assembly within a token budget. Recovers working context after /clear or new conversation.
Quick-capture thoughts to vault inbox. Zero-friction note creation without leaving the AI session.
Update an existing memory in place. Modify content, tags, or category without delete+recreate.
Merge two memories with tag union and audit trail. Deduplicate while preserving provenance.
Extract insights from AI coding session transcripts. Two-tier regex + LLM classification pipeline.
Run neurostack init to choose your mode, vault location, and set up Ollama models. It auto-installs Ollama if needed and skips models you already have. Re-run anytime to upgrade.
FTS5 keyword search, wiki-link graph, stale note detection, and MCP server. No GPU needed.
Adds semantic embeddings, LLM summaries, triple extraction, and Hopfield-grounded ranking. Requires Ollama.
Zero GPU required. Install in lite mode, log in via browser, push your vault for server-side Gemini-powered indexing. Full semantic search without local ML.
Community mode: Add topic cluster detection with neurostack init --mode community (+15MB, leidenalg GPL-3.0).
Upgrade anytime: Re-run neurostack init to switch modes (e.g., lite → full). Uninstall: neurostack uninstall removes everything. Config is preserved so reinstall picks up where you left off.
Skip local GPU setup. NeuroStack Cloud indexes your vault server-side with Google Gemini — embeddings, summaries, and triples generated in minutes. Your local CLI and MCP server work the same, backed by the cloud-indexed database.
Install in lite mode, log in via browser, push your vault for server-side indexing, then pull the indexed database.
Manage your vault, API keys, usage, and billing from the web dashboard.
NeuroStack works with any MCP-compatible client. Use the local MCP server for full-featured access, or the cloud endpoint for browser-based tools.
Auto-configure Claude Desktop with a single command. All 21 MCP tools available locally via stdio transport.
Or manually add to your Claude Desktop config:
Connect NeuroStack directly to Claude.ai in your browser via the remote MCP endpoint. No local install required — just a NeuroStack Cloud account.
neurostack cloud pushhttps://mcp.neurostack.sh/mcpAuto-configure Claude Code for local MCP access.
~/.claude/.mcp.jsonOne-command setup for Cursor, Windsurf, VS Code, and Gemini CLI.
Any MCP-compatible client can also use the remote endpoint at https://mcp.neurostack.sh/mcp with OAuth authentication.
NeuroStack complements Obsidian as your editor — it adds the AI search engine layer that Obsidian lacks.
| Feature | NeuroStack | Obsidian | Khoj | Notion AI | |
|---|---|---|---|---|---|
| Full ~210MB | Lite ~130MB | ||||
| Local-first | Yes | Yes | Yes | Partial | No |
| AI-provider agnostic | MCP | MCP | No | Partial | No |
| Full-text search | FTS5 | FTS5 | Yes | Yes | Yes |
| Semantic search | Hybrid | No | Plugin | Yes | Yes |
| Knowledge graph | PageRank | PageRank | Backlinks | No | No |
| Community detection | Attractor basin | No | No | No | No |
| Drift detection | Yes | Yes | No | No | No |
| Tiered retrieval | Auto | No | No | No | No |
| AI summaries & triples | Yes | No | No | Partial | Yes |
| Hopfield ranking | Yes | No | No | No | No |
| CLI | Yes | Yes | No | Yes | No |
| MCP server | Yes | Yes | No | No | No |
| Agent memory | Yes | Yes | No | No | No |
| Session lifecycle | Yes | Yes | No | No | No |
| LLM harvest | Yes | No | No | No | No |
| Open source | Apache-2.0 | Apache-2.0 | Core only | Yes | No |
| GPU required | Optional | No | No | Optional | No |
| Price | Free | Free | $50/yr | Free/paid | $10/mo |
All config values support NEUROSTACK_* env var overrides: