Supercharge your AI Agents with persistent memory
Your digital developer twinβBecause "Explain Our Setup Again" Gets Old Fast
cd install
chmod +x start-recallium.sh
./start-recallium.shThat's it! Visit http://localhost:9001 to complete setup.
cd install
chmod +x start-recallium.sh
./start-recallium.shNote: The scripts use IPv6 dual-stack binding for Safari compatibility. If you encounter port binding errors on Linux, Docker may not have IPv6 enabled. Either enable IPv6 in Docker (
/etc/docker/daemon.jsonβ"ipv6": true) or edit the script to use0.0.0.0instead of[::]for port bindings.
Visit http://localhost:9001 to complete setup.
Windows requires additional Ollama configuration for Docker connectivity.
Step 1: Configure Ollama (one-time setup)
- Open "Edit the system environment variables" (search in Start menu)
- Click "Environment Variables" β Under "System variables" β Click "New"
- Variable name:
OLLAMA_HOST/ Variable value:0.0.0.0:11434 - Click OK β OK β Restart Ollama
Step 2: Add Firewall Rule (run as Administrator)
netsh advfirewall firewall add rule name="Ollama" dir=in action=allow program="C:\Users\<YOUR_USERNAME>\AppData\Local\Programs\Ollama\ollama.exe" enable=yes profile=privateStep 3: Start Recallium
cd install
start-recallium.batThe script auto-detects your IP and configures Ollama connectivity. Visit http://localhost:9001 to complete setup.
For more control, use docker-compose directly:
cd install
docker compose --env-file recallium.env pull
docker compose --env-file recallium.env up -dMagic Command: Once connected, just say "recallium" in your IDE and your AI loads all project context instantly.
- The Problem
- The Solution
- What You Get
- Installation
- Setup Wizard
- Web Dashboard
- How It Works (with examples)
- What's New in v0.8.285
- Previous Releases
- Complete Feature Set
- 16 MCP Tools
- Pro Tips
- Built-in Help System
Everyone's doing it nowβvibe coding with AI assistants. That magical flow where you're thinking out loud, the AI gets it, and code just happens.
But here's the catch: Your AI forgets everything. Every conversation starts from zero. Every project feels like explaining yourself to a stranger who forgot your last 47 conversations. Enter **Context-Engineering"
- π Endless repetition β "Here's how we handle auth..." (for the 50th time)
- π€· Context amnesia β AI forgets your patterns, preferences, and hard-won decisions
- ποΈ Project islands β Learning from one project never benefits another
- π Knowledge leakage β That brilliant debugging insight? Gone tomorrow.
Recallium is the persistent memory layer your IDE agents desperately need.
Instead of this painful cycle:
You: "Hey AI, remember we use PostgreSQL with..."
You: "Oh, and we prefer functional components..."
You: "Also, we decided against Redis because..."
You get this:
You: "Build auth for the checkout flow"
AI: "Using your battle-tested pattern from payment-service-v2:
β JWT with refresh token rotation
β httpOnly cookies (per your security audit decision)
β Redis mutex to prevent race conditions (you debugged this for 6 hours)
β 15min access tokens, 7-day refresh tokens
Implementing now..."
Your code. Your patterns. Your decisions. Remembered. Connected. Evolved.
Just say "recallium" in your IDE and watch your AI instantly load:
- Recent session activity and where you left off
- Project briefs, PRDs, and implementation plans
- Pending tasks that need attention
- Global and project-specific rules
One word. Full context. Zero repetition.
| Feature | Description |
|---|---|
| One Memory System | Works across Cursor, Claude Desktop/Code, VS Code, Windsurf, JetBrains, Zedβany MCP-compatible tool |
| Cross-Project Intelligence | Lessons learned once β Applied everywhere automatically |
| Document Knowledge Base | Upload PDFs, docs, specs β Your AI understands your system instantly |
| Smart Project Linking | Connect v1 β v2 β v3, share patterns across related projects |
| Zero-Config Memory | Auto-tags, auto-clusters, auto-learns from every interaction |
| Privacy-First | Runs locally, your data never leaves your machine |
| Free Local Option | Use Ollama + built-in embeddings for $0 API costs |
Complete Installation Guide - Get running in under 2 minutes
TL;DR:
cd install
# Option 1: One-click start (pulls latest, handles cleanup)
./start-recallium.sh # Linux/macOS
start-recallium.bat # Windows
# Option 2: Docker Compose (more control)
docker compose --env-file recallium.env pull
docker compose --env-file recallium.env up -dThen configure your IDE (see install guide) and start coding with persistent memory!
Supported IDEs: Cursor β’ Claude Desktop β’ Claude Code β’ VS Code β’ Windsurf β’ Roo Code β’ Visual Studio 2022 β’ JetBrains β’ Zed β’ Cline β’ BoltAI β’ Augment Code β’ Warp β’ Amazon Q β’ AntiGravity β’ and more...
See install/README.md for detailed setup instructions for your IDE.
First-time setup takes 2 minutes via the web dashboard at http://localhost:9001
Setup Wizard: Guided configuration for first-time users
Provider Selection: Choose from Ollama (free), OpenAI, Anthropic, Gemini, or OpenRouter
Recallium works with any LLM providerβuse what you already have:
| Provider | Models | Notes |
|---|---|---|
| Anthropic | Claude 3.5 Sonnet, Claude 3 Opus/Sonnet/Haiku | Recommended for best results |
| OpenAI | GPT-4o, GPT-4 Turbo, GPT-3.5 Turbo | Function calling, streaming |
| Google Gemini | Gemini 1.5 Pro, Gemini 1.5 Flash | Multi-modal support |
| Ollama | Llama 3, Mistral, Qwen, any local model | Free, runs locally |
| OpenRouter | 100+ models via single API | Access any model |
Want to run completely free and private?
LLM: Ollama (local models like Llama 3, Mistral)
Embeddings: GTE-Large (built-in, runs locally)
Provider Priority: Set up automatic failover between providers
- Test before you commit β Verify API keys work before saving
- Switch providers anytime β Change LLM providers without losing memories
- Automatic failover β Configure backup providers for reliability
- Secure credential storage β API keys stored in encrypted vault
- Hot reload β Provider changes take effect immediately
Recallium uses GTE-Large (1024 dimensions) for embeddings:
- Runs locally on your machine
- No API costs, no rate limits
- Works on CPU, CUDA, or Apple Silicon (MPS)
Access the full-featured UI at http://localhost:9001
Dashboard: Your command center with system stats, recent activity, and quick insights
| Page | What You Can Do |
|---|---|
| Dashboard | Overview, stats, recent activity, quick insights |
| Memory Explorer | Search, filter, bulk operations, tag management |
| Projects | Briefs, documents, links, timeline, tasks |
| Insights | Pattern analysis, cluster visualization, recommendations |
| Analytics | Token usage, activity trends, cost tracking |
| Documents | Upload PDFs/DOCX, search chunks, manage files |
| Tasks | Create, track, link memories to tasks |
| Memory Graph | Visual relationship mapping between memories |
| Thinking | View reasoning sequences, explore branches |
| Providers | Configure LLM & embedding providers |
| Admin | System logs, configuration, maintenance |
Projects Page: Central hub for managing knowledge repositories with activity indicators, memory counts, and project linking
Memory Graph: Interactive force-directed visualization showing relationships between memories, clusters, and tags
You (in Cursor): "Search my memories about authentication patterns"
Cursor: "Found 12 memories across 3 projects:
payment-service-v2 (2 months ago):
β JWT with refresh token rotation
β Redis mutex prevents race conditions
β You debugged concurrent refresh calls for 6 hours
user-api-v1 (4 months ago):
β Security audit: never use localStorage for tokens
β Team decision: httpOnly cookies only
admin-dashboard (1 month ago):
β 15min access tokens, 7-day refresh tokens
β Auto-refresh 2 minutes before expiry"
You: "Build auth for my checkout flow"
Cursor: [Generates code using YOUR battle-tested patterns]
Memory Explorer: Search across all your stored decisions, patterns, and learnings
You (in Claude Desktop): "Get insights on database patterns across my projects"
Claude: "Analyzed 89 memories across 5 projects. Pattern detected:
You ALWAYS follow this progression:
1. Start with PostgreSQL (projects: 5/5)
2. Hit connection pool limits around 10K users (projects: 3/5)
3. Add PgBouncer to solve it (projects: 3/3 that hit the issue)
Current project: user-service-v3
Expected scale: 50K users
Recommendation: Include PgBouncer from day 1. You'll need it,
and adding it later caused 2 hours of downtime in admin-api."
Insights Dashboard: Discover patterns, recurring issues, and recommendations across projects
You (in Cursor): "Search my documents about webhook security"
Cursor: "Found in 'API Security Guidelines v2.pdf' (uploaded last week):
Page 47: Webhook Signature Verification
β MUST use HMAC-SHA256
β Signature in X-Webhook-Signature header
β Replay attack protection: 5-minute timestamp window
β Rate limiting: 100 webhooks/minute per endpoint
Also found related memory from payment-service-v1:
β You got burned by missing signature verification
β Production incident: fake webhook drained test account
I'll implement webhooks with proper security from the start."
You (in Claude Desktop): "Search my memories and documents about
error handling, include linked projects"
Claude: "Searching across:
- Current project: ecommerce-api
- Linked projects: payment-service-v2, order-management
- Uploaded documents: 3 PDFs, 12 markdown files
From documents:
β 'Engineering Standards.pdf': Error codes must follow
ERR-{DOMAIN}-{CODE} format (e.g., ERR-AUTH-001)
From payment-service-v2 memories:
β You built a standardized error response wrapper
β Includes: error code, message, trace_id, timestamp
β Logs to Sentry with context
From order-management memories:
β You added retry logic with exponential backoff
β Max 3 retries, 2^n second delays
I can implement error handling using all these patterns."
You (in Cursor): "Store a memory: We decided against microservices.
Team is too small (3 devs). Maintenance overhead
not worth it. Monolith with clear module boundaries
instead."
Cursor: "β Stored as decision with high importance.
I'll suggest monolith architecture in future projects."
You: "Link payment-service-v1 to checkout-api as dependency"
You: "Search my memories about Stripe webhooks, include linked projects"
Cursor: "Found context from payment-service-v1:
Decision (6 months ago):
β Stripe webhooks need idempotency keys
β You discovered this after processing same payment twice
β Solution: Store event IDs in Redis with 24hr TTL
Bug fix (3 months ago):
β Race condition with concurrent webhook deliveries
β Fixed with Redis distributed lock
Applying both lessons to checkout-api implementation..."
You: "Create task: Fix mobile login button not responding"
Cursor: "β Task created (ID: task_847)
Want to link related memories?"
You: "Yes, link to yesterday's auth debugging session"
---
Next day:
You: "Show my tasks"
Cursor: "1 pending task:
β Fix mobile login button not responding
β Related: Auth debugging session, JWT refresh race condition
Should I expand the debugging session memory?"
You: "Yes"
Cursor: "From debugging session:
β Issue: JWT refresh happens mid-form-submit
β Button disabled during API call
β Race condition prevents re-enable
β Solution discussed: Debounce + optimistic UI
Ready to implement the fix?"
You (in Claude Desktop): [Uploads "API Documentation v3.pdf" - 10MB]
Claude: "Processing... Done in 28 seconds.
Created 868 chunks from 247 pages.
I now understand:
β OAuth 2.0 implementation (section 3)
β Rate limiting rules (section 7.2)
β Webhook signature verification (section 9)
β Database sharding strategy (section 12)
β Error code taxonomy (appendix B)
All stored with semantic search enabled."
---
Later, in Cursor:
You: "Build the OAuth refresh token endpoint"
Cursor: [Generates code matching section 3 of your PDF exactly]
"Implementation follows your API docs:
β 15-minute access tokens
β 7-day refresh tokens
β Rotation on every refresh
β httpOnly, Secure, SameSite=Strict cookies"
Documents Page: Upload PDFs, specs, and docs that become instantly searchable by all your AI agents
You (starting a new session): "recallium"
Claude: "Welcome back! Loading your context...
π Project: ecommerce-api
π Recent: 12 memories in last 7 days
Last session (2 days ago):
β Implemented JWT refresh token rotation
β Fixed race condition in payment webhook
β Decision: Using Redis for session storage
π Pending Tasks:
β Add rate limiting to checkout endpoint
β Write tests for payment flow
π Active Rules:
β Always confirm destructive operations
β Search memories before implementing
Ready to continue. What would you like to work on?"
You: "What was I working on last week?"
Claude: "Here's your session recap for ecommerce-api:
Progress (7 days):
β 15 memories stored
β 3 tasks completed, 2 pending
β Most active: authentication module
Key Decisions:
β JWT with 15min access, 7-day refresh tokens
β Redis for session storage (not Memcached)
Recent Learnings:
β Webhook signature verification prevents replay attacks
β PgBouncer needed at 10K+ concurrent users
Shall I expand on any of these?"
You: "Search Redis caching patterns"
Claude: "Found 47 memories from user-service-v1.
But I also see newer memories showing you migrated
to Memcached in user-service-v2 last month.
Should I inactivate the old Redis memories?"
You: "Yes, inactivate them. Reason: migrated to Memcached"
Claude: "β Inactivated 47 Redis memories (soft delete, recoverable)
β Audit trail preserved
β I'll suggest Memcached patterns by default now
Your context stays current as your stack evolves."
Memory Management: Update, archive, or restore memories as your stack evolves
Bring your own LLM with Palantir Foundry AIP:
- Palantir Foundry AIP now supported as a Bring-Your-Own-LLM provider
- Seamlessly integrates with Recallium's multi-provider failover system
- Works with your existing Palantir enterprise deployment
Improved search accuracy and relevance:
- Smarter hybrid search combining semantic, keyword, and tag matching
- Faster results with reduced storage footprint
- More accurate ranking for better discovery
Native support for Claude Desktop with Extensions:
- One-click MCP connection at
http://localhost:8001/mcp - Works with 60+ IDEs including VS Code, JetBrains, Neovim, and more
- Seamless integration with Claude Desktop's new Extensions feature
Dashboard: Your command center for memories, projects, insights, and analytics
Use any LLM provider you wantβno lock-in:
- 5 Built-in Providers: Anthropic, OpenAI, Google Gemini, Ollama, OpenRouter
- Free Local Option: Ollama + GTE-Large = zero API costs
- Setup Wizard: Test and configure providers via web UI
- Hot Switching: Change providers anytime without losing data
- Automatic Failover: Configure backup providers for reliability
Full-featured UI at http://localhost:9001:
- Memory Explorer with advanced search
- Project management with briefs, docs, timeline
- Insights visualization and pattern analysis
- Analytics with token usage tracking
- Document library with chunk management
- Memory graph visualization
- Provider configuration and testing
Expanded from 10 to 16 tools:
recallium- Magic summon: load ALL context in one callsession_recap- Resume work with recent activity summaryhelp- 12 topics with progressive disclosurecheck_setup_status- Verify system readinessvalidate_intelligence_usage- Best practices complianceintelligence_health_check- Usage optimization tips
Direct HTTP connections for modern IDEsβno npm client needed for most tools.
What changed:
- Native HTTP/SSE transport for real-time streaming
- Cursor, VS Code, Claude Code, Windsurf work out of the box
- Optional npm client for command-only IDEs (Claude Desktop, Zed, JetBrains)
- Single unified MCP endpoint:
http://localhost:8001/mcp
Turn any PDF, DOCX, or code file into searchable agent knowledge.
What you get:
- Multi-format: PDF, DOCX, Markdown, CSV, Excel, all code files
- Lightning fast: 10MB PDF β searchable in < 30 seconds
- Smart chunking: 1000-token chunks with 128-token overlap
- Hybrid search: Vector similarity + keyword + tag matching
- Zero API costs: YAKE keyword extraction (9s for 868 chunks)
- Upload once: Every IDE agent gets instant access
Real impact:
- Architecture docs β Agent understands your system design
- API specs β Agent generates spec-compliant code
- Security guidelines β Agent follows your security rules
- Past project docs β Agent learns from your history
Connect projects. Share intelligence. Learn everywhere.
Link types:
dependency- Shared libraries (auth-lib used by 5 projects)parent/child- Version iterations (v1 β v2 β v3)fork- Experimental branches (compare approaches)template- Starter patterns (reusable scaffolds)related- Similar domains (all payment services)
What you get:
- Many-to-many relationships
- Recursive search up to 5 levels deep
- Auto-detection of related projects (finds shared patterns)
- Visibility controls (full access, metadata-only, hidden)
The power:
Project graph:
auth-library-v2 (dependency)
ββ Used by: ecommerce-api
ββ Used by: admin-dashboard
ββ Used by: mobile-backend
ββ Links to: auth-library-v1 (parent)
One lesson learned β Applied to 4 projects automatically
Your patterns evolve. Your memory should too.
Three actions:
update- Modify content, tags, importance, type (partial updates)inactivate- Soft delete with reason (recoverable, audit trail)reactivate- Restore archived memories
What you get:
- Full change history with timestamps
- Reasons for every inactivation
- Prevents knowledge drift as your stack evolves
- Keeps context clean and relevant
Set guardrails once. Every IDE agent follows them forever.
How it works:
- Store rules in special
__global__project - Rules auto-load during MCP handshake (before first user message)
- Apply to ALL projects, ALL agents, ALL time
4 default rules included:
-
Always Confirm Destructive Operations Ask before deleting files, dropping tables, force-pushing git
-
Never Implement Fallbacks Without Asking Report failures, don't substitute silently
-
Proactively Search Memories Before Implementing Auto-search for relevant context, past patterns, known issues
-
Load Rules on Initialization Ensure consistent behavior across all sessions
Know your knowledge costs. Optimize your memory footprint.
Dashboard shows:
- Total tokens across all memories and documents
- Per-project breakdown with memory counts
- Memory type distribution (decisions vs code vs docs)
- Growth trends over time
- Token-efficient vs token-heavy projects
Why it matters:
- Future-ready for OpenAI-compatible pricing
- Identify bloated projects needing cleanup
- Optimize what you store vs what you inactivate
- Understand where your context budget goes
- store_memory - Save decisions, patterns, learnings, bugs
- search_memories - Semantic + keyword + tag hybrid search
- expand_memories - Progressive disclosure (search IDs β full content)
- modify_memory - Update, inactivate, reactivate with audit trail
- get_insights - Cross-project patterns, quality analysis, technical debt
- start_thinking / add_thought - Structured reasoning sequences
- get_rules - Load global and project-specific guardrails
- Upload documents - PDF, DOCX, Markdown, CSV, Excel, code files
- Search documents - Hybrid vector + keyword + tag search
- Expand chunks - Get surrounding context (Β±1 or Β±2 chunks)
- Smart chunking - 1000 tokens with 128-token overlap
- manage_project - Briefs, PRDs, implementation plans
- Link projects - Dependencies, forks, parent/child, templates
- manage_tasks - Create, complete, list with linked memories
- Token analytics - Usage tracking per project
- 11 memory types - working-notes, decision, rule, code-snippet, learning, research, discussion, progress, task, debug, design
- 3 search modes - semantic (concepts), keyword (exact), hybrid (both)
- File path filtering - Find memories by code file patterns
- Linked project search - Discover context across relationships
- Search target - memories only, documents only, or both
- Recent filtering - Last 30 days only
- 4x larger chunks: 1000 tokens vs 250 (more context per chunk)
- 95% clean splits: <5% mid-sentence breaks (was ~30% in v0.2.5)
- Lightning tags: 9 seconds vs 15-43 minutes (YAKE-only approach)
- Indexed queries: Instant project link traversal
- Thread pooling: Concurrent embedding generation
- Three-table RAG: documents (metadata) β chunks (content) β memories (notes)
- Event-driven jobs: Async, non-blocking background processing
- Graph traversal: PostgreSQL recursive CTEs for project relationships
- HTTP/SSE transport: Real-time streaming with Server-Sent Events
- Enterprise foundation: Row-Level Security policies (multi-tenancy ready)
document_chunkstable with pgvector embeddingsproject_linkstable for many-to-many relationshipsmemory_audit_logfor complete change history- GIN index on tags for instant tag searches
- JSONB
token_metadatafor analytics - Lifecycle columns:
is_active,inactivated_at,inactivation_reason
Release History
| Version | Changes | Urgency | Date |
|---|---|---|---|
| main@2026-04-19 | Latest activity on main branch | High | 4/19/2026 |
| v0.8.368 | Latest release: v0.8.368 | High | 4/8/2026 |

