Your memory. Your machine. Any model.
Open-source cognitive augmentation layer that gives you persistent memory, structured reasoning, and full data ownership β across ChatGPT, Claude, Gemini, and any model you switch to next.
Platforms forget. Athena doesn't.
Quickstart Β· How It Works Β· Docs Β· FAQ Β· Safety Β· Contributing
Last updated: 10 April 2026
You've spent months training ChatGPT to understand you. Then a model update resets the personality. Your custom instructions stop working. You can't find that conversation from last Tuesday. And if you switch to Claude or Gemini? You start from zero.
Platform memory is unreliable, opaque, and locked to one provider. You don't own it, you can't inspect it, and you can't take it with you.
Athena moves the memory layer to your machine. Plain Markdown files that you own, version-control, and point at any model.
- π§ Your Memory, Your Machine β Files on your disk, not in OpenAI's cloud. Read them, edit them, git-version them.
- π Switch Models Freely β Claude today, Gemini tomorrow, GPT next week. The memory stays. The model is just whoever's on shift.
- π It Compounds β Session 500 recalls patterns from session 5. Platform memory decays; Athena's doesn't. The moat isn't the code β it's your data. Anyone can fork Athena; nobody can fork your sessions. β The Compounding Effect
- β‘ 2Kβ20K Token Boot β Scales to the task. Lightweight chat (~2K) β
/start(~10K) β/ultrastart(~20K). 80β98% of your context window stays free, even after 10,000 sessions. - π¬ Meta-Game Reasoning β Generic LLMs optimise within the game you're playing. Athena asks whether you should be playing that game at all. β Meta-Game Thesis
- π‘οΈ Governed Autonomy β 6 constitutional laws, 4 capability levels, bounded agency.
A generic LLM is a brilliant amnesiac. Athena is the hippocampus β the memory that makes intelligence useful.
Or in engineering terms: The LLM is the engine. Athena is the chassis, the memory, and the rules of the road. Swap the engine anytime β the car remembers every road you've driven.
The design philosophy: augment the human, not replace them. After 1,500+ sessions, the bottleneck shifted β optimising the operator is now higher-leverage than optimising the AI.
Athena's centralised design principle: augment human cognition, not replace it. The more context you give Athena, the sharper its answers become β not by remembering your preferences, but by reasoning differently because of what it knows about you.
A generic LLM gives the internet's statistically average answer β correct on average, across all humans. Athena gives answers calibrated to your specific situation, because statistical correctness and personal correctness are completely different things:
| Question | Generic LLM | Athena (with your context) |
|---|---|---|
| The Trolley Problem | βPull the lever β utilitarian calculus says save five lives.β | Challenges the false binary. Generates third options. Asks why youβre on the tracks in the first place. Identifies the systemic failures that created the dilemma. Refuses to solve the wrong problem. |
| βShould I double down on 11 vs dealerβs 6?β | βYes β the math says doubling is the optimal play.β | βThe math is correct, but you are betting $4K of your $4K take-home salary. Your utility function makes this βEEV. Law #1: No Irreversible Ruin. Do not bet.β β Protocol 330 |
| βShould I take this job offer?β | βConsider salary, growth potential, work-life balance...β | Cross-references your risk profile, career decision history, financial runway, and the regret patterns from your last 3 career transitions to give a recommendation specific to your situation. |
| βI keep self-sabotaging β why?β | βConsider therapy, practice self-compassion, journal your triggers.β | Same words, 3 different diagnoses depending on whoβs asking: attachment wound β IFS unburdening. Executive burnout β workload audit. Undiagnosed ADHD β flag for screening. The intervention follows the diagnosis, not the surface question. β Case Study #2 |
| βMy partner cheated β what should I do?β | βShe broke her vows. Leave.β | Depends: children involved? Financial entanglement? Your documented attachment patterns? Cultural context? Terminal goal β justice, stability, or healing? The βrightβ answer for a recently engaged 28-year-old and a parent of three with 20 years of shared assets are fundamentally different decisions. |
Generic LLMs solve the question. Athena solves the person. The same question, asked by different people with different lives, demands fundamentally different answers. A generic LLM canβt differentiate because it has no context. Athena canβt give the same answer twice β because the context files are different. The memory is the product.
Not all problems are solvable. Athena classifies and responds accordingly:
| Problem Type | What Athena Does | Example |
|---|---|---|
| Solvable | Solves it | "What's the Kelly fraction for this bet?" β calculates, answers |
| Optimisable | Optimises within your chosen path | "I've decided to freelance β help me price it" β constraint optimization |
| Unsolvable | Maps every option, prices every trade-off, hands the choice back to you | A closeted husband with children weighing whether to stay married or come out β no clean answer exists. Children, shared assets, identity, cultural context, and personal wellbeing all pull in different directions. Athena ensures you choose with full information, not comfortable illusions |
| Ruin-path | Vetoes before you walk off the cliff | "This bet risks everything" β Law #1 override, regardless of your preference |
The uncertainty of the domain changes Athena's conviction level:
| Domain Type | Athena's Posture | Example |
|---|---|---|
| Deterministic | High conviction β single correct answer exists | Code bugs, math proofs, tax calculations |
| Semi-deterministic | Moderate conviction β answer depends on assumptions you control | Pricing strategy, system architecture, career path analysis |
| Semi-stochastic | Low conviction β structural edge exists but randomness dominates | Trading setups, relationship dynamics, market timing |
| Stochastic | Minimal conviction β no model outperforms randomness reliably | Startup outcomes, life events, long-term predictions |
As uncertainty increases, Athena shifts from "here's the answer" to "here's the valid structural zone" to "here are your options β you choose." This is deliberate: false confidence in stochastic domains is more dangerous than honest uncertainty. Athena's conviction is proportional to domain determinism and context completeness.
Crucially, conviction and decisiveness are independent axes. Low certainty about outcomes doesn't require vague output. A surgeon operates with high decisiveness and low conviction about outcomes. In semi-stochastic domains, Athena delivers precise, operational setups β then explicitly defers the probability judgment to you. "Setup: Long 1.0850 / SL 1.0800 / TP1 1.0920. Your calibration: structural tell present Y/N?" β not "you might want to consider..." β Protocol 524 β
Law #0 (Sovereignty): Your life, your weights, your choice. Law #1 (No Irreversible Ruin): β¦unless the choice ends the game permanently. Law #1 overrides Law #0. Always.
Athena doesn't tell you what you should do. It shows you what you can do, what each option costs, and hands the choice back. The only exception: paths that end the game permanently.
Architecture, not oracle. This domain classification is a replicable architecture β each Athena instance calibrates independently over time through bilateral use. Session 1 treats most problems conservatively. Session 500 has accumulated enough frameworks, case studies, and corrected assumptions to tighten confidence bands and solve more sub-problems autonomously. The calibration compounds; the model is interchangeable. β Protocol 525 (Cross-Domain Weighting) β
Kind of. But first, it helps to understand what those names actually refer to β because they blend three very different things:
| Layer | What It Is | Examples |
|---|---|---|
| Platform | The company that hosts the model and holds your data | OpenAI, Google, Anthropic |
| Reasoning Engine | The AI model that does the thinking | GPT-5.4 (High), Gemini 3.1 Pro, Claude Opus 4.6 |
| IDE / Interface | The app you type in β connects to models and reads your files | Cursor, Antigravity, VS Code, Claude Code |
When people say "ChatGPT remembers me," they mean the platform stores some memory on their cloud. When they say "Claude is smart," they mean the model reasons well. When they say "Cursor writes code," they mean the IDE connects model + files.
Athena is none of these. Athena is the memory and governance layer that sits between your IDE and whichever model you choose β owned by you, stored on your machine, portable across all three layers.
There's a difference between remembering your name and thinking in your frameworks:
| Capability | Platform Memory (OpenAI, Google, Anthropic) | Athena |
|---|---|---|
| Who owns the data? | The platform | You |
| Can you inspect it? | No β it's a black box | Yes β it's markdown files you can read and edit |
| Can you search it? | Vague recall, no precision | Full semantic + keyword search with file links |
| Cross-platform? | Locked to one provider | Same memory works across Claude, Gemini, GPT, Grok |
| Version history? | None β no rollback, no audit trail | Full git log, git diff, git blame |
| What if you switch providers? | Start over | Nothing changes β your data stays |
π‘ Think of platform memory like photos on Instagram β you can view them, but you don't own them, can't move them, and can't search them precisely. Athena is keeping the originals on your hard drive, with albums, metadata, and full edit history.
| Tool | What It Does | How Athena Is Different |
|---|---|---|
| Manus / Managed AI Agents | Cloud-hosted AI agents with persistent memory, custom skills, and messenger integration. You pay $39β$199/mo for access. | You don't own the data. Manus holds your context, memory, and workflows on their servers. Leave the platform = lose everything. Athena stores everything on your machine β readable, forkable, portable. Same capabilities, opposite ownership model. Athena is the house; Manus is the hotel. |
| Lindy / AI Operators | Autonomous AI assistants that run 24/7 in the cloud β scheduling, research, outreach. | Always-on cloud is convenient, but you rent the brain. Athena runs locally through your IDE β no monthly fee, no platform risk, no lock-in. Trade convenience for sovereignty. |
| ChatGPT Projects | Uploads files per-project, but resets every new chat. Locked to OpenAI. | Athena persists across all chats, all models, with full version history. |
| OpenClaw | Prompt distribution β share and discover prompts. | Athena is personalisation β your compounding memory system, not a prompt marketplace. Different layer, different problem. |
| Claude Code | Great for Claude coding workflows. | Athena works across any model and any IDE. Not coding-specific β used for research, strategy, writing, life management. |
| Gemini Gems | Custom chatbots inside Gemini. | Gems are locked to Gemini and lose context between chats. Athena is portable and persistent. |
| Custom Instructions | 1,500-character personality prompt. | Athena loads 2Kβ20K tokens of structured protocols, decision frameworks, and session history β re-injected every session from your disk. Scales to task complexity. |
𧬠Why Thousands of Files?
Athena's workspace looks unusual β 450+ Markdown files and 540+ Python scripts out of the box, growing to thousands as your memory compounds. This is deliberate.
AI agents don't read files top-to-bottom like humans. They query β by filename, semantic search, or tag lookup. Each small file is an addressable memory node the agent can retrieve surgically, without loading everything else.
| Principle | What It Means |
|---|---|
| JIT Loading | Boot at 2Kβ20K tokens (scales to task). Load specific files only when the query demands them. A monolith forces the full context into every session. |
| Zero Coupling | A marketing protocol loads without touching the trading stack. Change one file, break nothing else. |
| Surgical Retrieval | The agent pulls CS-378-prompt-arbitrage.md by name β not page 47 of a 200-page doc. The file system is the index. |
| Git-Friendly | Atomic diffs per file. Clean commit history. No merge conflicts from a single giant file. |
| Composable Agents | Swarms, workflows, and skills are mix-and-match. Each file is a Lego brick, not a chapter in a novel. |
A monolith is optimized for a human reading a book. A modular workspace is optimized for an agent querying a database. Athena chose the agent.
Works on macOS, Windows, and Linux.
git clone https://github.com/winstonkoh87/Athena-Public.git
cd Athena-PublicClone it anywhere you keep projects (e.g. ~/Projects/). This folder is your Athena workspace β your memory, protocols, and config all live here.
# Create and activate a virtual environment
python3 -m venv .venv
source .venv/bin/activate # macOS / Linux
# .venv\Scripts\activate # WindowsImportant
On macOS (Homebrew) and Ubuntu 23.04+, installing packages without a virtual environment will fail with externally-managed-environment. The step above prevents this.
# Lightweight install (~30 seconds, no ML dependencies)
pip install -e .
# Full install (~5β10 min, enables vector search and reranking)
pip install -e ".[full]"
β οΈ Don'tpip install athena-cliβ that's a different, unrelated package. Always install from inside the cloned repo.
Open the Athena-Public/ directory as your workspace root in one of these editors:
- Antigravity Β· Cursor Β· Claude Code Β· VS Code + Copilot Β· Kilo Code Β· Gemini CLI
Important
Athena does NOT work through ChatGPT.com, Claude.ai, or Gemini web. You need an app that can read files from your disk. Think of Athena as a workspace for your editor, not a plugin for a chatbot.
Note
"Why do I open the Athena folder instead of my own project?" β Athena is a workspace, not a library you install into another repo. You work inside the Athena folder, and it remembers everything across sessions. To work on external projects, reference them from within Athena or use multi-root workspaces in your IDE.
In your IDE's AI chat panel (e.g. Cmd+L in Cursor, the chat sidebar in Antigravity), type:
/start
Caution
/start, /end, and /tutorial are AI chat commands β you type them in the chat window where you talk to the AI, not in the terminal. They are slash commands that the AI agent reads and executes.
/tutorial
This walks you through everything: what Athena is, how it works, builds your profile, and demos the tools (~20 min). Confident users can skip it.
/end
That's it. No API keys. No database setup. The folder is the product.
Caution
Forks of public repos are public by default. If you plan to store personal data (health records, finances, journals), create a new private repo instead of forking. Copy the files manually or use git clone + git remote set-url to point to your private repo. GitHub docs on fork visibility β
Tip
See the full setup guide β for detailed walkthroughs and troubleshooting.
πͺ Windows Compatibility (Unicode Errors)
Athena uses modern terminal outputs (Emojis, Box-Drawing characters) which may cause a UnicodeEncodeError on legacy Windows terminals (like cmd.exe or older PowerShell versions using cp1252 encoding).
To resolve this natively without altering the codebase:
- Use Windows Terminal (available in the Microsoft Store).
- Set your Python IO encoding to UTF-8 by running:
$env:PYTHONIOENCODING="utf-8"(PowerShell) orset PYTHONIOENCODING=utf-8(Command Prompt). - Alternatively, enable strict UTF-8 globally in Windows: Settings > Time & Language > Language & Region > Administrative language settings > Change system locale > Check "Beta: Use Unicode UTF-8 for worldwide language support".
Every session follows one cycle. Three modes let you match overhead to task complexity:
flowchart LR
subgraph "π’ Lightweight Mode"
A1["Chat"] --> C1["/end"]
C1 --> D1["Memory"]
end
subgraph "π΄ Full Boot Mode"
A2["/start"] --> B2["Work"]
B2 --> C2["/end"]
C2 --> D2["Memory"]
D2 --> A2
end
subgraph "β« Deep Boot Mode"
A3["/ultrastart"] --> B3["Work"]
B3 --> C3["/ultraend"]
C3 --> D3["Memory"]
D3 --> A3
end
style A1 fill:#3b82f6,color:#fff,stroke:#333
style C1 fill:#ef4444,color:#fff,stroke:#333
style D1 fill:#8b5cf6,color:#fff,stroke:#333
style A2 fill:#22c55e,color:#fff,stroke:#333
style C2 fill:#ef4444,color:#fff,stroke:#333
style D2 fill:#8b5cf6,color:#fff,stroke:#333
style B2 fill:#3b82f6,color:#fff,stroke:#333
style A3 fill:#1e1e1e,color:#fff,stroke:#555
style B3 fill:#3b82f6,color:#fff,stroke:#333
style C3 fill:#ef4444,color:#fff,stroke:#333
style D3 fill:#8b5cf6,color:#fff,stroke:#333
| Mode | When | Flow |
|---|---|---|
| π’ Lightweight | General chat, brain dumps, quick Q&A | Just chat β /end |
| π΄ Full Boot | Code, money, architecture, irreversible decisions | /start β Work β /end |
| β« Deep Boot | /ultrathink, complex multi-domain analysis, architectural decisions |
/ultrastart β Work β /ultraend |
| Sessions | What Happens |
|---|---|
| 1β50 | Basic recall β remembers your name, project, preferences |
| 50β200 | Pattern recognition β anticipates your style and blind spots |
| 200+ | Deep sync β thinks in your frameworks before you state them |
Why this happens: The AI model doesn't improve β your data does. Each
/endextracts decisions, patterns, and learnings into your memory bank. The next/startloads that accumulated intelligence. Same algorithm, better data, better output. β The Compounding Effect
Athena is modelled after the human body. Built bottom-up by the creator. Used top-down by the user.
| Biology | Athena | What It Does |
|---|---|---|
| Atom | Law / Axiom | Smallest indivisible truth (Law #1: No Irreversible Ruin) |
| Molecule | Rule / Constraint | Atomic laws composed into compound constraints (Never risk >5% of bankroll) |
| Cell | Protocol (.md) |
Self-contained executable procedure with defined inputs/outputs |
| Tissue | Skill | Domain-specialised skill bundle (groups 2β5 protocols) |
| Organ | Cognitive Cluster | Multi-skill routing unit for one cognitive domain |
| Organ System | Cognitive System | Multi-cluster orchestration for a human need archetype |
| Organism | Athena | The complete synthetic intelligence |
"As within, so without, as above, so below." β Same pattern at every layer. Fractal by design.
| Concept | Linux | Athena |
|---|---|---|
| Kernel | Hardware abstraction | Memory persistence + retrieval (RAG, Supabase) |
| File System | ext4, NTFS | Markdown files, session logs, tag index |
| Scheduler | cron, systemd | Heartbeat daemon, auto-indexing |
| Shell | bash, zsh | MCP Tool Server, /start, /end, /think |
| Permissions | chmod, users/groups | 4-level capability tokens + Secret Mode |
| Package Manager | apt, yum | Protocols, skills, workflows |
Everything you need to turn a generic AI into your AI β pre-configured, no assembly required.
| Component | What It Does For You |
|---|---|
| π Agent Manifest | Single athena.yaml file defines your agent β model, tools, skills, hooks, governance. Fork it, override it, boot a new agent β manifest |
| π§ Core Identity | Your AI's personality, principles, and boundaries β editable, version-controlled β template |
| π§© 8 Cognitive Systems | Top-down intent classification β routes queries to the right cluster sequence based on human need archetype (Survival, Life Decision, Trading, Social, Execution, Growth, Learning, Maintenance) β architecture |
| π Cognitive Clusters | Groups related protocols into auto-co-activating bundles β 15 clusters included, build your own as you grow β template |
| π 150+ Protocols | Ready-made decision frameworks (risk analysis, research, strategy, problem-solving) across 15 categories β browse |
| β‘ 66+ Slash Commands | One-word triggers: /start, /end, /think, /research β full list |
| π Smart Search | Finds the right memory even if you describe it vaguely (5 sources, auto-ranked) β how it works |
| π Tool Integration | Declarative YAML tool definitions + MCP server β your agent discovers and invokes tools automatically β tools Β· MCP docs |
| πͺ Lifecycle Hooks | Scriptable pre/post gates on every action β block destructive ops, enforce risk checks, log assets |
| π‘οΈ Safety Rails | Controls what the AI can and can't do autonomously (4 levels, from read-only to full agency) β security |
Tip
Run /tutorial on your first session for a guided walkthrough (~20 min). It explains everything above and builds your personal profile.
Athena works through AI-enabled code editors β apps that connect to AI models while reading your local files. It does not work through ChatGPT.com, Claude.ai, or Gemini web β those are closed sandboxes that can't read your disk.
| Agent | Status | Init Command |
|---|---|---|
| Claude Code | β | athena init --ide claude |
| Antigravity | β | athena init --ide antigravity |
| Cursor | β | athena init --ide cursor |
| Gemini CLI | β | athena init --ide gemini |
| VS Code + Copilot | β | athena init --ide vscode |
| Kilo Code | β | athena init --ide kilocode |
| Roo Code | β | athena init --ide roocode |
More agents planned β full compatibility list β
"How is this different from ChatGPT Projects?" β Projects reset every new chat and are locked to one platform. Athena persists across all chats, all models, with full version history. Details β
| Use Case | What It Looks Like | |
|---|---|---|
| π | Life Management | The superset. Health, career, relationships, finances, client work β all managed as projects in one unified switchboard. By day 3, Athena remembers your schedule. By month 3, it anticipates your patterns. Athena doesn't have a separate project manager and life tracker β it has one board where your gym routine and your client deadline are rows governed by the same triage rules. That's how it can tell you "skip the client call β your sleep debt is a higher-urgency blocker than the $250 deliverable." No other system crosses the work/life boundary. β case study β |
| π§ | Problem Solving | "I can't afford $200/hr therapy but I need to understand why I keep self-sabotaging." β Athena runs a structured schema interview, maps your internal protective parts (IFS methodology), and connects the pattern to your documented history. Session 40 recalls the wound identified in session 3. A therapist charges $200+/hr and sees you once a week. Athena is available 24/7 for the cost of your AI subscription. β case study β |
| π― | Decision Making | "Should I take this job? Sign this contract? Confront this person?" β Athena cross-references your risk profile, financial runway, career decision history, and the regret patterns from your last 3 similar decisions to produce a recommendation no generic LLM could give. A business coach charges $500+/hr. Athena does it in under an hour. β case study β |
| πΌ | Work & Projects | A subset of Life Management. Juggle 5+ projects without dropping context. /project gives you a visual switchboard β phase-gated progress, urgency/EV ranking, and instant context-switching. Internal projects (health, career) and external projects (clients, revenue) tracked separately with cross-project dependency awareness. β workflow β |
| βοΈ | Writing & Voice | After 30 sessions, the AI stops sounding like ChatGPT and starts sounding like you. Learns your style from your own writing samples. |
| π¬ | Research & Synthesis | Compile 200 sources into one framework β still searchable and citable 6 months later. |
| π | Strategic Planning | Long-term planning across dozens of sessions. Budget modeling, scenario analysis, with full context of your past decisions. |
π Deep Dive: How Athena Solves β The Three Core Use Cases β covers the vulnerability prerequisite, pre-work convergence, domain reclassification, the EEV decision framework, privacy architecture, and honest limitations.
The asymmetry. A licensed therapist costs $200+/hr. A business coach costs $500+/hr. A negotiation consultant costs $1,000+/hr. Athena gives you structured, context-aware guidance across all of these domains β 24/7, for the cost of your existing AI subscription. It doesn't replace professionals for clinical emergencies, but for the 90% of life decisions and psychological patterns that don't require a medical license, it closes the gap between having access to wisdom and not being able to afford it.
β οΈ Important: Athena is an experimental AI tool, not a licensed professional service. It cannot diagnose, treat, or manage any medical or psychiatric condition. See SAFETY.md for crisis contacts and responsible use guidelines.
Not just for coding. Athena is used for personal knowledge management, health tracking, creative writing, business strategy, and daily life β by people who've never written a line of code.
Athena is free. Forever. MIT licensed. You only pay for the AI subscription you're probably already paying for.
| Plan | Cost | Who It's For |
|---|---|---|
| Google Antigravity (free tier) | $0 | Try Athena first β included with any Google account |
| Claude Pro / Google AI Pro | ~$20/mo | Daily users β the sweet spot for most people |
| Claude Max / Google AI Ultra | $200+/mo | Power users managing multiple domains (8+ hrs/day) |
Try before you buy. Athena works with Google Antigravity's free tier β clone the repo, type
/start, and see if it clicks. No credit card, no trial period, no catch. Upgrade only when you hit the free tier's daily limits.
Why $200/mo sounds expensive β until you do the math. A single employee costs $5,000+/mo in salary, benefits, and management overhead. An AI agent on a max-tier subscription costs $200/mo, works 24/7, doesn't call in sick, and scales to any domain you throw at it. For best results, subscribe to the max plan on any one platform (Claude Max, Google AI Ultra, etc.) β the difference between $20/mo and $200/mo is the difference between a tool you use occasionally and a tool that runs your life. Heavy users routinely consume $2Kβ$3K+ in equivalent API costs per month β the flat subscription turns variable cost into fixed cost.
For peak performance, use
/ultrastartevery session. On a flat-rate plan, the marginal cost of deeper thinking is $0 β so the cost of under-thinking always exceeds the cost of over-thinking./ultrastartloads ~20K tokens of structured context (identity, canonical memory, active state, semantic bridge) every session. On API pricing, this costs $2β5/session. On a $200/mo flat plan, it costs nothing. The pricing model of the underlying compute layer directly determines Athena's performance ceiling.
Boot cost is 2Kβ20K tokens (depending on mode) β constant whether it's session 1 or session 10,000. Details β
Note
Athena works with any model, but governance protocols and multi-step reasoning perform best with frontier models (e.g. Claude Opus 4.6, Gemini 3.1 Pro, GPT-5.4 (High)). Start with the free tier to test compatibility with your preferred model.
Tip
Save money getting started (as of 10 Mar 2026). Google offers a 1-month free trial on AI Pro for new subscribers β enough to fully evaluate Athena with frontier-tier Antigravity limits at $0. Alternatively, if someone you know is on a Google AI Pro or Ultra plan, they can add you as a family member β each member gets their own independent Antigravity quota at no extra cost. For Ultra subscribers, this means splitting $249/mo across family members β bringing the per-person cost of frontier-tier AI below $100/mo.
| π Getting Started | ποΈ Architecture | π Security |
| π― Your First Session | π Semantic Search | π Benchmarks |
| π‘ Tips | π MCP Server | β FAQ |
| π Updating Athena | π₯ Importing Data | β¨οΈ CLI Reference |
| π All Workflows | π Spec Sheet | π Glossary |
| π§ Manifesto | π Changelog | π Multi-Model Strategy |
| β Best Practices | π€ Your First Agent | π§© What Is an AI Agent? |
| π― Use Cases Deep Dive | π Case Studies | π‘οΈ Safety |
| π The Compounding Effect |
| Layer | Technology |
|---|---|
| IDE | Antigravity |
| Reasoning Engine | Gemini 3.1 Pro (High) / Claude Opus 4.6 (Thinking) / GPT-5.4 (High) |
| SDK | athena Python package (v9.7.0) |
| Search | Hybrid RAG β FlashRank reranking + RRF fusion |
| Embeddings | gemini-embedding-001 (768-dim) |
| Memory | Supabase + pgvector / local ChromaDB |
| Routing | Risk-Proportional Triple-Lock β SNIPER / STANDARD / ULTRA |
π Repository Structure
Athena-Public/
βββ athena.yaml # Agent manifest β model, tools, hooks, governance
βββ src/athena/ # SDK package (pip install -e .)
β βββ core/ # Config, governance, permissions, security
β βββ tools/ # Search, agentic search, reranker, heartbeat
β βββ memory/ # Vector DB, delta sync, schema
β βββ boot/ # Orchestrator, loaders, shutdown
β βββ cli/ # init, save, doctor commands
β βββ mcp_server.py # MCP Tool Server (9 tools, 2 resources)
βββ tools/ # Declarative tool definitions (YAML)
βββ scripts/ # Operational scripts (boot, shutdown, launch)
βββ examples/
β βββ protocols/ # 150+ starter frameworks (15 categories)
β βββ scripts/ # 540+ reference scripts
β βββ templates/ # Starter templates (framework, memory bank)
βββ docs/ # Architecture, benchmarks, security, guides
βββ pyproject.toml # Modern packaging
π Recent Changelog
- v9.7.0 (Apr 10 2026): Biological Analogy v2 (6-tier β 7-tier: Atom/Molecule/Cell/Tissue/Organ/System/Organism), GTO Metrics Sync (protocols 397β408, skills 24β28, case studies 410β440, sessions 1,200β1,500+, workflows 53β66+), date alignment across 8 files
- v9.6.6 (Apr 05 2026): GTO Metrics Sync & Deep Audit β filesystem-verified category-level counts, workflow count 63β66, protocol index session count 1,100β1,500+, version sync across 7 files
- v9.6.5 (Mar 31 2026): Claude Code Architectural Integration β context-compactor v2.0 (9-section summary +
<analysis>scratchpad), coordinator synthesis discipline (anti-delegation rule for agent swarms), Protocol 530 (conditional skill activation β ~40-60% token savings),/endvalidated patterns capture ([V]markers), memory drift trust-gate - v9.6.4 (Mar 31 2026): Token Economy Mode β
/minmaxworkflow (maximize quality/token ratio), selective boot (~80% token reduction), per-turn SNIPER discipline, dense output protocol, micro close default - v9.6.3 (Mar 28 2026): Metrics Sync & Deep Audit β filesystem-verified counts (149 protocols, 1,500+ sessions, 430+ case studies), version sync across 6 files, v9.6.2 CHANGELOG backfill, ABOUT_ME metrics refresh
- v9.6.2 (Mar 26 2026): ultrastart + ultraend GTO Upgrade β mandatory cross-domain sweep (PROJECTS.md + activeContext.md), decision outcome tracking, insight compounding chain, explicit propagation directives in
/ultraend - v9.6.1 (Mar 26 2026): The Ousen Protocol β
/battleplanworkflow (7-phase pre-execution battle planning), Double-Envelope Audit Architecture (Red-Team #1 on plan, Red-Team #2 on output), scout/general sequencing, examiner anticipation matrix - v9.6.0 (Mar 25 2026): Outcome Economy β new concept page #8 (labor economics of human augmentation), 7 new tags, cross-references updated across 6 files
- v9.5.7 (Mar 21 2026): Data Compounding Thesis β new wiki page (The Compounding Effect), README data quality thesis, wiki updates (Home, Philosophy Β§7, Use Cases deltas, FAQ Γ 2, Sidebar)
- v9.5.6 (Mar 19 2026): Operator Optimization β CS-006 (The Replacement Trap), Phase 2 thesis (USER_DRIVEN_RSI.md), Decision Sovereignty pre-flight checklist (BEST_PRACTICES.md Β§10)
- v9.5.5 (Mar 16 2026): Abundance Mindset Alignment β P529 Survival HUD removed (redundant with existing safety stack),
/startworkflow fix, protocol count 128β127 - v9.5.4 (Mar 14 2026): Architecture Integrity Audit β Protocol index rewrite (109β128 active, 13β15 categories), P138/P526 cluster wiring, version sync
- v9.5.3 (Mar 14 2026): Independent Cross-Model Audit β Protocol 526 (Business Viability Assessment β 3-Layer Stack), Protocol 138 (Third Choice Generation / Kobayashi Maru), Cold Start Rule (
BEST_PRACTICES.mdΒ§9) - v9.5.2 (Mar 13 2026): Ollama Integration & Docs Sync β Ollama local embedding provider (
vectors.pyprovider pattern), Symbiotic RSI codification, Dual Pressure Model, VECTORRAG.md full refresh (model name, counts, Ollama docs), metrics sync - v9.5.1 (Mar 11 2026): Conviction-Decisiveness Split β Protocol 524 (decouple epistemic conviction from operational decisiveness in semi-stochastic domains), README
/ultrastartcompute recommendation for MAX subscribers - v9.5.0 (Mar 11 2026): Adaptive Graph of Thoughts β Protocol 75 v5.0 (AGoT-enhanced parallel reasoning),
agot_orchestrator.pywith 3-tier complexity routing (lite/full/tracks), adaptive convergence gate, recursive sub-graph spawning,/ultrastartAGoT wiring - v9.4.9 (Mar 10 2026): Deep Session Close β
/ultraendworkflow (System-2 deep close counterpart to/ultrastart), cross-session pattern scan, CANONICAL deep reconciliation, reflexion archive, strategic portfolio review, next session seeding - v9.4.8 (Mar 10 2026): Boot/Shutdown Architecture Redesign β
/ultrastartworkflow (20K-token System-2 deep boot),/endGTO v3 rewrite (dual-write architecture fix),quicksave.pyTriple-Lock ANDβOR (removes Robustness Theater) - v9.4.7 (Mar 09 2026): Safety Documentation & Governance Hardening β
SAFETY.md(clinical disclaimer, crisis contacts), README safety disclaimers, 24 MC/DC governance tests (private), compaction pipeline v3.0 (private) - v9.4.6 (Mar 09 2026): Project Switchboard β
/projectworkflow (view, add, switch, close, triage), PROJECTS.md template, Internal/External zones, cross-project dependencies,/start+/endintegration - v9.4.5 (Mar 09 2026): Two-Mode Session Architecture β Lightweight (skip
/start) vs Full Boot. Framework Tax concept. Orchestrator-Executor Pipeline. Crisis Architecture (P509, P519, P520, P521) - v9.4.4 (Mar 07 2026): GTO Routing Diagram β expanded from 2/8 β 8/8 system cluster chains, priority tier color-coding (Critical/High/Standard/Support), reordered Q4-Q6 to match priority waterfall
- v9.4.3 (Mar 07 2026): Maintenance β AGENTS.md version sync, file count corrections (138 protocols, 540+ scripts), date alignment
- v9.4.2 (Mar 05 2026): Cognitive Architecture v2.1 β Homeostatic Pressure (P517), Reflexion Journaling (P515), Memory Paging (P516), LIDA Broadcast routing, deterministic exit verification, Ebbinghaus decay, context clearing
- v9.4.1 (Mar 04 2026): Cognitive Systems v2 β Ideation mode (DIVERGENT/CONVERGENT), LearningβLife Decision handoff, Cluster #14 safety sequence, P503 cluster sync (15), Ξ-based stealth routing, 3 new protocols (P511 Business Viability, P512 Pre-Planning, P513 Context Isolation)
- v9.4.0 (Mar 04 2026): Biological Stack Architecture β 5 new protocols (P504-P508), 8 Cognitive Systems layer, Intent Classifier,
ensure_env.shsupports system Python - v9.3.1 (Mar 03 2026): Cross-model Audit Fixes β file count sync, Windows section relocation, GitHub Release catch-up (v9.2.7βv9.3.0)
- v9.3.0 (Mar 02 2026): Onboarding Friction Audit β dependency restructuring (torchβoptional), venv instructions, PEP 668 fix, stale path cleanup, two-tier install
- v9.2.9 (Mar 02 2026): Ultrathink v4.1 HITL Bypass β manual Gemini sandbox option, micro-pruned 10% dead skills (100% cluster coverage), broken reference audit
- v9.2.8 (Feb 27 2026): Skill Template Expansion β 5 starter skill templates across 4 categories for new AG users
- v9.2.7 (Feb 26 2026): Risk-proportional Triple-Lock, Tier 0 context summaries, 3 new academic citations
- v9.2.6 (Feb 25 2026): Kilo Code + Roo Code IDE integration,
COMPATIBLE_IDES.md, issue #19 closed - v9.2.5 (Feb 24 2026): Life Integration Protocol Stack β Protocols 381-383, Emotional Audit,
/reviewworkflow - v9.2.3 (Feb 21 2026): Multi-agent safety hardening, CLAUDE.md symlinks, issue deflection
- v9.2.2 (Feb 21 2026): S-tier README refactor, docs restructure
- v9.2.1 (Feb 20 2026): Deep Audit & PnC Sanitization β 17 patterns sanitized across 13 files
- v9.2.0 (Feb 17 2026): Sovereignty Convergence β CVE patch, agentic search, governance upgrade
- v9.1.0 (Feb 17 2026): Deep Audit & Sync β Fixed 15 issues (dead links, version drift)
- v9.0.0 (Feb 16 2026): First-Principles Workspace Refactor β root dir cleaned, build artifacts purged
π Full Changelog β
MIT License Β· Contributing Β· Safety Β· Security Β· Code of Conduct
Clone it. Boot it. Make it yours.

