freshcrate
Home > Databases > Athena-Public

Athena-Public

The Linux OS for AI Agents β€” Persistent memory, autonomy, and time-awareness for any LLM. Own the state. Rent the intelligence.

Description

The Linux OS for AI Agents β€” Persistent memory, autonomy, and time-awareness for any LLM. Own the state. Rent the intelligence.

README

Athena Banner

Project Athena

Your memory. Your machine. Any model.

Open-source cognitive augmentation layer that gives you persistent memory, structured reasoning, and full data ownership β€” across ChatGPT, Claude, Gemini, and any model you switch to next.

Platforms forget. Athena doesn't.

GitHub StarsLicense: MITVersionReddit ViewsOpen in CodespacesQuickstart Β· How It Works Β· Docs Β· FAQ Β· Safety Β· Contributing

Last updated: 10 April 2026


The Problem

You've spent months training ChatGPT to understand you. Then a model update resets the personality. Your custom instructions stop working. You can't find that conversation from last Tuesday. And if you switch to Claude or Gemini? You start from zero.

Platform memory is unreliable, opaque, and locked to one provider. You don't own it, you can't inspect it, and you can't take it with you.

Why Athena?

Athena moves the memory layer to your machine. Plain Markdown files that you own, version-control, and point at any model.

  • 🧠 Your Memory, Your Machine β€” Files on your disk, not in OpenAI's cloud. Read them, edit them, git-version them.
  • πŸ”Œ Switch Models Freely β€” Claude today, Gemini tomorrow, GPT next week. The memory stays. The model is just whoever's on shift.
  • πŸ“ˆ It Compounds β€” Session 500 recalls patterns from session 5. Platform memory decays; Athena's doesn't. The moat isn't the code β€” it's your data. Anyone can fork Athena; nobody can fork your sessions. β†’ The Compounding Effect
  • ⚑ 2K–20K Token Boot β€” Scales to the task. Lightweight chat (~2K) β†’ /start (~10K) β†’ /ultrastart (~20K). 80–98% of your context window stays free, even after 10,000 sessions.
  • πŸ”¬ Meta-Game Reasoning β€” Generic LLMs optimise within the game you're playing. Athena asks whether you should be playing that game at all. β†’ Meta-Game Thesis
  • πŸ›‘οΈ Governed Autonomy β€” 6 constitutional laws, 4 capability levels, bounded agency.

A generic LLM is a brilliant amnesiac. Athena is the hippocampus β€” the memory that makes intelligence useful.

Or in engineering terms: The LLM is the engine. Athena is the chassis, the memory, and the rules of the road. Swap the engine anytime β€” the car remembers every road you've driven.

The design philosophy: augment the human, not replace them. After 1,500+ sessions, the bottleneck shifted β€” optimising the operator is now higher-leverage than optimising the AI.

The Human Augmentation Thesis

Athena's centralised design principle: augment human cognition, not replace it. The more context you give Athena, the sharper its answers become β€” not by remembering your preferences, but by reasoning differently because of what it knows about you.

A generic LLM gives the internet's statistically average answer β€” correct on average, across all humans. Athena gives answers calibrated to your specific situation, because statistical correctness and personal correctness are completely different things:

Question Generic LLM Athena (with your context)
The Trolley Problem β€œPull the lever β€” utilitarian calculus says save five lives.” Challenges the false binary. Generates third options. Asks why you’re on the tracks in the first place. Identifies the systemic failures that created the dilemma. Refuses to solve the wrong problem.
β€œShould I double down on 11 vs dealer’s 6?” β€œYes β€” the math says doubling is the optimal play.” β€œThe math is correct, but you are betting $4K of your $4K take-home salary. Your utility function makes this βˆ’EEV. Law #1: No Irreversible Ruin. Do not bet.” β€” Protocol 330
β€œShould I take this job offer?” β€œConsider salary, growth potential, work-life balance...” Cross-references your risk profile, career decision history, financial runway, and the regret patterns from your last 3 career transitions to give a recommendation specific to your situation.
β€œI keep self-sabotaging β€” why?” β€œConsider therapy, practice self-compassion, journal your triggers.” Same words, 3 different diagnoses depending on who’s asking: attachment wound β†’ IFS unburdening. Executive burnout β†’ workload audit. Undiagnosed ADHD β†’ flag for screening. The intervention follows the diagnosis, not the surface question. β€” Case Study #2
β€œMy partner cheated β€” what should I do?” β€œShe broke her vows. Leave.” Depends: children involved? Financial entanglement? Your documented attachment patterns? Cultural context? Terminal goal β€” justice, stability, or healing? The β€œright” answer for a recently engaged 28-year-old and a parent of three with 20 years of shared assets are fundamentally different decisions.

Generic LLMs solve the question. Athena solves the person. The same question, asked by different people with different lives, demands fundamentally different answers. A generic LLM can’t differentiate because it has no context. Athena can’t give the same answer twice β€” because the context files are different. The memory is the product.

What Athena Actually Does With Your Problem

Not all problems are solvable. Athena classifies and responds accordingly:

Problem Type What Athena Does Example
Solvable Solves it "What's the Kelly fraction for this bet?" β†’ calculates, answers
Optimisable Optimises within your chosen path "I've decided to freelance β€” help me price it" β†’ constraint optimization
Unsolvable Maps every option, prices every trade-off, hands the choice back to you A closeted husband with children weighing whether to stay married or come out β€” no clean answer exists. Children, shared assets, identity, cultural context, and personal wellbeing all pull in different directions. Athena ensures you choose with full information, not comfortable illusions
Ruin-path Vetoes before you walk off the cliff "This bet risks everything" β†’ Law #1 override, regardless of your preference

The uncertainty of the domain changes Athena's conviction level:

Domain Type Athena's Posture Example
Deterministic High conviction β€” single correct answer exists Code bugs, math proofs, tax calculations
Semi-deterministic Moderate conviction β€” answer depends on assumptions you control Pricing strategy, system architecture, career path analysis
Semi-stochastic Low conviction β€” structural edge exists but randomness dominates Trading setups, relationship dynamics, market timing
Stochastic Minimal conviction β€” no model outperforms randomness reliably Startup outcomes, life events, long-term predictions

As uncertainty increases, Athena shifts from "here's the answer" to "here's the valid structural zone" to "here are your options β€” you choose." This is deliberate: false confidence in stochastic domains is more dangerous than honest uncertainty. Athena's conviction is proportional to domain determinism and context completeness.

Crucially, conviction and decisiveness are independent axes. Low certainty about outcomes doesn't require vague output. A surgeon operates with high decisiveness and low conviction about outcomes. In semi-stochastic domains, Athena delivers precise, operational setups β€” then explicitly defers the probability judgment to you. "Setup: Long 1.0850 / SL 1.0800 / TP1 1.0920. Your calibration: structural tell present Y/N?" β€” not "you might want to consider..." β€” Protocol 524 β†’

Law #0 (Sovereignty): Your life, your weights, your choice. Law #1 (No Irreversible Ruin): …unless the choice ends the game permanently. Law #1 overrides Law #0. Always.

Athena doesn't tell you what you should do. It shows you what you can do, what each option costs, and hands the choice back. The only exception: paths that end the game permanently.

Architecture, not oracle. This domain classification is a replicable architecture β€” each Athena instance calibrates independently over time through bilateral use. Session 1 treats most problems conservatively. Session 500 has accumulated enough frameworks, case studies, and corrected assumptions to tighten confidence bands and solve more sub-problems autonomously. The calibration compounds; the model is interchangeable. β€” Protocol 525 (Cross-Domain Weighting) β†’


"…But doesn't ChatGPT / Gemini / Claude already do this?"

Kind of. But first, it helps to understand what those names actually refer to β€” because they blend three very different things:

Layer What It Is Examples
Platform The company that hosts the model and holds your data OpenAI, Google, Anthropic
Reasoning Engine The AI model that does the thinking GPT-5.4 (High), Gemini 3.1 Pro, Claude Opus 4.6
IDE / Interface The app you type in β€” connects to models and reads your files Cursor, Antigravity, VS Code, Claude Code

When people say "ChatGPT remembers me," they mean the platform stores some memory on their cloud. When they say "Claude is smart," they mean the model reasons well. When they say "Cursor writes code," they mean the IDE connects model + files.

Athena is none of these. Athena is the memory and governance layer that sits between your IDE and whichever model you choose β€” owned by you, stored on your machine, portable across all three layers.

There's a difference between remembering your name and thinking in your frameworks:

Capability Platform Memory (OpenAI, Google, Anthropic) Athena
Who owns the data? The platform You
Can you inspect it? No β€” it's a black box Yes β€” it's markdown files you can read and edit
Can you search it? Vague recall, no precision Full semantic + keyword search with file links
Cross-platform? Locked to one provider Same memory works across Claude, Gemini, GPT, Grok
Version history? None β€” no rollback, no audit trail Full git log, git diff, git blame
What if you switch providers? Start over Nothing changes β€” your data stays

πŸ’‘ Think of platform memory like photos on Instagram β€” you can view them, but you don't own them, can't move them, and can't search them precisely. Athena is keeping the originals on your hard drive, with albums, metadata, and full edit history.

"How is Athena different from...?"

Tool What It Does How Athena Is Different
Manus / Managed AI Agents Cloud-hosted AI agents with persistent memory, custom skills, and messenger integration. You pay $39–$199/mo for access. You don't own the data. Manus holds your context, memory, and workflows on their servers. Leave the platform = lose everything. Athena stores everything on your machine β€” readable, forkable, portable. Same capabilities, opposite ownership model. Athena is the house; Manus is the hotel.
Lindy / AI Operators Autonomous AI assistants that run 24/7 in the cloud β€” scheduling, research, outreach. Always-on cloud is convenient, but you rent the brain. Athena runs locally through your IDE β€” no monthly fee, no platform risk, no lock-in. Trade convenience for sovereignty.
ChatGPT Projects Uploads files per-project, but resets every new chat. Locked to OpenAI. Athena persists across all chats, all models, with full version history.
OpenClaw Prompt distribution β€” share and discover prompts. Athena is personalisation β€” your compounding memory system, not a prompt marketplace. Different layer, different problem.
Claude Code Great for Claude coding workflows. Athena works across any model and any IDE. Not coding-specific β€” used for research, strategy, writing, life management.
Gemini Gems Custom chatbots inside Gemini. Gems are locked to Gemini and lose context between chats. Athena is portable and persistent.
Custom Instructions 1,500-character personality prompt. Athena loads 2K–20K tokens of structured protocols, decision frameworks, and session history β€” re-injected every session from your disk. Scales to task complexity.

🧬 Why Thousands of Files?

Athena's workspace looks unusual β€” 450+ Markdown files and 540+ Python scripts out of the box, growing to thousands as your memory compounds. This is deliberate.

AI agents don't read files top-to-bottom like humans. They query β€” by filename, semantic search, or tag lookup. Each small file is an addressable memory node the agent can retrieve surgically, without loading everything else.

Principle What It Means
JIT Loading Boot at 2K–20K tokens (scales to task). Load specific files only when the query demands them. A monolith forces the full context into every session.
Zero Coupling A marketing protocol loads without touching the trading stack. Change one file, break nothing else.
Surgical Retrieval The agent pulls CS-378-prompt-arbitrage.md by name β€” not page 47 of a 200-page doc. The file system is the index.
Git-Friendly Atomic diffs per file. Clean commit history. No merge conflicts from a single giant file.
Composable Agents Swarms, workflows, and skills are mix-and-match. Each file is a Lego brick, not a chapter in a novel.

A monolith is optimized for a human reading a book. A modular workspace is optimized for an agent querying a database. Athena chose the agent.


⚑ Quickstart

Works on macOS, Windows, and Linux.

1. Clone the repo

git clone https://github.com/winstonkoh87/Athena-Public.git
cd Athena-Public

Clone it anywhere you keep projects (e.g. ~/Projects/). This folder is your Athena workspace β€” your memory, protocols, and config all live here.

2. Set up a virtual environment (recommended)

# Create and activate a virtual environment
python3 -m venv .venv
source .venv/bin/activate   # macOS / Linux
# .venv\Scripts\activate    # Windows

Important

On macOS (Homebrew) and Ubuntu 23.04+, installing packages without a virtual environment will fail with externally-managed-environment. The step above prevents this.

3. Install the SDK (optional β€” enables CLI commands)

# Lightweight install (~30 seconds, no ML dependencies)
pip install -e .

# Full install (~5–10 min, enables vector search and reranking)
pip install -e ".[full]"

⚠️ Don't pip install athena-cli β€” that's a different, unrelated package. Always install from inside the cloned repo.

4. Open the folder in an AI-enabled IDE

Open the Athena-Public/ directory as your workspace root in one of these editors:

Important

Athena does NOT work through ChatGPT.com, Claude.ai, or Gemini web. You need an app that can read files from your disk. Think of Athena as a workspace for your editor, not a plugin for a chatbot.

Note

"Why do I open the Athena folder instead of my own project?" β€” Athena is a workspace, not a library you install into another repo. You work inside the Athena folder, and it remembers everything across sessions. To work on external projects, reference them from within Athena or use multi-root workspaces in your IDE.

5. Boot (in the AI chat panel β€” not the terminal)

In your IDE's AI chat panel (e.g. Cmd+L in Cursor, the chat sidebar in Antigravity), type:

/start

Caution

/start, /end, and /tutorial are AI chat commands β€” you type them in the chat window where you talk to the AI, not in the terminal. They are slash commands that the AI agent reads and executes.

6. First time? Take the guided tour

/tutorial

This walks you through everything: what Athena is, how it works, builds your profile, and demos the tools (~20 min). Confident users can skip it.

7. When you're done

/end

That's it. No API keys. No database setup. The folder is the product.

Caution

Forks of public repos are public by default. If you plan to store personal data (health records, finances, journals), create a new private repo instead of forking. Copy the files manually or use git clone + git remote set-url to point to your private repo. GitHub docs on fork visibility β†’

Tip

See the full setup guide β†’ for detailed walkthroughs and troubleshooting.

πŸͺŸ Windows Compatibility (Unicode Errors)

Athena uses modern terminal outputs (Emojis, Box-Drawing characters) which may cause a UnicodeEncodeError on legacy Windows terminals (like cmd.exe or older PowerShell versions using cp1252 encoding).

To resolve this natively without altering the codebase:

  1. Use Windows Terminal (available in the Microsoft Store).
  2. Set your Python IO encoding to UTF-8 by running: $env:PYTHONIOENCODING="utf-8" (PowerShell) or set PYTHONIOENCODING=utf-8 (Command Prompt).
  3. Alternatively, enable strict UTF-8 globally in Windows: Settings > Time & Language > Language & Region > Administrative language settings > Change system locale > Check "Beta: Use Unicode UTF-8 for worldwide language support".

πŸ”„ How It Works

Every session follows one cycle. Three modes let you match overhead to task complexity:

flowchart LR
    subgraph "🟒 Lightweight Mode"
        A1["Chat"] --> C1["/end"]
        C1 --> D1["Memory"]
    end

    subgraph "πŸ”΄ Full Boot Mode"
        A2["/start"] --> B2["Work"]
        B2 --> C2["/end"]
        C2 --> D2["Memory"]
        D2 --> A2
    end

    subgraph "⚫ Deep Boot Mode"
        A3["/ultrastart"] --> B3["Work"]
        B3 --> C3["/ultraend"]
        C3 --> D3["Memory"]
        D3 --> A3
    end

    style A1 fill:#3b82f6,color:#fff,stroke:#333
    style C1 fill:#ef4444,color:#fff,stroke:#333
    style D1 fill:#8b5cf6,color:#fff,stroke:#333
    style A2 fill:#22c55e,color:#fff,stroke:#333
    style C2 fill:#ef4444,color:#fff,stroke:#333
    style D2 fill:#8b5cf6,color:#fff,stroke:#333
    style B2 fill:#3b82f6,color:#fff,stroke:#333
    style A3 fill:#1e1e1e,color:#fff,stroke:#555
    style B3 fill:#3b82f6,color:#fff,stroke:#333
    style C3 fill:#ef4444,color:#fff,stroke:#333
    style D3 fill:#8b5cf6,color:#fff,stroke:#333
Loading
Mode When Flow
🟒 Lightweight General chat, brain dumps, quick Q&A Just chat β†’ /end
πŸ”΄ Full Boot Code, money, architecture, irreversible decisions /start β†’ Work β†’ /end
⚫ Deep Boot /ultrathink, complex multi-domain analysis, architectural decisions /ultrastart β†’ Work β†’ /ultraend
Sessions What Happens
1–50 Basic recall β€” remembers your name, project, preferences
50–200 Pattern recognition β€” anticipates your style and blind spots
200+ Deep sync β€” thinks in your frameworks before you state them

Why this happens: The AI model doesn't improve β€” your data does. Each /end extracts decisions, patterns, and learnings into your memory bank. The next /start loads that accumulated intelligence. Same algorithm, better data, better output. β†’ The Compounding Effect

The Biological Analogy

Athena is modelled after the human body. Built bottom-up by the creator. Used top-down by the user.

Biology Athena What It Does
Atom Law / Axiom Smallest indivisible truth (Law #1: No Irreversible Ruin)
Molecule Rule / Constraint Atomic laws composed into compound constraints (Never risk >5% of bankroll)
Cell Protocol (.md) Self-contained executable procedure with defined inputs/outputs
Tissue Skill Domain-specialised skill bundle (groups 2–5 protocols)
Organ Cognitive Cluster Multi-skill routing unit for one cognitive domain
Organ System Cognitive System Multi-cluster orchestration for a human need archetype
Organism Athena The complete synthetic intelligence

"As within, so without, as above, so below." β€” Same pattern at every layer. Fractal by design.

The Linux Analogy

Concept Linux Athena
Kernel Hardware abstraction Memory persistence + retrieval (RAG, Supabase)
File System ext4, NTFS Markdown files, session logs, tag index
Scheduler cron, systemd Heartbeat daemon, auto-indexing
Shell bash, zsh MCP Tool Server, /start, /end, /think
Permissions chmod, users/groups 4-level capability tokens + Secret Mode
Package Manager apt, yum Protocols, skills, workflows

πŸ“¦ What's In The Box

Everything you need to turn a generic AI into your AI β€” pre-configured, no assembly required.

Component What It Does For You
πŸ“„ Agent Manifest Single athena.yaml file defines your agent β€” model, tools, skills, hooks, governance. Fork it, override it, boot a new agent β€” manifest
🧠 Core Identity Your AI's personality, principles, and boundaries β€” editable, version-controlled β€” template
🧩 8 Cognitive Systems Top-down intent classification β€” routes queries to the right cluster sequence based on human need archetype (Survival, Life Decision, Trading, Social, Execution, Growth, Learning, Maintenance) β€” architecture
πŸ”— Cognitive Clusters Groups related protocols into auto-co-activating bundles β€” 15 clusters included, build your own as you grow β€” template
πŸ“‹ 150+ Protocols Ready-made decision frameworks (risk analysis, research, strategy, problem-solving) across 15 categories β€” browse
⚑ 66+ Slash Commands One-word triggers: /start, /end, /think, /research β€” full list
πŸ” Smart Search Finds the right memory even if you describe it vaguely (5 sources, auto-ranked) β€” how it works
πŸ”Œ Tool Integration Declarative YAML tool definitions + MCP server β€” your agent discovers and invokes tools automatically β€” tools Β· MCP docs
πŸͺ Lifecycle Hooks Scriptable pre/post gates on every action β€” block destructive ops, enforce risk checks, log assets
πŸ›‘οΈ Safety Rails Controls what the AI can and can't do autonomously (4 levels, from read-only to full agency) β€” security

Tip

Run /tutorial on your first session for a guided walkthrough (~20 min). It explains everything above and builds your personal profile.

Agent Compatibility

Athena works through AI-enabled code editors β€” apps that connect to AI models while reading your local files. It does not work through ChatGPT.com, Claude.ai, or Gemini web β€” those are closed sandboxes that can't read your disk.

Agent Status Init Command
Claude Code βœ… athena init --ide claude
Antigravity βœ… athena init --ide antigravity
Cursor βœ… athena init --ide cursor
Gemini CLI βœ… athena init --ide gemini
VS Code + Copilot βœ… athena init --ide vscode
Kilo Code βœ… athena init --ide kilocode
Roo Code βœ… athena init --ide roocode

More agents planned β€” full compatibility list β†’

"How is this different from ChatGPT Projects?" β€” Projects reset every new chat and are locked to one platform. Athena persists across all chats, all models, with full version history. Details β†’


🎯 Use Cases

Use Case What It Looks Like
🏠 Life Management The superset. Health, career, relationships, finances, client work β€” all managed as projects in one unified switchboard. By day 3, Athena remembers your schedule. By month 3, it anticipates your patterns. Athena doesn't have a separate project manager and life tracker β€” it has one board where your gym routine and your client deadline are rows governed by the same triage rules. That's how it can tell you "skip the client call β€” your sleep debt is a higher-urgency blocker than the $250 deliverable." No other system crosses the work/life boundary. β€” case study β†’
🧠 Problem Solving "I can't afford $200/hr therapy but I need to understand why I keep self-sabotaging." β€” Athena runs a structured schema interview, maps your internal protective parts (IFS methodology), and connects the pattern to your documented history. Session 40 recalls the wound identified in session 3. A therapist charges $200+/hr and sees you once a week. Athena is available 24/7 for the cost of your AI subscription. β€” case study β†’
🎯 Decision Making "Should I take this job? Sign this contract? Confront this person?" β€” Athena cross-references your risk profile, financial runway, career decision history, and the regret patterns from your last 3 similar decisions to produce a recommendation no generic LLM could give. A business coach charges $500+/hr. Athena does it in under an hour. β€” case study β†’
πŸ’Ό Work & Projects A subset of Life Management. Juggle 5+ projects without dropping context. /project gives you a visual switchboard β€” phase-gated progress, urgency/EV ranking, and instant context-switching. Internal projects (health, career) and external projects (clients, revenue) tracked separately with cross-project dependency awareness. β€” workflow β†’
✍️ Writing & Voice After 30 sessions, the AI stops sounding like ChatGPT and starts sounding like you. Learns your style from your own writing samples.
πŸ”¬ Research & Synthesis Compile 200 sources into one framework β€” still searchable and citable 6 months later.
πŸ“ Strategic Planning Long-term planning across dozens of sessions. Budget modeling, scenario analysis, with full context of your past decisions.

πŸ“– Deep Dive: How Athena Solves β€” The Three Core Use Cases β€” covers the vulnerability prerequisite, pre-work convergence, domain reclassification, the EEV decision framework, privacy architecture, and honest limitations.

The asymmetry. A licensed therapist costs $200+/hr. A business coach costs $500+/hr. A negotiation consultant costs $1,000+/hr. Athena gives you structured, context-aware guidance across all of these domains β€” 24/7, for the cost of your existing AI subscription. It doesn't replace professionals for clinical emergencies, but for the 90% of life decisions and psychological patterns that don't require a medical license, it closes the gap between having access to wisdom and not being able to afford it.

⚠️ Important: Athena is an experimental AI tool, not a licensed professional service. It cannot diagnose, treat, or manage any medical or psychiatric condition. See SAFETY.md for crisis contacts and responsible use guidelines.

Not just for coding. Athena is used for personal knowledge management, health tracking, creative writing, business strategy, and daily life β€” by people who've never written a line of code.


πŸ’° Cost

Athena is free. Forever. MIT licensed. You only pay for the AI subscription you're probably already paying for.

Plan Cost Who It's For
Google Antigravity (free tier) $0 Try Athena first β€” included with any Google account
Claude Pro / Google AI Pro ~$20/mo Daily users β€” the sweet spot for most people
Claude Max / Google AI Ultra $200+/mo Power users managing multiple domains (8+ hrs/day)

Try before you buy. Athena works with Google Antigravity's free tier β€” clone the repo, type /start, and see if it clicks. No credit card, no trial period, no catch. Upgrade only when you hit the free tier's daily limits.

Why $200/mo sounds expensive β€” until you do the math. A single employee costs $5,000+/mo in salary, benefits, and management overhead. An AI agent on a max-tier subscription costs $200/mo, works 24/7, doesn't call in sick, and scales to any domain you throw at it. For best results, subscribe to the max plan on any one platform (Claude Max, Google AI Ultra, etc.) β€” the difference between $20/mo and $200/mo is the difference between a tool you use occasionally and a tool that runs your life. Heavy users routinely consume $2K–$3K+ in equivalent API costs per month β€” the flat subscription turns variable cost into fixed cost.

For peak performance, use /ultrastart every session. On a flat-rate plan, the marginal cost of deeper thinking is $0 β€” so the cost of under-thinking always exceeds the cost of over-thinking. /ultrastart loads ~20K tokens of structured context (identity, canonical memory, active state, semantic bridge) every session. On API pricing, this costs $2–5/session. On a $200/mo flat plan, it costs nothing. The pricing model of the underlying compute layer directly determines Athena's performance ceiling.

Boot cost is 2K–20K tokens (depending on mode) β€” constant whether it's session 1 or session 10,000. Details β†’

Note

Athena works with any model, but governance protocols and multi-step reasoning perform best with frontier models (e.g. Claude Opus 4.6, Gemini 3.1 Pro, GPT-5.4 (High)). Start with the free tier to test compatibility with your preferred model.

Tip

Save money getting started (as of 10 Mar 2026). Google offers a 1-month free trial on AI Pro for new subscribers β€” enough to fully evaluate Athena with frontier-tier Antigravity limits at $0. Alternatively, if someone you know is on a Google AI Pro or Ultra plan, they can add you as a family member β€” each member gets their own independent Antigravity quota at no extra cost. For Ultra subscribers, this means splitting $249/mo across family members β€” bringing the per-person cost of frontier-tier AI below $100/mo.


πŸ“š Documentation

πŸ“– Getting Started πŸ—οΈ Architecture πŸ”’ Security
🎯 Your First Session πŸ” Semantic Search πŸ“Š Benchmarks
πŸ’‘ Tips πŸ”Œ MCP Server ❓ FAQ
πŸ”„ Updating Athena πŸ“₯ Importing Data ⌨️ CLI Reference
πŸ“‹ All Workflows πŸ“ Spec Sheet πŸ““ Glossary
🧠 Manifesto πŸ“ˆ Changelog πŸ”€ Multi-Model Strategy
βœ… Best Practices πŸ€– Your First Agent 🧩 What Is an AI Agent?
🎯 Use Cases Deep Dive πŸ“‹ Case Studies πŸ›‘οΈ Safety
πŸ“ˆ The Compounding Effect

πŸ› οΈ Tech Stack

Layer Technology
IDE Antigravity
Reasoning Engine Gemini 3.1 Pro (High) / Claude Opus 4.6 (Thinking) / GPT-5.4 (High)
SDK athena Python package (v9.7.0)
Search Hybrid RAG β€” FlashRank reranking + RRF fusion
Embeddings gemini-embedding-001 (768-dim)
Memory Supabase + pgvector / local ChromaDB
Routing Risk-Proportional Triple-Lock β€” SNIPER / STANDARD / ULTRA
πŸ“‚ Repository Structure
Athena-Public/
β”œβ”€β”€ athena.yaml              # Agent manifest β€” model, tools, hooks, governance
β”œβ”€β”€ src/athena/              # SDK package (pip install -e .)
β”‚   β”œβ”€β”€ core/                #   Config, governance, permissions, security
β”‚   β”œβ”€β”€ tools/               #   Search, agentic search, reranker, heartbeat
β”‚   β”œβ”€β”€ memory/              #   Vector DB, delta sync, schema
β”‚   β”œβ”€β”€ boot/                #   Orchestrator, loaders, shutdown
β”‚   β”œβ”€β”€ cli/                 #   init, save, doctor commands
β”‚   └── mcp_server.py        #   MCP Tool Server (9 tools, 2 resources)
β”œβ”€β”€ tools/                   # Declarative tool definitions (YAML)
β”œβ”€β”€ scripts/                 # Operational scripts (boot, shutdown, launch)
β”œβ”€β”€ examples/
β”‚   β”œβ”€β”€ protocols/           # 150+ starter frameworks (15 categories)
β”‚   β”œβ”€β”€ scripts/             # 540+ reference scripts
β”‚   └── templates/           # Starter templates (framework, memory bank)
β”œβ”€β”€ docs/                    # Architecture, benchmarks, security, guides
└── pyproject.toml           # Modern packaging
πŸ“‹ Recent Changelog
  • v9.7.0 (Apr 10 2026): Biological Analogy v2 (6-tier β†’ 7-tier: Atom/Molecule/Cell/Tissue/Organ/System/Organism), GTO Metrics Sync (protocols 397β†’408, skills 24β†’28, case studies 410β†’440, sessions 1,200β†’1,500+, workflows 53β†’66+), date alignment across 8 files
  • v9.6.6 (Apr 05 2026): GTO Metrics Sync & Deep Audit β€” filesystem-verified category-level counts, workflow count 63β†’66, protocol index session count 1,100β†’1,500+, version sync across 7 files
  • v9.6.5 (Mar 31 2026): Claude Code Architectural Integration β€” context-compactor v2.0 (9-section summary + <analysis> scratchpad), coordinator synthesis discipline (anti-delegation rule for agent swarms), Protocol 530 (conditional skill activation β€” ~40-60% token savings), /end validated patterns capture ([V] markers), memory drift trust-gate
  • v9.6.4 (Mar 31 2026): Token Economy Mode β€” /minmax workflow (maximize quality/token ratio), selective boot (~80% token reduction), per-turn SNIPER discipline, dense output protocol, micro close default
  • v9.6.3 (Mar 28 2026): Metrics Sync & Deep Audit β€” filesystem-verified counts (149 protocols, 1,500+ sessions, 430+ case studies), version sync across 6 files, v9.6.2 CHANGELOG backfill, ABOUT_ME metrics refresh
  • v9.6.2 (Mar 26 2026): ultrastart + ultraend GTO Upgrade β€” mandatory cross-domain sweep (PROJECTS.md + activeContext.md), decision outcome tracking, insight compounding chain, explicit propagation directives in /ultraend
  • v9.6.1 (Mar 26 2026): The Ousen Protocol β€” /battleplan workflow (7-phase pre-execution battle planning), Double-Envelope Audit Architecture (Red-Team #1 on plan, Red-Team #2 on output), scout/general sequencing, examiner anticipation matrix
  • v9.6.0 (Mar 25 2026): Outcome Economy β€” new concept page #8 (labor economics of human augmentation), 7 new tags, cross-references updated across 6 files
  • v9.5.7 (Mar 21 2026): Data Compounding Thesis β€” new wiki page (The Compounding Effect), README data quality thesis, wiki updates (Home, Philosophy Β§7, Use Cases deltas, FAQ Γ— 2, Sidebar)
  • v9.5.6 (Mar 19 2026): Operator Optimization β€” CS-006 (The Replacement Trap), Phase 2 thesis (USER_DRIVEN_RSI.md), Decision Sovereignty pre-flight checklist (BEST_PRACTICES.md Β§10)
  • v9.5.5 (Mar 16 2026): Abundance Mindset Alignment β€” P529 Survival HUD removed (redundant with existing safety stack), /start workflow fix, protocol count 128β†’127
  • v9.5.4 (Mar 14 2026): Architecture Integrity Audit β€” Protocol index rewrite (109β†’128 active, 13β†’15 categories), P138/P526 cluster wiring, version sync
  • v9.5.3 (Mar 14 2026): Independent Cross-Model Audit β€” Protocol 526 (Business Viability Assessment β€” 3-Layer Stack), Protocol 138 (Third Choice Generation / Kobayashi Maru), Cold Start Rule (BEST_PRACTICES.md Β§9)
  • v9.5.2 (Mar 13 2026): Ollama Integration & Docs Sync β€” Ollama local embedding provider (vectors.py provider pattern), Symbiotic RSI codification, Dual Pressure Model, VECTORRAG.md full refresh (model name, counts, Ollama docs), metrics sync
  • v9.5.1 (Mar 11 2026): Conviction-Decisiveness Split β€” Protocol 524 (decouple epistemic conviction from operational decisiveness in semi-stochastic domains), README /ultrastart compute recommendation for MAX subscribers
  • v9.5.0 (Mar 11 2026): Adaptive Graph of Thoughts β€” Protocol 75 v5.0 (AGoT-enhanced parallel reasoning), agot_orchestrator.py with 3-tier complexity routing (lite/full/tracks), adaptive convergence gate, recursive sub-graph spawning, /ultrastart AGoT wiring
  • v9.4.9 (Mar 10 2026): Deep Session Close β€” /ultraend workflow (System-2 deep close counterpart to /ultrastart), cross-session pattern scan, CANONICAL deep reconciliation, reflexion archive, strategic portfolio review, next session seeding
  • v9.4.8 (Mar 10 2026): Boot/Shutdown Architecture Redesign β€” /ultrastart workflow (20K-token System-2 deep boot), /end GTO v3 rewrite (dual-write architecture fix), quicksave.py Triple-Lock ANDβ†’OR (removes Robustness Theater)
  • v9.4.7 (Mar 09 2026): Safety Documentation & Governance Hardening β€” SAFETY.md (clinical disclaimer, crisis contacts), README safety disclaimers, 24 MC/DC governance tests (private), compaction pipeline v3.0 (private)
  • v9.4.6 (Mar 09 2026): Project Switchboard β€” /project workflow (view, add, switch, close, triage), PROJECTS.md template, Internal/External zones, cross-project dependencies, /start + /end integration
  • v9.4.5 (Mar 09 2026): Two-Mode Session Architecture β€” Lightweight (skip /start) vs Full Boot. Framework Tax concept. Orchestrator-Executor Pipeline. Crisis Architecture (P509, P519, P520, P521)
  • v9.4.4 (Mar 07 2026): GTO Routing Diagram β€” expanded from 2/8 β†’ 8/8 system cluster chains, priority tier color-coding (Critical/High/Standard/Support), reordered Q4-Q6 to match priority waterfall
  • v9.4.3 (Mar 07 2026): Maintenance β€” AGENTS.md version sync, file count corrections (138 protocols, 540+ scripts), date alignment
  • v9.4.2 (Mar 05 2026): Cognitive Architecture v2.1 β€” Homeostatic Pressure (P517), Reflexion Journaling (P515), Memory Paging (P516), LIDA Broadcast routing, deterministic exit verification, Ebbinghaus decay, context clearing
  • v9.4.1 (Mar 04 2026): Cognitive Systems v2 β€” Ideation mode (DIVERGENT/CONVERGENT), Learningβ†’Life Decision handoff, Cluster #14 safety sequence, P503 cluster sync (15), Ξ›-based stealth routing, 3 new protocols (P511 Business Viability, P512 Pre-Planning, P513 Context Isolation)
  • v9.4.0 (Mar 04 2026): Biological Stack Architecture β€” 5 new protocols (P504-P508), 8 Cognitive Systems layer, Intent Classifier, ensure_env.sh supports system Python
  • v9.3.1 (Mar 03 2026): Cross-model Audit Fixes β€” file count sync, Windows section relocation, GitHub Release catch-up (v9.2.7–v9.3.0)
  • v9.3.0 (Mar 02 2026): Onboarding Friction Audit β€” dependency restructuring (torchβ†’optional), venv instructions, PEP 668 fix, stale path cleanup, two-tier install
  • v9.2.9 (Mar 02 2026): Ultrathink v4.1 HITL Bypass β€” manual Gemini sandbox option, micro-pruned 10% dead skills (100% cluster coverage), broken reference audit
  • v9.2.8 (Feb 27 2026): Skill Template Expansion β€” 5 starter skill templates across 4 categories for new AG users
  • v9.2.7 (Feb 26 2026): Risk-proportional Triple-Lock, Tier 0 context summaries, 3 new academic citations
  • v9.2.6 (Feb 25 2026): Kilo Code + Roo Code IDE integration, COMPATIBLE_IDES.md, issue #19 closed
  • v9.2.5 (Feb 24 2026): Life Integration Protocol Stack β€” Protocols 381-383, Emotional Audit, /review workflow
  • v9.2.3 (Feb 21 2026): Multi-agent safety hardening, CLAUDE.md symlinks, issue deflection
  • v9.2.2 (Feb 21 2026): S-tier README refactor, docs restructure
  • v9.2.1 (Feb 20 2026): Deep Audit & PnC Sanitization β€” 17 patterns sanitized across 13 files
  • v9.2.0 (Feb 17 2026): Sovereignty Convergence β€” CVE patch, agentic search, governance upgrade
  • v9.1.0 (Feb 17 2026): Deep Audit & Sync β€” Fixed 15 issues (dead links, version drift)
  • v9.0.0 (Feb 16 2026): First-Principles Workspace Refactor β€” root dir cleaned, build artifacts purged

πŸ‘‰ Full Changelog β†’


🌟 Star History

Star History Chart

MIT License Β· Contributing Β· Safety Β· Security Β· Code of Conduct

Clone it. Boot it. Make it yours.

Release History

VersionChangesUrgencyDate
v9.6.5## Highlights (v9.4.9 β†’ v9.6.5) This release consolidates **7 versions** of improvements since v9.4.8. ### πŸ—οΈ Architecture - **Claude Code Architectural Integration** (v9.6.5): Context Compactor v2.0 (9-section summary + `<analysis>` scratchpad), Protocol 530 (conditional skill activation β€” ~40-60% token savings), coordinator synthesis discipline, validated patterns capture (`[V]` markers) - **Adaptive Graph of Thoughts** (v9.5.0): Protocol 75 v5.0 (AGoT-enhanced parallel reasoning), 3-tier cMedium3/31/2026

Dependencies & License Audit

Loading dependencies...

Similar Packages

langgraphBuild resilient language agents as graphs.1.1.8
langchainThe agent engineering platformlangchain-core==1.3.0
codeweaverSemantic code search for AI agents β€” 166+ languages, hybrid search, works offlinev0.1.2
aiagentflowA local-first, CLI-driven multi-agent AI software engineering workflow orchestrator with feed specs, PRDs, and guidelines to auto-generate implementation plans and code.v1.0.2
clonemeCloneMe is an advanced AI platform that builds your digital twinβ€”an AI that chats like you, remembers details, and supports multiple platforms. Customizable, memory-driven, and hot-reloadable, it's th0.0.0