Turn requirements into tested, production-ready code โ with quality gates your AI can't skip.
SkillFoundry is an AI engineering framework with a centralized MCP server, 22 real tool agents, and a skill factory. v5 adds learning-driven intelligence: secret guard, deviation enforcement (161 rules), import validation, correction feedback loops, and per-project health scores โ all derived from analysis of 2,792 artifacts across 49 projects. One server serves 124+ skills to any IDE via MCP with Playwright, Semgrep, a knowledge base that learns from every failure, and an iznir-powered skill factory.
- Works where you already code โ Claude Code, Cursor, GitHub Copilot, OpenAI Codex, Google Gemini. No new IDE to learn.
- Quality gates your AI can't bypass โ The Anvil runs 8 tiers of checks (correctness contracts, banned patterns, type checking, tests, security, build, scope, deploy pre-flight) between every agent handoff. Code that fails doesn't ship.
- Persistent memory across sessions โ Decisions, errors, and patterns are stored in
memory_bank/and recalled automatically. Your AI doesn't repeat the same mistakes. - PRD-first, not vibe-coding โ Every feature starts with a Product Requirements Document. The framework validates it before writing a single line of code.
- 6 AI providers, one workflow โ Anthropic, OpenAI, xAI, Google, Ollama, LM Studio. Switch providers without changing how you work.
Parallel Dispatch Engine + Natural Language Cron
v5.6.0 adds two execution infrastructure features:
- Parallel Dispatch Engine (
sf_parallel_analyze) โ DAG-based dependency analysis for multi-story pipelines. Detects independent stories, generates execution waves, calculates critical path and estimated speedup. File conflict and migration ordering detection prevents parallel execution when it would cause merge issues. Typical speedup: 2-5x on multi-story PRDs. - Natural Language Cron (
sf_cron_compile) โ Converts English scheduling descriptions into cron expressions. Supports 20+ patterns: "every morning at 9am", "weekdays at 2:30pm", "every 30 minutes", "first day of every month". Returns cron expression, human description, next 5 run times, and confidence level. Includes cron validator.
- LLM Context Summarizer (Haiku-powered), Memory Nudge System (13 tools monitored), FTS5 Session Search (sf_memory_search).
- Response Optimizer (JSON compaction, concise mode, output truncation), Token Tracker (SQLite-persisted), sf_token_report tool.
- GateGuard (force-read-before-edit), Config Protection (30+ linter/formatter configs), Session Quality Report, Session Lifecycle hooks.
- Auto-injected security stories (SEC-001โSEC-006), Multi-Tenant Isolation Gate (7 checks), 10 deviation patterns (MT-001โMT-010).
- 22 tool agents, Secret Guard, Deviation Enforcer (161 rules), Import Validator, Correction Loop, Health Scores, harness engineering upgrade.
Connect from any IDE:
"mcpServers": { "skillfoundry": { "url": "http://localhost:9877/mcp/sse" } }
# npm (recommended)
npm install -g skillfoundry
# Homebrew (macOS)
brew install samibs/tap/skillfoundry
# One-liner (Linux/macOS)
curl -fsSL https://raw.githubusercontent.com/samibs/skillfoundry/main/scripts/install-global.sh | bash
# npx (no install needed)
npx skillfoundry initFive ways to install โ pick what fits your workflow:
# Option A: npx (quickest โ no clone needed)
cd ~/my-project
npx skillfoundry init
# Option B: npm global install (persistent CLI)
npm install -g skillfoundry
cd ~/my-project && skillfoundry init
# Option C: Homebrew (macOS)
brew install samibs/tap/skillfoundry
cd ~/my-project && skillfoundry init
# Option D: curl one-liner (Linux/macOS โ installs via npm)
curl -fsSL https://raw.githubusercontent.com/samibs/skillfoundry/main/scripts/install-global.sh | bash
cd ~/my-project && skillfoundry init
# Option E: git clone (full source โ for contributors and power users)
git clone https://github.com/samibs/skillfoundry.git ~/dev-tools/skillfoundry
cd ~/my-project && ~/dev-tools/skillfoundry/install.shThen use it in your AI IDE:
/prd "add user authentication" # write requirements
/forge # build everything with quality gatesThat's it. The installer copies agents and skills into your project, builds the CLI, and adds sf to your PATH.
Windows (PowerShell)
# Option A: npx
cd C:\MyProject
npx skillfoundry init
# Option B: npm global
npm install -g skillfoundry
cd C:\MyProject; skillfoundry init
# Option C: git clone
git clone https://github.com/samibs/skillfoundry.git C:\DevTools\skillfoundry
cd C:\MyProject
C:\DevTools\skillfoundry\install.ps1Requires Node.js v20+ for the standalone CLI. IDE skills work without Node.js. Cross-platform: Works on Linux, macOS, Windows (native, Git Bash, and WSL). Quality gates and anvil scripts auto-detect the environment.
New to SkillFoundry? See the Todo API example โ a complete project built from a single PRD in two commands. For model recommendations, see Model Compatibility.
/prd "feature" Write requirements (saved to genesis/)
โ
/forge The full pipeline:
โ
โโโ Validate PRD Are requirements complete?
โโโ Generate stories Break into implementable units
โโโ Implement Architect โ Coder โ Tester pipeline
โโโ Quality gates The Anvil (T0-T7) + Micro-Gates (MG0-MG3)
โโโ Circuit breaker Halt on repeated systemic errors
โโโ Security audit OWASP scan + dependency CVEs + credential check
โโโ Harvest knowledge Save lessons to memory_bank/
โโโ Quality metrics Track gate pass rates, trends, and industry baselines
Or use autonomous mode โ just type what you want in plain English:
/autonomous on
> "add dark mode to the dashboard" โ auto-classified as FEATURE โ full pipeline runs
> "the login is broken" โ auto-classified as BUG โ debugger + fixer dispatched
SkillFoundry has two independent systems. They share the same agents and philosophy, but work differently:
| IDE Skills (64 skills) | Standalone CLI (sf) |
|
|---|---|---|
| What it is | Markdown instruction files your AI reads | Terminal app with its own AI connection |
| Runs inside | Claude Code, Copilot, Cursor, Codex, Gemini | Your terminal (any OS) |
| Full pipeline | /forge, /go, /goma (all 64 skills) |
/forge, /plan, /gates (23 commands) |
| Autonomous mode | /goma โ full autonomous with safety gates |
Not available |
| Provider switching | Uses your IDE's provider | Built-in: 6 providers, switch at runtime |
| Budget controls | Not available | Per-run and monthly cost caps |
| Persistent memory | /memory, /gohm |
/memory, /lessons |
| Requires | An AI coding tool | Node.js v20+ |
Most users should start with IDE skills. The CLI is for standalone use or when you want provider switching and budget controls.
64 skills install directly into your AI coding tool. This is the full SkillFoundry experience โ all agents, all orchestration, autonomous mode, everything.
| Platform | Invocation | Example |
|---|---|---|
| Claude Code | /command |
/forge, /go, /goma, /review |
| GitHub Copilot | @agent in chat |
@forge, @coder, @tester |
| Cursor | Auto-loaded rules | Rules activate on context |
| OpenAI Codex | $command |
$forge, $go, $review |
| Google Gemini | Skill invocation | forge, go, review |
# Full pipeline โ works in Claude Code, Copilot, Cursor, Codex, Gemini
/prd "add user authentication" # create requirements
/forge # validate โ implement โ gate โ audit โ harvest
/goma # autonomous mode: just describe what you want
# Individual agents
/coder # code implementation
/review # code review
/security audit # security scan
/memory recall "auth" # recall lessons from previous sessionsA separate terminal app with its own AI connection. Useful for provider switching, budget controls, and working outside an IDE. Has 23 native commands (not all 64 skills).
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โ SkillFoundry CLI anthropic:claude-sonnet โ team:dev โ $0.00 โ 14.2k tok โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โธ sf:coder> I'll help you add dark mode. Let me look at the existing
โ dashboard code...
โ
โ โธ bash npm test โ 0.8s
โ โ read src/styles/theme.ts โ 0.1s
โ โ write src/styles/dark.ts โ 0.1s
โ
โ โ routed:high โ 142 in / 387 out โ $0.0045
โญโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฎ
โ โซ looks good, now review for accessibility issues โ
โฐโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฏ
โ โธ sf:review> I'll review the dark mode implementation for accessibility...
Summon a team once, and messages auto-route to the best agent for the job. No manual switching.
/team dev โ coder, tester, fixer, review, debugger
/team security โ security, review, tester
/team fullstack โ architect, coder, tester, review, debugger, docs
/team ops โ devops, sre, performance, security
/team ship โ coder, tester, review, release, docs
/team custom coder review tester โ build your own roster
Routing is keyword-based with weighted patterns โ no extra LLM calls, deterministic and fast.
Every handoff passes through a 7-tier quality pipeline plus AI-powered micro-gates:
โ The Anvil
โฃโ T0 โ Correctness Contract 0.1s
โฃโ T1 โ Banned Patterns & Syntax 0.2s
โฃโ T2 โ Type Check 1.1s
โฃโ T3 โ Tests 3.4s
โฃโ T4 โ Security Scan 0.8s
โฃโ T5 โ Build 2.1s
โโ T6 โ Scope Validation 0.3s
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ โ VERDICT: PASS 6P 0F 0W 1S (8.0s) โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ Micro-Gates (per story)
โฃโ MG0 โ AC Validation (static) PASS
โฃโ MG1 โ Security Review (AI) PASS
โฃโ MG1.5 โ Test Documentation (AI) PASS
โฃโ MG2 โ Standards Review (AI) PASS
โโ MG3 โ Cross-Story Review (AI) PASS (advisory)
Micro-gates are lightweight reviews at pipeline handoff points. MG0 validates acceptance criteria are objectively testable (static, no AI). MG1 (security) and MG2 (standards) run after each story โ if they fail, the fixer is triggered automatically. MG1.5 checks test documentation quality (re-triggers tester on failure). MG3 reviews cross-story consistency before the TEMPER phase (advisory only).
Every non-trivial feature starts with a Product Requirements Document. No PRD = no implementation.
/prd "user authentication with OAuth2" โ Creates genesis/2026-02-23-auth.md
/go โ Validates PRDs โ generates stories โ implements
/forge โ Real AI pipeline: PRDs โ stories โ implement โ gates โ report
/forge --dry-run โ Read-only scan (no AI execution)
Switch between 6 AI providers without changing your workflow:
| Provider | Env Variable | Default Model |
|---|---|---|
| Anthropic Claude | ANTHROPIC_API_KEY |
claude-sonnet-4 |
| OpenAI | OPENAI_API_KEY |
gpt-4o |
| xAI Grok | XAI_API_KEY |
grok-3 |
| Google Gemini | GOOGLE_API_KEY |
gemini-2.5-flash |
| Ollama (local) | OLLAMA_BASE_URL |
llama3.1 |
| LM Studio (local) | LMSTUDIO_BASE_URL |
qwen2.5-coder-7b |
sf setup --provider anthropic --key sk-ant-... # persistent storage
/provider set openai # switch at runtime
/provider set lmstudio # use LM Studio locallyPer-run and monthly cost caps with real-time tracking:
# .skillfoundry/config.toml
[budget]
monthly_limit_usd = 50.00
per_run_limit_usd = 2.00Token usage and cost are shown live in the header during streaming.
Every pipeline run and interactive session produces structured JSONL logs for troubleshooting. One JSON object per line โ parseable, grep-friendly, and timestamped.
.skillfoundry/logs/
โโโ session.log # Rolling log for interactive mode (last 1000 lines)
โโโ forge-20260305-abc123.log # Per-pipeline-run log
โโโ ... # Auto-cleanup keeps last 20
{"ts":"2026-03-05T12:07:14.123Z","level":"INFO","category":"pipeline","event":"phase_start","data":{"phase":"FORGE"}}
{"ts":"2026-03-05T12:07:15.456Z","level":"ERROR","category":"provider","event":"connection_error","data":{"provider":"openai","error":"ECONNREFUSED"}}
{"ts":"2026-03-05T12:07:16.789Z","level":"WARN","category":"provider","event":"retry_attempt","data":{"attempt":2,"delayMs":1000}}Configure the log level in .skillfoundry/config.toml:
log_level = "info" # debug | info | warn | errorSet to debug for per-turn token counts and tool call tracing. Default info captures phase transitions, gate results, and errors.
/forge tracks story completion and resumes where it left off. If a run is interrupted or partially fails, re-running /forge skips already-completed stories and implements only the remaining ones.
Implementing 2 stories (3 already done, skipped)
Stories are marked status: DONE in their .md files after successful completion, including passing micro-gates and fixer loops.
Use local models (Ollama, LM Studio) for free, offline AI โ with automatic fallback to cloud when needed.
# .skillfoundry/config.toml
[routing]
route_local_first = true # Enable local-first routing
local_provider = "ollama" # or "lmstudio"
local_model = "llama3.1" # your preferred local model
context_window = 0 # 0 = auto-detect from modelWhat happens when enabled:
- Simple tasks (docs, formatting, boilerplate) route to your local model (free)
- Complex tasks (architecture, security, refactoring) route to cloud (paid)
- Context compaction automatically fits prompts within local model limits (4K-32K)
- If the local model is offline, cloud fallback activates with a warning
/cost โ Shows local vs cloud token breakdown + estimated savings
/config route_local_first true โ Enable routing
/provider set lmstudio โ Switch to LM Studio
Tool outputs are automatically compressed before reaching the AI context โ saving 60-90% of tokens on common operations. No configuration needed.
| Command | Before | After | Savings |
|---|---|---|---|
git status |
~2,000 tokens | ~400 tokens | -80% |
git push/commit |
~1,600 tokens | ~120 tokens | -92% |
npm test (all pass) |
~25,000 tokens | ~2,500 tokens | -90% |
tsc --noEmit |
~8,000 tokens | ~1,600 tokens | -80% |
npm install |
~3,000 tokens | ~450 tokens | -85% |
The compressor detects command type via regex (<1ms), applies type-specific filtering (strip noise, group errors, deduplicate, show failures only), and falls back to standard truncation for unrecognized commands. Full output is preserved on errors for debugging context.
The AI executes tools with permission controls and dangerous command blocking:
| Tool | Icon | Purpose |
|---|---|---|
bash |
โธ |
Run shell commands |
read |
โ |
Read files with line numbers |
write |
โ |
Create or overwrite files |
glob |
โถ |
Find files by pattern |
grep |
โฃ |
Search file contents |
debug_start |
Start a debug session (Node.js) | |
debug_breakpoint |
Set/remove/list breakpoints | |
debug_inspect |
Inspect scope, callstack, variables | |
debug_evaluate |
Evaluate expressions in paused frame | |
debug_step |
Step over/into/out/continue/pause | |
debug_stop |
Terminate debug session |
Permission modes: auto (read auto-approved, write asks), ask (prompt every time), trusted (allow all), deny (block all).
AI agents can debug your code with real breakpoints, variable inspection, and expression evaluation โ using the Chrome DevTools Protocol (CDP) over WebSocket.
/debug src/server.ts # Start a debug session, paused at entry
/debug src/server.ts:42 # Start and set breakpoint at line 42
/debug test src/auth.test.ts # Debug a test file via test runner
The debugger:
- Spawns
node --inspect-brk=0with source map support - Connects via a custom MinimalWebSocket (Node 20 compatible, no polyfills)
- Enforces singleton sessions (one at a time) with configurable timeout (default 60s, max 5min)
- Localhost-only WebSocket connections (security hardened)
- SIGTERM โ SIGKILL escalation for process cleanup
SkillFoundry remembers across sessions. Every decision, error, and pattern is stored in memory_bank/ using append-only JSONL with weighted relevance ranking. Agents query this memory automatically โ so they don't repeat mistakes or forget conventions.
memory_bank/
โโโ knowledge/
โ โโโ bootstrap.jsonl Pre-seeded framework knowledge
โ โโโ facts.jsonl Verified technical facts
โ โโโ decisions.jsonl Design decisions with rationale
โ โโโ errors.jsonl Error patterns and their solutions
โ โโโ preferences.jsonl User preferences and conventions
โโโ relationships/
โ โโโ knowledge-graph.json Node/edge relationship graph
โ โโโ lineage.json Correction chains and lineage
โโโ retrieval/
โโโ query-cache.json Recent query cache
โโโ weights.json Weight adjustment history
Automatic pipeline harvesting: Every /forge run automatically writes knowledge entries to memory_bank/knowledge/*.jsonl โ run summaries, failed story errors, micro-gate findings, and gate verdicts. No manual invocation needed.
In the CLI:
/memory stats Show memory bank statistics
/memory recall "authentication" Find relevant lessons
In any platform (Claude Code, Cursor, etc.):
/gohm Harvest lessons from current session
/memory recall "JWT" Recall what you learned about JWT
Via shell scripts:
scripts/memory.sh remember "Use RS256 for JWT, never HS256" decision
scripts/memory.sh recall "database migration"
scripts/memory.sh statusWeight system: Entries start at 0.5 weight. Validated-by-test entries gain +0.2. Retrieved-and-used entries gain +0.1. Corrected entries drop to 0.3 while the correction starts at 0.7. Higher weight = higher retrieval priority.
Knowledge doesn't stay locked in one project. The sync daemon pushes lessons to a central GitHub repository and pulls global lessons back โ so patterns learned in project A are available in project B.
# One-time setup: connect a global knowledge repo
scripts/knowledge-sync.sh init https://github.com/you/dev-memory.git
# Start the background sync daemon (syncs every 5 minutes)
scripts/knowledge-sync.sh start
# Manual sync
scripts/knowledge-sync.sh sync
# Promote recurring error patterns to global lessons
scripts/knowledge-sync.sh promote
# Register a new project for cross-project sync
scripts/knowledge-sync.sh register /path/to/projectHarvest engine: After a session, scripts/harvest.sh extracts decisions, errors, and patterns from one or all registered projects into the central memory_bank/knowledge/ universal files. Entries that repeat 3+ times across projects get auto-promoted to global lessons.
Toggle autonomous mode and stop typing commands โ just describe what you want in plain English. SkillFoundry classifies your intent and routes to the correct pipeline automatically.
/autonomous on Enable autonomous mode
Once active, every message is classified and routed:
| You Type | Classified As | Pipeline |
|---|---|---|
| "add dark mode to the dashboard" | FEATURE | Architect โ Coder โ Tester โ Gate-Keeper |
| "the login is broken" | BUG | Debugger โ Fixer โ Tester |
| "clean up the auth module" | REFACTOR | Architect โ Coder โ Tester |
| "how does the payment flow work?" | QUESTION | Explain (read-only, no file changes) |
| "deploy to staging" | OPS | Ship / DevOps pipeline |
| "remember: we use RS256 for JWT" | MEMORY | Write to memory_bank/ |
Complex features automatically get a PRD generated in genesis/, stories broken out, and the full agent pipeline executed โ with quality gates between every handoff.
/autonomous off Back to manual command mode
/autonomous status Check if autonomous mode is active
SkillFoundry ships a native VS Code extension that brings quality gates, telemetry, and forge runs directly into your editor โ no terminal switching required.
โโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโ
โ Explorer โ Editor โ SF Sidebar โ
โ โ โ โ
โ โ src/auth.ts โ โ Dashboard โ
โ โ โโโโโโโโโโโโโโโโโ โ Pass Rate 94%โ
โ โ 1 โ import { hash } โ Last Forge โ โ
โ โ 2 โ // TODO: add rate โ โ CVEs: 2 high โ
โ โ โ โธ Run T1 (Patterns) โ โ
โ โ โ โธ Run T4 (Security) โ โ Gate Status โ
โ โ โ T0-T7 resultsโ
โ โ โ โ
โ โ โ โ Forge โ
โ โ โ Phase: TEMPERโ
โ โ โ Story: 5/8 โ
โโโโโโโโโโโโดโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโดโโโโโโโโโโโโโโโโค
โ SF: 94% gates โ sf:coder โ $0.12 Output โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Features:
- Sidebar dashboard โ gate pass rate, security findings, telemetry trends, dependency CVEs
- Inline diagnostics โ gate findings appear as squiggly underlines (like ESLint)
- CodeLens โ "Run T3 (Tests)" above test files, "Run T1/T4" above source files
- Command palette โ 12 commands: gate, forge, metrics, report, memory recall, PRD, hooks, benchmark
- Forge monitor โ real-time phase progress in sidebar, runs in integrated terminal
- Status bar โ gate pass rate with color coding, click to open metrics
- File watcher โ auto-refreshes dashboard when telemetry updates
# From VS Code Marketplace (recommended)
# Search "SkillFoundry" in the Extensions panel, or:
code --install-extension skillfoundry.skillfoundry
# Or build from source
cd skillfoundry-vscode && npm install && npm run build
code --install-extension skillfoundry-0.1.0.vsixAgents improve over time through a debate-implement-iterate loop:
scripts/evolve.sh debate Agents debate improvements
scripts/evolve.sh implement --auto-fix Apply winning proposals
scripts/evolve.sh iterate Refine through multiple rounds
scripts/evolve.sh run Full evolution cycleThese work inside the sf terminal app:
| Command | Purpose |
|---|---|
/help |
List available commands |
/setup |
Configure API keys |
/status |
Session info, provider, budget |
/team <name> |
Summon a team (dev, security, ops, fullstack, ship) |
/agent <name> |
Activate a single agent (coder, review, tester, etc.) |
/plan <task> |
Generate an implementation plan |
/apply [plan-id] |
Execute a plan with quality gates |
/gates [target] |
Run The Anvil quality gates (T1-T7) |
/gate <t0-t7|all> |
Run a single quality gate or all gates |
/forge |
Pipeline: validate PRDs โ implement โ gate โ report |
/forge --dry-run |
Read-only scan without execution |
/provider [set <name>] |
Switch AI provider |
/model [model-name] |
List or switch the AI model |
/cost |
Token usage and cost report |
/memory [stats|recall] |
Query or record knowledge |
/config [key] [value] |
View or edit configuration |
/metrics [--window N] |
Quality metrics dashboard with trends |
/report [--format md|json] |
Generate exportable quality report |
/benchmark |
Compare quality against industry baselines |
/hook install|uninstall|status |
Manage git hook integration for quality gates |
/runtime |
Runtime intelligence status (message bus, agent pool, vector store) |
/prd review <path> |
Score a PRD on 4 dimensions with actionable feedback |
/lessons |
Query and manage knowledge bank entries |
These work inside your AI coding tool, not in the sf CLI:
| Skill | Purpose |
|---|---|
/forge |
Full 6-phase pipeline (Ignite โ Forge โ Temper โ Inspect โ Remember โ Debrief) |
/go |
PRD-first orchestrator: validate โ stories โ implement |
/goma |
Autonomous mode: classify intent, route to pipeline, execute |
/prd "idea" |
Create a Product Requirements Document |
/coder |
Code implementation agent |
/tester |
Test generation and validation |
/review |
Code review |
/security |
Security audit (OWASP, credentials, banned patterns) |
/architect |
System design and architecture |
/debug |
Interactive debugger (breakpoints, scope, evaluate) |
/layer-check |
Three-layer validation (DB โ Backend โ Frontend) |
/memory |
Knowledge management |
/gohm |
Harvest lessons from current session |
/autonomous |
Toggle autonomous developer loop |
| ...and 50 more | See /help in your IDE for the full list |
Note:
/forgeexists in both systems but they are different implementations. The IDE skill orchestrates sub-agents; the CLI command runs a self-contained pipeline.
The framework generates platform-specific configurations during install. Each platform gets the same 64 skills adapted to its native format:
| Platform | What Gets Installed | How to Invoke | Notes |
|---|---|---|---|
| Claude Code | .claude/commands/ (64 skills) |
/command |
Slash commands in Claude Code CLI |
| GitHub Copilot | .copilot/custom-agents/ (60 agents) |
@agent in chat |
Custom agents in Copilot Chat |
| Cursor | .cursor/rules/ (60 rules) |
Auto-loaded | Rules activate based on context |
| OpenAI Codex | .agents/skills/ (64 skills) |
$command |
Dollar-prefix commands in Codex CLI |
| Google Gemini | .gemini/skills/ (64 skills) |
Skill invocation | Available in Gemini sessions |
Install multiple platforms at once:
./install.sh --platform="claude,cursor,copilot" # Linux/macOS
./install.ps1 -Platform "claude,cursor,copilot" # WindowsAll 64 skills work identically across platforms. The installer translates agent contracts from agents/ into each platform's native format. When you update the framework, update.sh / update.ps1 regenerates all platform files.
skillfoundry/
โโโ sf_cli/ Interactive CLI (Node.js + React/Ink)
โ โโโ src/core/ Provider adapters, tools, permissions, gates, micro-gates, budget, compaction, output compression, debugger (CDP), message bus, agent pool, embedding service, vector store, weight learner, dependency scanner, report generator, gitleaks scanner, checkov scanner, semantic search, PRD scorer
โ โโโ src/components/ Terminal UI: Header, Input, Message, GateTimeline, ...
โ โโโ src/commands/ Slash command handlers (/team, /agent, /plan, ...)
โ โโโ src/hooks/ Session state and streaming hooks
โโโ skillfoundry-vscode/ VS Code Extension (bridge to sf_cli)
โ โโโ src/extension.ts Activation, lifecycle, file watcher
โ โโโ src/bridge.ts Thin adapter calling sf_cli core modules
โ โโโ src/providers/ Dashboard, gate timeline, dependency CVE tree, forge monitor, diagnostics, CodeLens, status bar
โ โโโ src/commands/ Gate, forge, memory, PRD, metrics, report commands
โโโ agents/ 56 agent contracts and orchestration protocols
โโโ genesis/ PRD templates and your feature documents
โโโ memory_bank/ Persistent knowledge across sessions
โ โโโ knowledge/ facts, decisions, errors, preferences (JSONL)
โ โโโ relationships/ Knowledge graph and lineage tracking
โโโ scripts/ Shell tooling (works without the CLI)
โ โโโ env-preflight.sh Environment audit โ JSON (interpreters, deps, .env safety)
โ โโโ auto-harvest-cron.sh Cron sweep: harvest all registered projects on schedule
โ โโโ setup-auto-harvest.sh One-command installer for cron + Claude Code hooks
โ โโโ session-monitor.sh Real-time agent behavior monitor (PostToolUse hook)
โ โโโ memory.sh Remember, recall, correct knowledge
โ โโโ harvest.sh Extract lessons from projects
โ โโโ knowledge-sync.sh Sync daemon for cross-project learning
โ โโโ anvil.sh Quality gate runner
โ โโโ evolve.sh Agent evolution (debate โ implement โ iterate)
โ โโโ session-init.sh Session startup (pull knowledge, start daemon)
โ โโโ session-close.sh Session teardown (harvest, sync, stop daemon)
โโโ compliance/ GDPR, HIPAA, SOC2 profiles and automated checks
โโโ observability/ Audit logging, metrics collection, trace viewer
โ
โ Platform skill files (generated by installer):
โโโ .claude/commands/ Claude Code (64 skills)
โโโ .copilot/custom-agents/ GitHub Copilot (60 agents)
โโโ .cursor/rules/ Cursor (60 rules)
โโโ .agents/skills/ OpenAI Codex (64 skills)
โโโ .gemini/skills/ Google Gemini (64 skills)
Release History
| Version | Changes | Urgency | Date |
|---|---|---|---|
| v2.0.61 | ## Documentation & Help Sync for 8-Gate Pipeline Comprehensive update syncing all documentation, in-app help, and VS Code extension to reflect the new T7 deployment pre-flight gate. ### Updated Surfaces - **README** โ badge, What's New, command reference, ASCII mockup (T0-T7) - **VS Code Extension v1.2.0** โ T7 in gate timeline reference, tier picker, changelog - **site-docs** โ architecture.md gate table, getting-started.md pipeline output - **site** โ version references updated - **sf_cli** | Low | 3/17/2026 |
| v2.0.60 | ## T7: Deployment Pre-Flight Gate New gate that catches deployment-environment issues **before** they hit production โ preventing the exact class of bugs that burned 4+ debugging rounds in a real session: ### What it catches | Check | Example Issue | Verdict | |-------|--------------|---------| | DB migration state | Alembic heads โ current | FAIL | | Missing DATABASE_URL | `.env` without it | FAIL | | CORS www variant | `example.com` without `www.example.com` | WARN | | Hardcoded API URL | ` | Low | 3/17/2026 |
| v2.0.59 | ## What's New - **CLI Jurisdiction Config** โ `data_jurisdiction` and `quality_fallback` now editable via `/config` - **Status Dashboard** โ `/status` shows active jurisdiction mode and quality fallback state - **Site & Docs** โ Version bumped, README updated ### Full Changelog https://github.com/samibs/skillfoundry/blob/main/CHANGELOG.md | Low | 3/16/2026 |
| v2.0.57 | ## What's Changed (v2.0.54 โ v2.0.57) ### VS Code Extension: Live Gate Execution The VS Code extension now runs **real quality gates, dependency scans, reports, and metrics** โ no more empty panels or static references. **sf-runner.mjs โ ESM/CJS Bridge:** - New `sf_cli/bin/sf-runner.mjs` wrapper that bridges VS Code's CJS bundle to sf_cli's ESM core modules - Supports `--gate-all`, `--gate T1`, `--scan-deps`, `--report`, `--metrics` with `--workdir` and `--target` options - Sets `SF_FRAMEWORK | Low | 3/16/2026 |
| v2.0.54 | ## What's Changed (v2.0.43 โ v2.0.54) This release covers 12 versions of improvements since v2.0.42. Major highlights: ### Phase 1: Make It Reachable (v2.0.52) - **GitHub Actions release workflow** โ automated npm publish with provenance + GitHub Releases on semver tag push - **Homebrew formula** โ `brew install samibs/tap/skillfoundry` - **curl|bash installer** โ one-liner install with Node.js version check - **Docusaurus 3 documentation site** โ full docs with dark mode, offline search, GitH | Low | 3/16/2026 |
| v2.0.42 | ## What's New This release addresses adoption feedback from an [external review](https://github.com/samibs/skillfoundry/issues/7), focusing on lowering the barrier to entry. ### Lightweight Installation (#7) Three ways to install โ no git clone required: ```bash npx skillfoundry init # quickest npm install -g skillfoundry # persistent CLI git clone ... && ./install.sh # contributors ``` New root `package.json` and `bin/skillfoundry.js` wrapper dele | Low | 3/13/2026 |

