Search results for "kimi"
5 results found (Go)
βΎοΈ Private Agent Fleet with Spec Coding. Each agent gets their own GPU-accelerated desktop. Run Claude, Codex, Gemini and open models on a full private AI Stack βΎοΈ
Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.
β‘οΈ Open-source AI Gateway β Use any SDK to call 100+ LLMs. Built-in failover, load balancing, cost control & end-to-end tracing.
mkdir beats vector DB. B-tree NeuronFS: 0-byte folders govern AI β β©0 infrastructure, ~200x token efficiency. OS-native constraint engine for LLM agents.
One API for 25+ LLMs, OpenAI, Anthropic, Bedrock, Azure. Caching, guardrails & cost controls. Go-native LiteLLM & Kong AI Gateway alternative.
