Search results for "infrastructure"
Framework for AI Backend. Build and run AI agents like microservices - scalable, observable, and identity-aware from day one.
An API server that implements the official MCP Registry API, providing standardised access to MCP servers from multiple backends, including file-based and other API-compliant registries.
Run a fleet of AI agents on Kubernetes. Administer your cluster agentically
#1 Terminal Benchmark 2.0 β AI that ships your tickets.
LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.
Security scanner for AI-generated ("vibe-coded") code. Runs SAST, DAST, and sandboxed exploit simulation across 15+ languages using 30+ tools. Catches what LLMs introduce before it ships β wit
βΎοΈ Private Agent Fleet with Spec Coding. Each agent gets their own GPU-accelerated desktop. Run Claude, Codex, Gemini and open models on a full private AI Stack βΎοΈ
Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.
mkdir beats vector DB. B-tree NeuronFS: 0-byte folders govern AI β β©0 infrastructure, ~200x token efficiency. OS-native constraint engine for LLM agents.
The Maestro App Factory: a highly-opinionated multi-agent orchestration tool for app development that emulates the workflow of high-functioning human development teams using AI agents
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a c
Kelos - The Kubernetes-native framework for orchestrating autonomous AI coding agents.
One API for 25+ LLMs, OpenAI, Anthropic, Bedrock, Azure. Caching, guardrails & cost controls. Go-native LiteLLM & Kong AI Gateway alternative.
Open-source, self-improving autonomous agent swarmπ
Agent Sandbox is an E2B compatible, enterprise-grade ai-first, cloud-native runtime environment for AI Agents. Allows Agents to securely run untrusted LLM-generated Code, Browser use, Computer use, an
An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare an
LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
Run AI coding agents in hardened container sandboxes.
Mattermost Agents plugin supporting multiple LLMs
Automatically generate server and client framework code based on descriptive files (proto/api/sql), while using built-in jzero-skills to empower AI to generate production-ready business code adhering
A fully autonomous (no human-in-loop) agentic based project design and coding machine
A Slack bot and MCP client acts as a bridge between Slack and Model Context Protocol (MCP) servers. Using Slack as the interface, it enables large language models (LLMs) to connect and interact with v
