Search results for "llm"
Fastest enterprise AI gateway (50x faster than LiteLLM) with adaptive load balancer, cluster mode, guardrails, 1000+ models support & <100 Β΅s overhead at 5k RPS.
β‘οΈ Open-source AI Gateway β Use any SDK to call 100+ LLMs. Built-in failover, load balancing, cost control & end-to-end tracing.
The ultimate LLM/AI application development framework in Go.
A unified AI model hub for aggregation & distribution. It supports cross-converting various LLMs into OpenAI-compatible, Claude-compatible, or Gemini-compatible formats. A centralized gateway for pers
Enterprise Security Data Pipeline Platform (SDPP) β Real-Time Threat Detection + Deeply Integrated LLM Agents
LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.
Nornicdb is a low-latency, Graph + Vector, Temporal MVCC with all sub-ms HNSW search, graph traversal, and writes. Uses Neo4j Bolt/Cypher and qdrant's gRPC drivers so you can switch with no changes. T
βΎοΈ Private Agent Fleet with Spec Coding. Each agent gets their own GPU-accelerated desktop. Run Claude, Codex, Gemini and open models on a full private AI Stack βΎοΈ
Framework for AI Backend. Build and run AI agents like microservices - scalable, observable, and identity-aware from day one.
A modular MCP server that provides commonly used developer tools for AI coding agents
A lightweight, embeddable vector database library for Go AI projects.
Container-free, deny-by-default sandbox for AI coding agents. Kernel-enforced filesystem, network, and syscall isolation for Linux and macOS
Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.
mkdir beats vector DB. B-tree NeuronFS: 0-byte folders govern AI β β©0 infrastructure, ~200x token efficiency. OS-native constraint engine for LLM agents.
A Go implementation of the Model Context Protocol (MCP), enabling seamless integration between LLM applications and external data sources and tools.
The Maestro App Factory: a highly-opinionated multi-agent orchestration tool for app development that emulates the workflow of high-functioning human development teams using AI agents
Open-source Agentic AI framework in Go for building, orchestrating, and deploying intelligent agents. LLM-agnostic, event-driven, with multi-agent workflows, MCP tool discovery, and production-grade o
Go framework for agentic AI app with MCP and built-in tools
Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a c
Agent Sandbox is an E2B compatible, enterprise-grade ai-first, cloud-native runtime environment for AI Agents. Allows Agents to securely run untrusted LLM-generated Code, Browser use, Computer use, an
One API for 25+ LLMs, OpenAI, Anthropic, Bedrock, Azure. Caching, guardrails & cost controls. Go-native LiteLLM & Kong AI Gateway alternative.
A repository of servers and clients from the Model Context Protocol tutorials
OasisDB: A minimal and lightweight vector database
An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare an
trpc-agent-go is a powerful Go framework for building intelligent agent systems using large language models (LLMs) and tools.
LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
Artifical Ecology For Thought and Emergent Reasoning. The Colony That Builds With You.
Your AI forgets everything between sessions. SAME fixes that. Local-first, no API keys, single binary.
The cognitive database. A new class of data storage. Not a vector store, not a graph DB, not a RAG wrapper. Ebbinghaus decay, Hebbian learning, and Bayesian confidence are engine-native primitives.
Model Context Protocol (MCP) server for Kubernetes and OpenShift
Mattermost Agents plugin supporting multiple LLMs
Type-safe AI agents for Go. Suricata combines LLM intelligence with Goβs strong typing, declarative YAML specs, and code generation to build safe, maintainable, and production-ready AI agents.
A minimal, lightweight structured data store designed for small applications, scripts and automation workflows. Built for simplicity, portability and low overhead.
A selective learning and memory substrate for agentic systems β typed, revisable, decayable memory with competence learning and trust-aware retrieval.
Autonomous local AI assistant in Go β 40+ tools, 20+ LLM providers, multi-agent orchestration, self-improving
π Ask questions about your shop data in natural language and get instant answers about appointments, customers, and repair orders with Tekmetric MCP.
A fully autonomous (no human-in-loop) agentic based project design and coding machine
A tool supports OPENAI and other LLMs with Claude Skills, you can also use it as a subagent
Multi-LLM agent orchestration TUI β parallel Claude/Gemini/Codex sessions, 126 MCP tools
π‘οΈ Enable secure, read-only SSH access for LLM agents to audit servers, run diagnostics, and inspect logs without risking data changes.
Manage and automate tasks in Yandex Tracker using a robust MCP integration for efficient issue tracking and project control.
A Slack bot and MCP client acts as a bridge between Slack and Model Context Protocol (MCP) servers. Using Slack as the interface, it enables large language models (LLMs) to connect and interact with v
Decentralized coordination protocol for autonomous agents
π‘ Enforce security policies, redact data, sandbox processes, and verify integrity for Model Context Protocol (MCP) server communication.
Convert any URL into LLM-friendly formats using a lightweight CLI tool for reading and searching web content efficiently.
π Process JSON data in batches with `llm-batch`, leveraging sequential or parallel modes for efficient interaction with LLMs.
Self-hosted AI workflow orchestration server. Runs multi-phase LLM pipelines (Director β Architect β Implementer β QA) and delivers structured artifacts via PR, webhook, or bundle.
