Search results for "streaming"
Framework for AI Backend. Build and run AI agents like microservices - scalable, observable, and identity-aware from day one.
No description
Mattermost Agents plugin supporting multiple LLMs
Fastest enterprise AI gateway (50x faster than LiteLLM) with adaptive load balancer, cluster mode, guardrails, 1000+ models support & <100 ยตs overhead at 5k RPS.
The ultimate LLM/AI application development framework in Go.
GoClaw - GoClaw is OpenClaw rebuilt in Go โ with multi-tenant isolation, 5-layer security, and native concurrency. Deploy AI agent teams at scale without compromising on safety.
Nornicdb is a low-latency, Graph + Vector, Temporal MVCC with all sub-ms HNSW search, graph traversal, and writes. Uses Neo4j Bolt/Cypher and qdrant's gRPC drivers so you can switch with no changes. T
A unified AI model hub for aggregation & distribution. It supports cross-converting various LLMs into OpenAI-compatible, Claude-compatible, or Gemini-compatible formats. A centralized gateway for pers
Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search
LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.
Design-first Go framework that generates API code, documentation, and clients. Define once in an elegant DSL, deploy as HTTP and gRPC services with zero drift between code and docs.
Open-source Agentic AI framework in Go for building, orchestrating, and deploying intelligent agents. LLM-agnostic, event-driven, with multi-agent workflows, MCP tool discovery, and production-grade o
Privacy-first LLM proxy and AI gateway โ load balancing, multi-provider routing, API key management, usage tracking, rate limiting. Self-hosted. Zero knowledge of your prompts.
One API for 25+ LLMs, OpenAI, Anthropic, Bedrock, Azure. Caching, guardrails & cost controls. Go-native LiteLLM & Kong AI Gateway alternative.
An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare an
trpc-agent-go is a powerful Go framework for building intelligent agent systems using large language models (LLMs) and tools.
LocalAI is the open-source AI engine. Run any model - LLMs, vision, voice, image, video - on any hardware. No GPU required.
A fully autonomous (no human-in-loop) agentic based project design and coding machine
ToolHive's registry catalog of MCP servers
High-performance zero-dependency L4/L7 load balancer written in Go. Single binary with Web UI, clustering, MCP/AI integration. 8.5K RPS, 39 E2E tests.
Zero trust LLM gateway. OpenAI-compatible proxy with semantic routing and load balancing across OpenAI, Anthropic, Ollama, vLLM, and any compatible backend. Identity-based access, virtual A
A Slack bot and MCP client acts as a bridge between Slack and Model Context Protocol (MCP) servers. Using Slack as the interface, it enables large language models (LLMs) to connect and interact with v
soulshack, an irc chatbot. openai/ollama/gemini/anthropic apis. basic shell tools and mcp server support.
A minimal, lightweight structured data store designed for small applications, scripts and automation workflows. Built for simplicity, portability and low overhead.
A self-evolving coding agent written in Go. Reads its own source, decides what to improve, writes code, runs tests, and commits โ autonomously.
Multi-LLM agent orchestration TUI โ parallel Claude/Gemini/Codex sessions, 126 MCP tools
Self-hosted AI workflow orchestration server. Runs multi-phase LLM pipelines (Director โ Architect โ Implementer โ QA) and delivers structured artifacts via PR, webhook, or bundle.
๐ Process JSON data in batches with `llm-batch`, leveraging sequential or parallel modes for efficient interaction with LLMs.
