Search results for "duration"
Framework for AI Backend. Build and run AI agents like microservices - scalable, observable, and identity-aware from day one.
eBPF-based GPU causal observability agent
#1 Terminal Benchmark 2.0 — AI that ships your tickets.
LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.
Privacy-first LLM proxy and AI gateway — load balancing, multi-provider routing, API key management, usage tracking, rate limiting. Self-hosted. Zero knowledge of your prompts.
An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare an
Model Context Protocol (MCP) server for Kubernetes and OpenShift
Zero trust LLM gateway. OpenAI-compatible proxy with semantic routing and load balancing across OpenAI, Anthropic, Ollama, vLLM, and any compatible backend. Identity-based access, virtual A
Apache Arrow Flight clustered vector cache for high throughput Agent memory sharing
