freshcrate

Search results for "inference"

Clear filters
12 results found (Go)
sympozium📁v0.9.1🌿 Growing417

Run a fleet of AI agents on Kubernetes. Administer your cluster agentically

WeKnora📁v0.4.0🌳 Mature13,819

LLM-powered framework for deep document understanding, semantic retrieval, and context-aware answers using RAG paradigm.

vibescan📁0.0.0🌱 Seedling48

Security scanner for AI-generated ("vibe-coded") code. Runs SAST, DAST, and sandboxed exploit simulation across 15+ languages using 30+ tools. Catches what LLMs introduce before it ships — wit

ollama📁v0.21.0🌿 Growing168,597

Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.

weaviate📁v1.37.1🌿 Growing15,988

Weaviate is an open-source vector database that stores both objects and vectors, allowing for the combination of vector search with structured filtering with the fault tolerance and scalability of a c

inference-gateway📁v0.23.6🌱 Seedling109

An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare an

go-apispec📁v0.4.7🌱 Seedling25

Generate OpenAPI 3.1 specs from Go source code via static analysis — zero annotations, automatic framework detection

AgenticGoKit📁v0.5.9🌿 Growing134

Open-source Agentic AI framework in Go for building, orchestrating, and deploying intelligent agents. LLM-agnostic, event-driven, with multi-agent workflows, MCP tool discovery, and production-grade o

ai-gateway📁v1.0.4🌿 Growing59

One API for 25+ LLMs, OpenAI, Anthropic, Bedrock, Azure. Caching, guardrails & cost controls. Go-native LiteLLM & Kong AI Gateway alternative.

Aether📁v1.0.17🌱 Seedling8

Artifical Ecology For Thought and Emergent Reasoning. The Colony That Builds With You.

chroma-go📁v0.4.1🌱 Seedling202

The Go client for Chroma vector database

llm-batch📁main@2026-04-21🌱 Seedling1

🚀 Process JSON data in batches with `llm-batch`, leveraging sequential or parallel modes for efficient interaction with LLMs.