Search results for "gemma"
FlashInfer: Kernel Library for LLM Serving
Deploy any AI model, agent, database, RAG, and pipeline locally or remotely in minutes
Autonomous AI agent with persistent memory, self-learning, and earned autonomy. Cognitive partner that remembers, learns, and evolves.
β‘ Lightweight offline AI agent for local models. No cloud, no API keys β just your GPU.
A high-throughput and memory-efficient inference and serving engine for LLMs
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX bac
vMLX - Home of JANG_Q - Cont Batch, Prefix, Paged, KV Cache Quant, VL - Powers MLX Studio. Image gen/edit, OpenAI/Anth
π‘ All-in-one AI framework for semantic search, LLM orchestration and language model workflows
A thin cython wrapper around llama.cpp, whisper.cpp and stable-diffusion.cpp
autonomous AI agent that builds full-stack apps. local models. no cloud. no API keys. runs on your hardware.
Curated list of the best truly open-source AI projects, models, tools, and infrastructure.
3-tier agentic ChatOps (n8n + GPT-4o + Claude Code) implementing all 21 patterns from "Agentic Design Patterns" β solo operator managing 137 devices
Open-Source Intelligent Command Layer
OllamaFreeAPI: Free Distributed API for Ollama LLMs Public gateway to our managed Ollama servers with: - Zero-configuration access to 50+ models - Auto load-balanced across global nodes - Free tier w
π€ The most comprehensive directory of AI agent frameworks, platforms, tools, and resources - hundreds of curated entries covering open-source, no-code, enterprise, and autonomous solutions. NEW Boil
Local AI server with persistent memory, RAG, and multi-backend inference (MLX / llama.cpp / Ollama). Runs entirely on your machine β zero data sent to external services.
The API layer for AI agents. Dashboard + 22K APIs + 18 Direct Call providers. MCP native.
Autonomous VAPT platform. Give it a target (FQDN, IP, CIDR) β it hunts, it reports. Inspired by the Obsidian Order.
π€ Build a smart AI assistant that learns from any website using a Retrieval-Augmented Generation framework with local models powered by Ollama.
π§ PromptDrifter β oneβcommand CI guardrail that catches prompt drift and fails the build when your LLM answers change.
