freshcrate

Search results for "gemma"

Clear filters
21 results found (Python)
flashinfer-pythonπŸ“0.6.8.post1πŸ›οΈ Flagship⭐5,467

FlashInfer: Kernel Library for LLM Serving

llamafarmπŸ“v0.0.31🌳 Mature⭐819

Deploy any AI model, agent, database, RAG, and pipeline locally or remotely in minutes

GENesis-AGIπŸ“v3.0a8🌱 Seedling⭐22

Autonomous AI agent with persistent memory, self-learning, and earned autonomy. Cognitive partner that remembers, learns, and evolves.

qwe-qweπŸ“v0.17.6🌱 Seedling⭐35

⚑ Lightweight offline AI agent for local models. No cloud, no API keys β€” just your GPU.

vllmπŸ“v0.19.1πŸ›οΈ Flagship⭐77,587

A high-throughput and memory-efficient inference and serving engine for LLMs

AReaLπŸ“v1.0.3πŸ›οΈ Flagship⭐5,075

Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.

vllm-mlxπŸ“v0.2.8🌳 Mature⭐917

OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX bac

vmlxπŸ“v1.3.34🌿 Growing⭐348

vMLX - Home of JANG_Q - Cont Batch, Prefix, Paged, KV Cache Quant, VL - Powers MLX Studio. Image gen/edit, OpenAI/Anth

txtaiπŸ“v9.7.0πŸ›οΈ Flagship⭐12,412

πŸ’‘ All-in-one AI framework for semantic search, LLM orchestration and language model workflows

cyllamaπŸ“0.2.11🌱 Seedling⭐25

A thin cython wrapper around llama.cpp, whisper.cpp and stable-diffusion.cpp

tsunamiπŸ“main@2026-04-21🌱 Seedling⭐16

autonomous AI agent that builds full-stack apps. local models. no cloud. no API keys. runs on your hardware.

awesome-opensource-aiπŸ“main@2026-04-20🌿 Growing⭐2,849

Curated list of the best truly open-source AI projects, models, tools, and infrastructure.

agentic-chatopsπŸ“main@2026-04-20🌿 Growing⭐100

3-tier agentic ChatOps (n8n + GPT-4o + Claude Code) implementing all 21 patterns from "Agentic Design Patterns" β€” solo operator managing 137 devices

ollamafreeapiπŸ“main@2026-04-15🌿 Growing⭐172

OllamaFreeAPI: Free Distributed API for Ollama LLMs Public gateway to our managed Ollama servers with: - Zero-configuration access to 50+ models - Auto load-balanced across global nodes - Free tier w

Ultimate-Agent-DirectoryπŸ“0.0.0🌱 Seedling⭐51

πŸ€– The most comprehensive directory of AI agent frameworks, platforms, tools, and resources - hundreds of curated entries covering open-source, no-code, enterprise, and autonomous solutions. NEW Boil

server-nexeπŸ“v1.0.2-beta🌱 Seedling⭐9

Local AI server with persistent memory, RAG, and multi-backend inference (MLX / llama.cpp / Ollama). Runs entirely on your machine β€” zero data sent to external services.

apiclawπŸ“v2.0.0🌱 Seedling⭐7

The API layer for AI agents. Dashboard + 22K APIs + 18 Direct Call providers. MCP native.

vikramadityaπŸ“main@2026-04-20🌱 Seedling⭐5

Autonomous VAPT platform. Give it a target (FQDN, IP, CIDR) β€” it hunts, it reports. Inspired by the Obsidian Order.

OllamaRAGπŸ“main@2026-04-21🌱 Seedling⭐5

πŸ€– Build a smart AI assistant that learns from any website using a Retrieval-Augmented Generation framework with local models powered by Ollama.

PromptDrifterπŸ“main@2026-04-19🌱 Seedling⭐8

🧭 PromptDrifter – one‑command CI guardrail that catches prompt drift and fails the build when your LLM answers change.