Search results for "query"
Conversation memory SDK — query meeting transcripts, decisions, and action items from any AI agent or application
A portable accelerated SQL query, search, and LLM-inference engine, written in Rust, for data-grounded AI apps and agents.
A jq-like Markdown query language for command-line processing
Local knowledge graph for AI agents. Hybrid search + MCP server for Obsidian vaults.
Every meeting, every idea, every voice note — searchable by your AI. Open-source, privacy-first conversation memory layer.
EdgeCrab 🦀 A Super Powerful Personal Assistant inspired by NousHermes and OpenClaw — Rust-native, blazing-fast terminal UI, ReAct tool loop, multi-provider LLM support, ACP protocol, gateway adapters
LeanKG: Stop Burning Tokens. Start Coding Lean.
Fast, cross-platform, real-time token usage tracker and cost monitor for Gemini CLI / Claude Code / Codex CLI / Qwen Code / Cline / Roo Code / Kilo Code / GitHub Copilot / OpenCode / Pi Agent / Piebal
Markdown and OFM SDK w/ MCP server that transforms your Obsidian vault into an intelligent knowledge system
SeekStorm: vector & lexical search - in-process library & multi-tenancy server, in Rust.
EdegQuake 🌋 High-performance GraphRAG inspired from LightRag written in Rust; Transform documents into intelligent knowledge graphs for superior retrieval and generation
Memory that lasts and compounds. MentisDB gives agents durable memory so they do not just remember, they improve over time. It stores append-only thought chains plus a Git-like skills registry, lett
Plano is an AI-native proxy and data plane for agentic apps — with built-in orchestration, safety, observability, and smart LLM routing so you stay focused on your agents core logic.
Data infrastructure for AI
It is a simple, fast, and hard-durable embedded database designed specifically for AI agent memory. It provides a single-file-like experience (no server required) but with native support for vectors,
BioMCP: Biomedical Model Context Protocol
The official Rust SDK for the Model Context Protocol
Pragmatic AI Labs MCP Agent Toolkit - An MCP Server designed to make code with agents more deterministic
High-performance capability authorization engine for AI agents. Cryptographically attenuated warrants, task-scoped authority, verifiable offline. Rust core.
Semantic code searcher and codebase utility
NextPlaid, ColGREP: Multi-vector search, from database to coding agents.
Cognitive memory for AI agents — learns from use, forgets what's irrelevant, strengthens what matters. Single binary, fully offline.
OramaCore is the complete runtime you need for your projects, answer engines, copilots, and search. It includes a fully-fledged full-text search engine, vector database, LLM interface, and many more u
An AI agent for teams, communities, and multi-user environments.
HelixDB is an open-source graph-vector database built from scratch in Rust.
Uni is a modern, embedded database that combines property graph (OpenCypher), vector search, and columnar storage (Lance) into a single, cohesive engine. It is designed for applications requiring loca
Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
Git primitives for autonomous coding agents
Scalable graph analytics database powered by a multithreaded, vectorized temporal engine, written in Rust
Embedded database for agentic memory — relational, graph, and vector under unified MVCC transactions
Official repository of the Seismic library.
A high-performance, in-memory vector database written in Rust, designed for semantic search and top-k nearest neighbor queries in AI-driven applications, with binary file persistence for durability.
Hermes Agent rewritten in Rust: production-grade multi-platform AI agent runtime with gateway adapters, tool orchestration, MCP, memory plugins, and cost-safe autonomous loops.
Fast, streaming indexing, query, and agentic LLM applications in Rust
Next-Gen AI-Aware Git
Rust framework for building Model Context Protocol servers with cancel-correct async, zero-copy serialization, and first-class tool/resource/prompt abstractions
The graph-native hybrid retrieval engine for AI and GraphRAG. Graph + Vector + Full-Text in a single transactional engine.
Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.
Local-first knowledge base with semantic search for AI coding assistants. Fewer tokens, faster context.
SDK to build Agents in rust. Build and manage complex, self-correcting agent teams. Launch multi-agent systems via configuration files or using an Agent Factory. It's the robust framework for orchestr
⚡💾 Vectro — Compress LLM embeddings 🧠🚀 Save memory, speed up retrieval, and keep semantic accuracy 🎯✨ Lightning-fast quantization for Python + Mojo, vector DB friendly 🗄️, and perfect for RAG pip
Lean Rust AI agent: 6MB binary, 7.9MB RAM. OpenClaw replacement. Telegram + Discord + GitHub auto-PR. Ollama/Anthropic support.
Add formal type safety (10 levels, dependent/linear/session types) to any query language — SQL, GraphQL, Cypher, SPARQL, VQL
开源牛子 Rust-first local coding-agent CLI with a local /v1/threads server, plugins/skills, and Python/TypeScript SDK alphas.
Augment any database with VeriSimDB octad capabilities — drift detection, provenance, temporal versioning, modality overlays
Enable AI agents with fast, local semantic memory to search and recall knowledge from text files without servers or complex setup.
Complete open-source AI collaboration suite and multi-agent platform featuring LLM orchestration, automation, and virtual assistants. Scales seamlessly from small deployments to large enterprise envir
