Search results for "retrieval"
SeekStorm: vector & lexical search - in-process library & multi-tenancy server, in Rust.
EdegQuake 🌋 High-performance GraphRAG inspired from LightRag written in Rust; Transform documents into intelligent knowledge graphs for superior retrieval and generation
Memory that lasts and compounds. MentisDB gives agents durable memory so they do not just remember, they improve over time. It stores append-only thought chains plus a Git-like skills registry, lett
Local knowledge graph for AI agents. Hybrid search + MCP server for Obsidian vaults.
Every meeting, every idea, every voice note — searchable by your AI. Open-source, privacy-first conversation memory layer.
The fastest PDF library for Python and Rust. Text extraction, image extraction, markdown conversion, PDF creation & editing. 0.8ms mean, 5× faster than industry leaders, 100% pass rate on 3,830 PDFs.
A portable accelerated SQL query, search, and LLM-inference engine, written in Rust, for data-grounded AI apps and agents.
LeanKG: Stop Burning Tokens. Start Coding Lean.
It is a simple, fast, and hard-durable embedded database designed specifically for AI agent memory. It provides a single-file-like experience (no server required) but with native support for vectors,
Semantic code searcher and codebase utility
NextPlaid, ColGREP: Multi-vector search, from database to coding agents.
An AI agent for teams, communities, and multi-user environments.
TensorZero is an open-source LLMOps platform that unifies an LLM gateway, observability, evaluation, optimization, and experimentation.
Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
Official repository of the Seismic library.
Scalable, fast, and disk-friendly vector search in Postgres, the successor of pgvecto.rs.
The graph-native hybrid retrieval engine for AI and GraphRAG. Graph + Vector + Full-Text in a single transactional engine.
Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.
⚡💾 Vectro — Compress LLM embeddings 🧠🚀 Save memory, speed up retrieval, and keep semantic accuracy 🎯✨ Lightning-fast quantization for Python + Mojo, vector DB friendly 🗄️, and perfect for RAG pip
Lightweight, modular AI agent runtime — thinks (Hrafn) and remembers (MuninnDB) 🐦⬛
Open-source autonomous AI assistant with 5-tier security, 62 tools, 14 LLM providers. Written in Rust. Single binary.
Lean Rust AI agent: 6MB binary, 7.9MB RAM. OpenClaw replacement. Telegram + Discord + GitHub auto-PR. Ollama/Anthropic support.
Enable autonomous AI workflows with a local-first, zero-trust Rust framework for high-performance multi-agent orchestration and deterministic execution.
🚀 Build memory and retrieval infrastructure for ReasonKit, enhancing data management and access for your applications with ease and efficiency.
🧠 Capture and manage your team's knowledge effortlessly with Eywa, ensuring no valuable memory is ever lost.
