Search results for "benchmark"
The fastest PDF library for Python and Rust. Text extraction, image extraction, markdown conversion, PDF creation & editing. 0.8ms mean, 5× faster than industry leaders, 100% pass rate on 3,830 PDFs.
Token-efficient navigation substrate for AI coding agents. Index code once and retrieve only the symbols you need.
LeanKG: Stop Burning Tokens. Start Coding Lean.
Distributed AI/LLM for the people. Share compute privately or publicly to power your agents and chat.
SeekStorm: vector & lexical search - in-process library & multi-tenancy server, in Rust.
Fast, small, and fully autonomous AI personal assistant infrastructure, ANY OS, ANY PLATFORM — deploy anywhere, swap anything 🦀
Memory that lasts and compounds. MentisDB gives agents durable memory so they do not just remember, they improve over time. It stores append-only thought chains plus a Git-like skills registry, lett
It is a simple, fast, and hard-durable embedded database designed specifically for AI agent memory. It provides a single-file-like experience (no server required) but with native support for vectors,
BioMCP: Biomedical Model Context Protocol
Pragmatic AI Labs MCP Agent Toolkit - An MCP Server designed to make code with agents more deterministic
Semantic code searcher and codebase utility
TensorZero is an open-source LLMOps platform that unifies an LLM gateway, observability, evaluation, optimization, and experimentation.
A high-performance, in-memory vector database written in Rust, designed for semantic search and top-k nearest neighbor queries in AI-driven applications, with binary file persistence for durability.
A functional programming language optimized for LLM code generation. Compiles to Rust and WebAssembly.
A unified SDK for IronCore encryption operations including vector encryption for AI data
The graph-native hybrid retrieval engine for AI and GraphRAG. Graph + Vector + Full-Text in a single transactional engine.
Local AI anywhere, for everyone — LLM inference, chat UI, voice, agents, workflows, RAG, and image generation. No cloud, no subscriptions.
⚡💾 Vectro — Compress LLM embeddings 🧠🚀 Save memory, speed up retrieval, and keep semantic accuracy 🎯✨ Lightning-fast quantization for Python + Mojo, vector DB friendly 🗄️, and perfect for RAG pip
High-performance crystal structure modeling and DFT/MD file preparation. Native desktop app fusing a Rust/C++ physics kernel, a GPU-accelerated Metal/Vulkan renderer, and an AI-driven command bus for
