Current

Endee: Billion-Scale Local Vector Search

Endee is an open-source, hardware-native vector database engine designed to scale semantic search to one billion vectors on self-hosted infrastructure.

Signal

Scale your semantic search to one billion vectors on your own hardware · opensourceprojects · 2026-04-06

Endee introduces an open-source vector database engine optimized for scaling semantic retrieval to one billion vectors on self-hosted hardware. The project addresses the cost and data residency constraints of cloud vector services by implementing storage and indexing strategies tailored to local machine architectures. Source code is available via the endee-io/endee repository, positioning the tool as a drop-in infrastructure component for teams requiring high-capacity embedding search without external API dependencies.

Context

Production-grade semantic search typically relies on managed cloud vector databases to handle memory overhead, distributed indexing, and query routing at scale. Local alternatives frequently struggle with RAM limitations, slow disk I/O, or inefficient indexing structures when datasets exceed tens of millions of vectors. Endee attempts to bridge this gap by restructuring how vector data is paged, compressed, and queried, enabling billion-scale workloads to run on dedicated on-premises servers or high-end consumer hardware.

Relevance

This project extends the local-first infrastructure pattern by decoupling high-capacity semantic retrieval from cloud providers. It directly supports RAG pipelines, long-term agent memory, and dense embedding workflows that require strict data locality, predictable latency, and elimination of egress costs. By treating vector search as ordinary local infrastructure, it aligns with broader efforts to consolidate AI stack dependencies into self-managed environments.

Current State

Endee is in early open-source release with a public GitHub repository. The project claims billion-vector capacity on self-hosted hardware, but independent benchmarks, detailed architectural documentation, and hardware requirement specifications are not yet published. Integration guides for existing LLM frameworks, agent runtimes, or standard embedding pipelines remain absent, indicating the project is currently in an experimental or foundational stage.

Open Questions

What specific indexing architecture and on-disk storage format enable the claimed scale without prohibitive memory consumption? How does query latency, recall accuracy, and concurrent request handling compare to established local or cloud vector stores at the 100M and 1B vector thresholds? What are the minimum hardware specifications for stable production deployment, and does the engine support incremental indexing or real-time vector updates?

Connections

Operates within the local retrieval and agent memory infrastructure layer. Complements hybrid state managers like NornicDB by providing a dedicated high-scale semantic index, while aligning with local-first data ingestion patterns that prioritize on-premises storage and reduced cloud egress for autonomous workflows.

Connections

  • NornicDB - Complementary self-hosted vector storage layer; NornicDB focuses on hybrid graph/vector state management for agents, while Endee targets high-throughput billion-scale semantic retrieval on local hardware. (Current · en)

External references

Mediation note

Tooling: OpenRouter / qwen/qwen3.6-plus:free

Use: drafted entry from external signal, assessed linkage against existing knowledge base

Human role: review, edit, and approve before publication

Limits: signal content may be incomplete; verify primary sources before publishing