Merge commit 'd803bfe2b1fe7f5e219e50ac20d6801a0a58ac75' as 'vendor/ruvector'

This commit is contained in:
ruv
2026-02-28 14:39:40 -05:00
7854 changed files with 3522914 additions and 0 deletions

View File

@@ -0,0 +1,393 @@
# Ruvector System Architecture Overview
## Introduction
Ruvector is a high-performance vector database built in Rust, designed to deliver 10-100x performance improvements over Python/TypeScript implementations while maintaining full AgenticDB API compatibility.
## Architecture Principles
### 1. **Performance First**
- Zero-cost abstractions via Rust
- SIMD-optimized distance calculations
- Lock-free concurrent data structures
- Memory-mapped I/O for instant loading
### 2. **Multi-Platform**
- Single codebase deploys everywhere
- Rust native, Node.js via NAPI-RS, Browser via WASM
- CLI for standalone operation
### 3. **Production Ready**
- Memory safety without garbage collection
- ACID transactions via redb
- Crash recovery and data durability
- Extensive test coverage
### 4. **Extensible**
- Trait-based abstractions
- Pluggable distance metrics and indexes
- Advanced features as opt-in modules
## System Layers
```
┌─────────────────────────────────────────────────────────────────┐
│ Application Layer │
│ (AgenticDB API, VectorDB API, CLI Commands, MCP Tools) │
└─────────────────────────────────────────────────────────────────┘
┌─────────────────────────────────────────────────────────────────┐
│ Query Engine │
│ • Parallel search (rayon) │
│ • SIMD distance calculations (SimSIMD) │
│ • Filtered search (pre/post) │
│ • Hybrid search (vector + BM25) │
│ • MMR diversity │
└─────────────────────────────────────────────────────────────────┘
┌─────────────────────────────────────────────────────────────────┐
│ Index Layer │
│ • HNSW (hnsw_rs): O(log n) approximate search │
│ • Flat index: Brute force for small datasets │
│ • Quantized indexes: Compressed search │
└─────────────────────────────────────────────────────────────────┘
┌─────────────────────────────────────────────────────────────────┐
│ Storage Layer │
│ • Vector storage: memmap2 (zero-copy) │
│ • Metadata: redb (ACID transactions) │
│ • Index persistence: rkyv (zero-copy serialization) │
│ • AgenticDB tables: Specialized storage │
└─────────────────────────────────────────────────────────────────┘
```
## Core Components
### 1. Storage Layer
**Purpose**: Persist vectors and metadata with ACID guarantees and instant loading.
**Technologies**:
- **redb**: LMDB-inspired embedded database for metadata
- ACID transactions
- Crash recovery
- Zero-copy reads
- Pure Rust (no C dependencies)
- **memmap2**: Memory-mapped vector storage
- Zero-copy access
- OS-managed caching
- Instant loading (no deserialization)
- Supports datasets larger than RAM
- **rkyv**: Zero-copy serialization for index persistence
- Direct pointer access to serialized data
- No deserialization overhead
- Sub-second loading for billion-scale indexes
**Data Layout**:
```
vectors.db/
├── metadata.redb # redb database (vector IDs, metadata, config)
├── vectors.bin # Memory-mapped vectors (aligned f32 arrays)
├── index.rkyv # Serialized HNSW graph
└── agenticdb/ # AgenticDB specialized tables
├── reflexion.redb
├── skills.redb
├── causal.redb
└── learning.redb
```
### 2. Index Layer
**Purpose**: Fast approximate nearest neighbor (ANN) search.
**Primary: HNSW (Hierarchical Navigable Small World)**
- **Complexity**: O(log n) search, O(n log n) build
- **Recall**: 95%+ with proper tuning
- **Memory**: ~640 bytes per vector (M=32, 128D vectors)
- **Parameters**:
- `m`: Connections per node (16-64)
- `ef_construction`: Build quality (100-400)
- `ef_search`: Query-time quality (50-500)
**Implementation**: Uses `hnsw_rs` crate with custom optimizations:
- Parallel construction via rayon
- SIMD distance calculations
- Lock-free concurrent search
- Custom quantization integration
**Alternative: Flat Index**
- Brute-force exact search
- Optimal for < 10K vectors
- 100% recall
- Simple fallback when HNSW overhead not justified
### 3. Query Engine
**Purpose**: Execute searches efficiently with various strategies.
**Components**:
a) **Distance Calculation**
- **SimSIMD**: Production-ready SIMD kernels
- L2 (Euclidean)
- Cosine similarity
- Dot product
- Manhattan (L1)
- **Speedup**: 4-16x vs scalar implementations
- **Architecture support**: AVX2, AVX-512, ARM NEON/SVE
b) **Parallel Execution**
- **rayon**: Data parallelism for CPU-bound operations
- Batch inserts
- Parallel queries
- Index construction
- **Scaling**: Near-linear to CPU core count
c) **Advanced Search Strategies**
- **Filtered Search**: Metadata-based constraints
- Pre-filtering: Apply before graph traversal
- Post-filtering: Apply after retrieval
- **Hybrid Search**: Vector + keyword (BM25)
- **MMR**: Maximal Marginal Relevance for diversity
### 4. Application Layer
**Purpose**: Provide user-facing APIs across platforms.
**APIs**:
a) **Core VectorDB API**
```rust
pub trait VectorDB {
fn insert(&self, entry: VectorEntry) -> Result<VectorId>;
fn insert_batch(&self, entries: Vec<VectorEntry>) -> Result<Vec<VectorId>>;
fn search(&self, query: &SearchQuery) -> Result<Vec<SearchResult>>;
fn delete(&self, id: &VectorId) -> Result<()>;
}
```
b) **AgenticDB API** (5-table schema)
- `vectors_table`: Core embeddings
- `reflexion_episodes`: Self-critique memory
- `skills_library`: Consolidated patterns
- `causal_edges`: Cause-effect hypergraphs
- `learning_sessions`: RL training data
c) **Platform Bindings**
- **Rust**: Native library
- **Node.js**: NAPI-RS bindings with TypeScript definitions
- **WASM**: wasm-bindgen for browser
- **CLI**: clap-based command-line interface
- **MCP**: Model Context Protocol tools
## Data Flow
### Insert Operation
```
Application
↓ insert(vector, metadata)
VectorDB
↓ assign ID
↓ store metadata → redb
↓ append vector → memmap
↓ add to index → HNSW
↓ [optional] quantize
↓ persist index → rkyv
Return ID
```
**Optimizations**:
- Batch inserts amortize transaction overhead
- Parallel index updates
- Lazy quantization (on first search if enabled)
### Search Operation
```
Application
↓ search(query, k, filters)
VectorDB
↓ [optional] apply pre-filters
↓ normalize query (if cosine)
Query Engine
↓ HNSW graph traversal
↓ ├─ Start at entry point
↓ ├─ Greedy search per layer
↓ └─ Refine at bottom layer
↓ SIMD distance calculations
↓ [optional] apply post-filters
↓ [optional] re-rank with full precision
↓ top-k selection
Return results
```
**Optimizations**:
- Quantized search for initial retrieval
- Full-precision re-ranking
- SIMD vectorization
- Lock-free graph reads
## Performance Characteristics
### Time Complexity
| Operation | Complexity | Notes |
|-----------|-----------|-------|
| Insert (HNSW) | O(log n) | Amortized per insertion |
| Batch insert | O(n log n) | Parallelized across cores |
| Search (HNSW) | O(log n) | With 95% recall |
| Search (Flat) | O(n) | Exact search |
| Delete | O(log n) | Mark deleted in HNSW |
### Space Complexity
| Component | Memory per vector | Notes |
|-----------|------------------|-------|
| Full precision (128D) | 512 bytes | 128 × 4 bytes |
| HNSW graph (M=32) | ~640 bytes | M × 2 layers × 10 bytes/edge |
| Scalar quantization | 128 bytes | 4x compression |
| Product quantization | 16 bytes | 32x compression (16 subspaces) |
| Metadata | Variable | Stored in redb |
**Total for 1M vectors (128D, HNSW M=32, scalar quant)**:
- Vectors: 128 MB (quantized)
- HNSW: 640 MB
- Metadata: ~50 MB
- **Total**: ~818 MB vs ~1.2 GB uncompressed
### Latency Characteristics
**1M vectors, 128D, HNSW (M=32, ef_search=100)**:
- p50: 0.8ms
- p95: 2.1ms
- p99: 4.5ms
**Factors affecting latency**:
- Vector dimensionality (linear impact)
- Dataset size (logarithmic impact with HNSW)
- HNSW ef_search parameter (linear impact)
- Quantization (0.8-1.2x slower, but cache-friendly)
- SIMD availability (4-16x speedup)
## Concurrency Model
### Read Operations
- **Lock-free**: Multiple concurrent searches
- **Mechanism**: Arc<RwLock<T>> with read locks
- **Scalability**: Linear with CPU cores
### Write Operations
- **Exclusive lock**: Single writer at a time
- **Mechanism**: RwLock write lock
- **Batch optimization**: Amortize lock overhead
### Mixed Workloads
- Readers don't block readers
- Writers block all operations
- Read-heavy workloads scale well (typical for vector DB)
## Memory Management
### Zero-Copy Patterns
1. **Memory-mapped vectors**: OS manages paging
2. **rkyv serialization**: Direct pointer access
3. **NAPI-RS buffers**: Share TypedArrays with Node.js
4. **WASM memory**: Direct ArrayBuffer access
### Memory Safety
- Rust's ownership system prevents:
- Use-after-free
- Double-free
- Data races
- Buffer overflows
- No garbage collection overhead
### Resource Limits
- **Max vectors**: Configurable (default 10M)
- **Max dimensions**: Theoretically unlimited (practical limit ~4096)
- **Memory-mapped limit**: OS-dependent (typically 128TB on 64-bit)
## Extensibility Points
### 1. Distance Metrics
```rust
pub trait DistanceMetric: Send + Sync {
fn distance(&self, a: &[f32], b: &[f32]) -> f32;
fn batch_distance(&self, a: &[f32], batch: &[&[f32]]) -> Vec<f32>;
}
```
### 2. Index Structures
```rust
pub trait IndexStructure: Send + Sync {
fn insert(&mut self, id: VectorId, vector: &[f32]) -> Result<()>;
fn search(&self, query: &[f32], k: usize) -> Result<Vec<SearchResult>>;
fn delete(&mut self, id: VectorId) -> Result<()>;
}
```
### 3. Quantization Methods
```rust
pub trait Quantizer: Send + Sync {
type Quantized;
fn quantize(&self, vector: &[f32]) -> Self::Quantized;
fn distance(&self, a: &Self::Quantized, b: &Self::Quantized) -> f32;
}
```
## Security Considerations
### Memory Safety
- Rust prevents entire classes of vulnerabilities
- No buffer overflows, use-after-free, or data races
### Input Validation
- Vector dimension checks
- ID format validation
- Metadata size limits
- Query parameter bounds
### Resource Limits
- Maximum query size
- Rate limiting (application-level)
- Memory quotas
- Disk space monitoring
### Data Privacy
- On-premises deployment option
- No telemetry by default
- Memory zeroing on delete
- Encrypted storage (via OS-level encryption)
## Future Architecture Enhancements
### Phase 1 (Current)
- HNSW indexing
- Scalar & product quantization
- AgenticDB compatibility
- Multi-platform bindings
### Phase 2 (Near-term)
- Distributed query processing
- Horizontal scaling with sharding
- GPU acceleration for distance calculations
- Learned index structures (hybrid with HNSW)
### Phase 3 (Long-term)
- Hypergraph structures for n-ary relationships
- Temporal indexes for time-series embeddings
- Neural hash functions for improved compression
- Neuromorphic hardware support (Intel Loihi)
## Related Documentation
- [Storage Layer](STORAGE_LAYER.md) - Detailed storage architecture
- [Index Structures](INDEX_STRUCTURES.md) - HNSW and flat indexes
- [Quantization](QUANTIZATION.md) - Compression techniques
- [Performance](../optimization/PERFORMANCE_TUNING_GUIDE.md) - Optimization guide
- [API Reference](../api/) - Complete API documentation