Files
wifi-densepose/crates/ruQu/docs/RESEARCH_DISCOVERIES.md
ruv d803bfe2b1 Squashed 'vendor/ruvector/' content from commit b64c2172
git-subtree-dir: vendor/ruvector
git-subtree-split: b64c21726f2bb37286d9ee36a7869fef60cc6900
2026-02-28 14:39:40 -05:00

211 lines
7.9 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Research Discoveries for ruQu Enhancement
*Compiled: January 2026*
This document captures state-of-the-art research findings that can inform further improvements to ruQu's coherence gate architecture.
---
## 1. Real-Time Decoding at Scale
### DECONET System (April 2025)
**Source**: [arXiv:2504.11805](https://arxiv.org/abs/2504.11805)
DECONET is a first-of-its-kind decoding system that scales to **thousands of logical qubits** with lattice surgery support. Key innovations:
- **Network-integrated hybrid tree-grid structure**: O(log(l)) latency increase as system grows
- **Resource scaling**: O(l × log(l)) compute, O(l) I/O for l logical qubits
- **Union-Find decoder**: 100× higher accuracy than greedy algorithms
- **Prototype**: 100 logical qubits on 5 VMK-180 FPGAs
**Relevance to ruQu**: Our `ParallelFabric` uses flat parallelism. Consider hierarchical tree-grid topology for 1000+ tile scaling.
### Google Below-Threshold (2025)
**Source**: [Nature 2024](https://www.nature.com/articles/s41586-024-08449-y)
Google achieved Λ = 2.14 ± 0.02 error suppression when increasing code distance by 2, with a 101-qubit distance-7 code achieving **0.143% error per cycle**.
**Relevance to ruQu**: Our three-filter decision pipeline should target similar sub-0.2% false positive rates.
---
## 2. Hardware-Accelerated Decoding
### Riverlane Collision Clustering Decoder
**Source**: [Riverlane Blog](https://www.riverlane.com/news/introducing-the-world-s-first-low-latency-qec-experiment)
| Platform | Qubits | Latency | Power |
|----------|--------|---------|-------|
| FPGA | 881 | 810 ns | - |
| ASIC | 1,057 | **240 ns** | 8 mW |
The ASIC fits in 0.06 mm² - suitable for cryogenic deployment.
**Relevance to ruQu**: Our coherence simulation achieves 468ns P99. ASIC compilation of the hot path could reach 240ns.
### QASBA: Sparse Blossom on FPGA
**Source**: [ACM TRETS](https://dl.acm.org/doi/10.1145/3723168)
- **25× performance** vs software baseline
- **304× energy efficiency** improvement
**Relevance to ruQu**: Our min-cut computation is the hot path. FPGA synthesis of `SubpolynomialMinCut` could yield similar gains.
---
## 3. Adaptive Syndrome Extraction
### PRX Quantum (July 2025)
**Source**: [PRX Quantum](https://doi.org/10.1103/ps3r-wf84)
Adaptive syndrome extraction measures **only stabilizers likely to provide useful information**:
- **10× lower logical error rates** vs non-adaptive
- Fewer CNOT gates and physical qubits
- Uses [[4,2,2]] concatenated with hypergraph product code
**Relevance to ruQu**: This validates our coherence gate philosophy - don't process everything, focus on what matters. Consider:
- Tracking which detectors fire frequently (already in `stim.rs`)
- Skip syndrome processing for "quiet" regions
- Adaptive measurement scheduling
### Multi-Agent RL for QEC
**Source**: [arXiv:2509.03974](https://arxiv.org/pdf/2509.03974)
Uses **reinforcement learning bandits** to:
- Evaluate fidelity after recovery
- Determine when retraining is necessary
- Optimize encoder, syndrome measurement, and recovery jointly
**Relevance to ruQu**: Our `AdaptiveThresholds` uses EMA-based learning. Consider upgrading to bandit-based exploration for threshold optimization.
### Window-Based Drift Estimation (Nov 2025)
**Source**: [arXiv:2511.09491](https://arxiv.org/html/2511.09491)
Estimates noise drift profiles **from syndrome data alone**, then adapts decoder parameters.
**Relevance to ruQu**: Integrate drift detection into `adaptive.rs`:
```rust
pub fn detect_drift(&mut self, window: &[SyndromeStats]) -> Option<DriftProfile> {
// Detect if noise characteristics are shifting
// Adjust thresholds proactively
}
```
---
## 4. Mixture-of-Depths for Efficiency
### MoD (DeepMind, 2024)
**Source**: [arXiv:2404.02258](https://arxiv.org/html/2404.02258v1)
- **50% FLOPs reduction** while matching dense transformer performance
- Per-token dynamic routing (skip middle layers for "resolved" tokens)
- Different from early-exit: tokens can skip middle layers then attend
**Status**: Already implemented in `attention.rs` via `MincutDepthRouter` integration.
### Mixture-of-Recursions (NeurIPS 2025)
**Source**: [arXiv:2507.10524](https://arxiv.org/html/2507.10524v1)
Combines parameter sharing + adaptive computation:
- Reuses shared layer stack across recursion steps
- Lightweight routers assign recursion depth per-token
- Token-level early exiting for simple predictions
**Relevance to ruQu**: Consider recursive tile processing:
```rust
pub fn process_recursive(&mut self, syndrome: &SyndromeDelta, max_depth: usize) -> GateDecision {
for depth in 0..max_depth {
let decision = self.process_at_depth(syndrome, depth);
if decision.confidence > EARLY_EXIT_THRESHOLD {
return decision; // Exit early for clear cases
}
}
decision
}
```
---
## 5. Fusion Blossom Performance
### Fusion Blossom Decoder
**Source**: [arXiv:2305.08307](https://arxiv.org/abs/2305.08307), [GitHub](https://github.com/yuewuo/fusion-blossom)
- **1 million measurement rounds/second** at d=33
- **0.7 ms latency** in stream mode at d=21
- **58 ns per non-trivial measurement** on 64-core machine
- O(N) complexity for defect vertices N
**Status**: Already integrated via `decoder.rs` feature. Consider:
- Enabling parallel fusion mode in production
- Streaming mode for real-time applications
### PyMatching V2 Comparison
PyMatching V2 achieves 5-20× single-thread speedup over Fusion Blossom. The algorithms are compatible - combining them could yield another 5-20× improvement.
---
## 6. Graph Neural Networks for QEC
### QSeer (May 2025)
**Source**: [arXiv:2505.06810](https://arxiv.org/abs/2505.06810)
GNN for QAOA parameter prediction:
- 6-68% improvement in approximation ratio
- 5-10× convergence speedup
- Supports variable-depth circuits and weighted Max-Cut
**Relevance to ruQu**: Train a small GNN to predict optimal thresholds from syndrome graph structure:
```rust
pub struct ThresholdPredictor {
model: OnnxModel, // Export trained model
}
impl ThresholdPredictor {
pub fn predict(&self, graph_embedding: &[f32]) -> GateThresholds {
// Use learned model for threshold prediction
}
}
```
---
## Implementation Priority Matrix
| Enhancement | Impact | Effort | Priority |
|-------------|--------|--------|----------|
| Hierarchical tree-grid topology | High | High | P2 |
| Drift detection in adaptive.rs | High | Medium | P1 |
| Recursive early-exit processing | Medium | Low | P1 |
| Bandit-based threshold exploration | Medium | Medium | P2 |
| FPGA synthesis of min-cut | Very High | Very High | P3 |
| GNN threshold predictor | Medium | High | P3 |
| Streaming Fusion mode | High | Low | P1 |
---
## Immediate Next Steps
1. **Drift Detection**: Add window-based drift estimation to `adaptive.rs`
2. **Early-Exit Depth**: Implement confidence-based early exit in tile processing
3. **Streaming Decoder**: Enable Fusion Blossom streaming mode for <1ms latency
4. **Parallel Fusion**: Configure parallel fusion on 64+ core systems
---
## References
1. DECONET: [arxiv.org/abs/2504.11805](https://arxiv.org/abs/2504.11805)
2. Google Below-Threshold: [nature.com/articles/s41586-024-08449-y](https://www.nature.com/articles/s41586-024-08449-y)
3. Riverlane CC Decoder: [riverlane.com](https://www.riverlane.com/news/introducing-the-world-s-first-low-latency-qec-experiment)
4. Adaptive Syndrome Extraction: [doi.org/10.1103/ps3r-wf84](https://doi.org/10.1103/ps3r-wf84)
5. Multi-Agent RL QEC: [arxiv.org/pdf/2509.03974](https://arxiv.org/pdf/2509.03974)
6. Drift Estimation: [arxiv.org/html/2511.09491](https://arxiv.org/html/2511.09491)
7. Mixture-of-Depths: [arxiv.org/html/2404.02258v1](https://arxiv.org/html/2404.02258v1)
8. Mixture-of-Recursions: [arxiv.org/html/2507.10524v1](https://arxiv.org/html/2507.10524v1)
9. Fusion Blossom: [arxiv.org/abs/2305.08307](https://arxiv.org/abs/2305.08307)
10. QSeer GNN: [arxiv.org/abs/2505.06810](https://arxiv.org/abs/2505.06810)
11. QASBA FPGA: [dl.acm.org/doi/10.1145/3723168](https://dl.acm.org/doi/10.1145/3723168)