git-subtree-dir: vendor/ruvector git-subtree-split: b64c21726f2bb37286d9ee36a7869fef60cc6900
7.9 KiB
Research Discoveries for ruQu Enhancement
Compiled: January 2026
This document captures state-of-the-art research findings that can inform further improvements to ruQu's coherence gate architecture.
1. Real-Time Decoding at Scale
DECONET System (April 2025)
Source: arXiv:2504.11805
DECONET is a first-of-its-kind decoding system that scales to thousands of logical qubits with lattice surgery support. Key innovations:
- Network-integrated hybrid tree-grid structure: O(log(l)) latency increase as system grows
- Resource scaling: O(l × log(l)) compute, O(l) I/O for l logical qubits
- Union-Find decoder: 100× higher accuracy than greedy algorithms
- Prototype: 100 logical qubits on 5 VMK-180 FPGAs
Relevance to ruQu: Our ParallelFabric uses flat parallelism. Consider hierarchical tree-grid topology for 1000+ tile scaling.
Google Below-Threshold (2025)
Source: Nature 2024
Google achieved Λ = 2.14 ± 0.02 error suppression when increasing code distance by 2, with a 101-qubit distance-7 code achieving 0.143% error per cycle.
Relevance to ruQu: Our three-filter decision pipeline should target similar sub-0.2% false positive rates.
2. Hardware-Accelerated Decoding
Riverlane Collision Clustering Decoder
Source: Riverlane Blog
| Platform | Qubits | Latency | Power |
|---|---|---|---|
| FPGA | 881 | 810 ns | - |
| ASIC | 1,057 | 240 ns | 8 mW |
The ASIC fits in 0.06 mm² - suitable for cryogenic deployment.
Relevance to ruQu: Our coherence simulation achieves 468ns P99. ASIC compilation of the hot path could reach 240ns.
QASBA: Sparse Blossom on FPGA
Source: ACM TRETS
- 25× performance vs software baseline
- 304× energy efficiency improvement
Relevance to ruQu: Our min-cut computation is the hot path. FPGA synthesis of SubpolynomialMinCut could yield similar gains.
3. Adaptive Syndrome Extraction
PRX Quantum (July 2025)
Source: PRX Quantum
Adaptive syndrome extraction measures only stabilizers likely to provide useful information:
- 10× lower logical error rates vs non-adaptive
- Fewer CNOT gates and physical qubits
- Uses 4,2,2 concatenated with hypergraph product code
Relevance to ruQu: This validates our coherence gate philosophy - don't process everything, focus on what matters. Consider:
- Tracking which detectors fire frequently (already in
stim.rs) - Skip syndrome processing for "quiet" regions
- Adaptive measurement scheduling
Multi-Agent RL for QEC
Source: arXiv:2509.03974
Uses reinforcement learning bandits to:
- Evaluate fidelity after recovery
- Determine when retraining is necessary
- Optimize encoder, syndrome measurement, and recovery jointly
Relevance to ruQu: Our AdaptiveThresholds uses EMA-based learning. Consider upgrading to bandit-based exploration for threshold optimization.
Window-Based Drift Estimation (Nov 2025)
Source: arXiv:2511.09491
Estimates noise drift profiles from syndrome data alone, then adapts decoder parameters.
Relevance to ruQu: Integrate drift detection into adaptive.rs:
pub fn detect_drift(&mut self, window: &[SyndromeStats]) -> Option<DriftProfile> {
// Detect if noise characteristics are shifting
// Adjust thresholds proactively
}
4. Mixture-of-Depths for Efficiency
MoD (DeepMind, 2024)
Source: arXiv:2404.02258
- 50% FLOPs reduction while matching dense transformer performance
- Per-token dynamic routing (skip middle layers for "resolved" tokens)
- Different from early-exit: tokens can skip middle layers then attend
Status: Already implemented in attention.rs via MincutDepthRouter integration.
Mixture-of-Recursions (NeurIPS 2025)
Source: arXiv:2507.10524
Combines parameter sharing + adaptive computation:
- Reuses shared layer stack across recursion steps
- Lightweight routers assign recursion depth per-token
- Token-level early exiting for simple predictions
Relevance to ruQu: Consider recursive tile processing:
pub fn process_recursive(&mut self, syndrome: &SyndromeDelta, max_depth: usize) -> GateDecision {
for depth in 0..max_depth {
let decision = self.process_at_depth(syndrome, depth);
if decision.confidence > EARLY_EXIT_THRESHOLD {
return decision; // Exit early for clear cases
}
}
decision
}
5. Fusion Blossom Performance
Fusion Blossom Decoder
Source: arXiv:2305.08307, GitHub
- 1 million measurement rounds/second at d=33
- 0.7 ms latency in stream mode at d=21
- 58 ns per non-trivial measurement on 64-core machine
- O(N) complexity for defect vertices N
Status: Already integrated via decoder.rs feature. Consider:
- Enabling parallel fusion mode in production
- Streaming mode for real-time applications
PyMatching V2 Comparison
PyMatching V2 achieves 5-20× single-thread speedup over Fusion Blossom. The algorithms are compatible - combining them could yield another 5-20× improvement.
6. Graph Neural Networks for QEC
QSeer (May 2025)
Source: arXiv:2505.06810
GNN for QAOA parameter prediction:
- 6-68% improvement in approximation ratio
- 5-10× convergence speedup
- Supports variable-depth circuits and weighted Max-Cut
Relevance to ruQu: Train a small GNN to predict optimal thresholds from syndrome graph structure:
pub struct ThresholdPredictor {
model: OnnxModel, // Export trained model
}
impl ThresholdPredictor {
pub fn predict(&self, graph_embedding: &[f32]) -> GateThresholds {
// Use learned model for threshold prediction
}
}
Implementation Priority Matrix
| Enhancement | Impact | Effort | Priority |
|---|---|---|---|
| Hierarchical tree-grid topology | High | High | P2 |
| Drift detection in adaptive.rs | High | Medium | P1 |
| Recursive early-exit processing | Medium | Low | P1 |
| Bandit-based threshold exploration | Medium | Medium | P2 |
| FPGA synthesis of min-cut | Very High | Very High | P3 |
| GNN threshold predictor | Medium | High | P3 |
| Streaming Fusion mode | High | Low | P1 |
Immediate Next Steps
- Drift Detection: Add window-based drift estimation to
adaptive.rs - Early-Exit Depth: Implement confidence-based early exit in tile processing
- Streaming Decoder: Enable Fusion Blossom streaming mode for <1ms latency
- Parallel Fusion: Configure parallel fusion on 64+ core systems
References
- DECONET: arxiv.org/abs/2504.11805
- Google Below-Threshold: nature.com/articles/s41586-024-08449-y
- Riverlane CC Decoder: riverlane.com
- Adaptive Syndrome Extraction: doi.org/10.1103/ps3r-wf84
- Multi-Agent RL QEC: arxiv.org/pdf/2509.03974
- Drift Estimation: arxiv.org/html/2511.09491
- Mixture-of-Depths: arxiv.org/html/2404.02258v1
- Mixture-of-Recursions: arxiv.org/html/2507.10524v1
- Fusion Blossom: arxiv.org/abs/2305.08307
- QSeer GNN: arxiv.org/abs/2505.06810
- QASBA FPGA: dl.acm.org/doi/10.1145/3723168