git-subtree-dir: vendor/ruvector git-subtree-split: b64c21726f2bb37286d9ee36a7869fef60cc6900
31 lines
1.1 KiB
Rust
31 lines
1.1 KiB
Rust
//! Information Geometry
|
|
//!
|
|
//! Information geometry treats probability distributions as points on a curved manifold,
|
|
//! enabling geometry-aware optimization and analysis.
|
|
//!
|
|
//! ## Core Concepts
|
|
//!
|
|
//! - **Fisher Information Matrix (FIM)**: Measures curvature of probability space
|
|
//! - **Natural Gradient**: Gradient descent that respects the manifold geometry
|
|
//! - **K-FAC**: Kronecker-factored approximation for efficient natural gradient
|
|
//!
|
|
//! ## Benefits for Vector Search
|
|
//!
|
|
//! 1. **Faster Index Optimization**: 3-5x fewer iterations vs Adam
|
|
//! 2. **Better Generalization**: Follows geodesics in parameter space
|
|
//! 3. **Stable Continual Learning**: Information-aware regularization
|
|
//!
|
|
//! ## References
|
|
//!
|
|
//! - Amari & Nagaoka (2000): Methods of Information Geometry
|
|
//! - Martens & Grosse (2015): Optimizing Neural Networks with K-FAC
|
|
//! - Pascanu & Bengio (2013): Natural Gradient Works Efficiently in Learning
|
|
|
|
mod fisher;
|
|
mod kfac;
|
|
mod natural_gradient;
|
|
|
|
pub use fisher::FisherInformation;
|
|
pub use kfac::KFACApproximation;
|
|
pub use natural_gradient::NaturalGradient;
|