Files
wifi-densepose/rust-port/wifi-densepose-rs/crates/wifi-densepose-nn/Cargo.toml
rUv 9bbe95648c feat: ADR-024 Contrastive CSI Embedding Model — all 7 phases (#52)
Full implementation of Project AETHER — Contrastive CSI Embedding Model.

## Phases Delivered
1. ProjectionHead (64→128→128) + L2 normalization
2. CsiAugmenter (5 physically-motivated augmentations)
3. InfoNCE contrastive loss + SimCLR pretraining
4. FingerprintIndex (4 index types: env, activity, temporal, person)
5. RVF SEG_EMBED (0x0C) + CLI integration
6. Cross-modal alignment (PoseEncoder + InfoNCE)
7. Deep RuVector: MicroLoRA, EWC++, drift detection, hard-negative mining, SEG_LORA

## Stats
- 276 tests passing (191 lib + 51 bin + 16 rvf + 18 vitals)
- 3,342 additions across 8 files
- Zero unsafe/unwrap/panic/todo stubs
- ~55KB INT8 model for ESP32 edge deployment

Also fixes deprecated GitHub Actions (v3→v4) and adds feat/* branch CI triggers.

Closes #50
2026-03-01 01:44:38 -05:00

62 lines
1.5 KiB
TOML

[package]
name = "wifi-densepose-nn"
version.workspace = true
edition.workspace = true
authors.workspace = true
license.workspace = true
repository.workspace = true
documentation.workspace = true
keywords = ["neural-network", "onnx", "inference", "densepose", "deep-learning"]
categories = ["science", "computer-vision"]
description = "Neural network inference for WiFi-DensePose pose estimation"
readme = "README.md"
[features]
default = ["onnx"]
onnx = ["ort"]
tch-backend = ["tch"]
candle-backend = ["candle-core", "candle-nn"]
cuda = ["onnx"]
tensorrt = ["onnx"]
all-backends = ["onnx", "tch-backend", "candle-backend"]
[dependencies]
# Core utilities
thiserror.workspace = true
anyhow.workspace = true
serde.workspace = true
serde_json.workspace = true
tracing.workspace = true
# Tensor operations
ndarray.workspace = true
num-traits.workspace = true
# ONNX Runtime (default)
ort = { workspace = true, optional = true }
# PyTorch backend (optional)
tch = { workspace = true, optional = true }
# Candle backend (optional)
candle-core = { workspace = true, optional = true }
candle-nn = { workspace = true, optional = true }
# Async runtime
tokio = { workspace = true, features = ["sync", "rt"] }
# Additional utilities
parking_lot = "0.12"
once_cell = "1.19"
memmap2 = "0.9"
[dev-dependencies]
criterion.workspace = true
proptest.workspace = true
tokio = { workspace = true, features = ["rt-multi-thread", "macros"] }
tempfile = "3.10"
[[bench]]
name = "inference_bench"
harness = false