You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
docs: Enrich Models & Training section with RuVector repo links
- ToC: Add ruvector GitHub link and integration point count
- RVF Container: Add deployment targets table (ESP32 0.7MB to server
50MB), link to rvf crate family on GitHub
- Training: Add RuVector column to pipeline table showing which crate
powers each phase, add SONA component breakdown table, link arXiv
- RuVector Crates: Split into 5 directly-used (with integration
points mapped to exact .rs files) and 6 additional vendored, add
crates.io and GitHub source links for all 11
Co-Authored-By: claude-flow <[email protected]>
The neural pipeline uses a graph transformer with cross-attention to map CSI feature matrices to 17 COCO body keypoints and DensePose UV coordinates. Models are packaged as single-file `.rvf` containers with progressive loading (Layer A instant, Layer B warm, Layer C full). SONA (Self-Optimizing Neural Architecture) enables continuous on-device adaptation via micro-LoRA + EWC++ without catastrophic forgetting.
280
+
The neural pipeline uses a graph transformer with cross-attention to map CSI feature matrices to 17 COCO body keypoints and DensePose UV coordinates. Models are packaged as single-file `.rvf` containers with progressive loading (Layer A instant, Layer B warm, Layer C full). SONA (Self-Optimizing Neural Architecture) enables continuous on-device adaptation via micro-LoRA + EWC++ without catastrophic forgetting. Signal processing is powered by 5 [RuVector](https://github.com/ruvnet/ruvector) crates (v2.0.4) with 7 integration points across the Rust workspace, plus 6 additional vendored crates for inference and graph intelligence.
281
281
282
282
| Section | Description | Docs |
283
283
|---------|-------------|------|
284
284
|[RVF Model Container](#rvf-model-container)| Binary packaging with Ed25519 signing, progressive 3-layer loading, SIMD quantization |[ADR-023](docs/adr/ADR-023-trained-densepose-model-ruvector-pipeline.md)|
@@ -522,7 +522,7 @@ See [ADR-014](docs/adr/ADR-014-sota-signal-processing.md) for full mathematical
522
522
<details>
523
523
<summary><aid="rvf-model-container"></a><strong>📦 RVF Model Container</strong> — Single-file deployment with progressive loading</summary>
524
524
525
-
The RuVector Format (RVF) packages an entire trained model — weights, HNSW indexes, quantization codebooks, SONA adaptation deltas, and WASM inference runtime — into a single self-contained binary file. No external dependencies are needed at deployment time.
525
+
The [RuVector Format (RVF)](https://github.com/ruvnet/ruvector/tree/main/crates/rvf) packages an entire trained model — weights, HNSW indexes, quantization codebooks, SONA adaptation deltas, and WASM inference runtime — into a single self-contained binary file. No external dependencies are needed at deployment time.
526
526
527
527
**Container structure:**
528
528
@@ -548,14 +548,23 @@ The RuVector Format (RVF) packages an entire trained model — weights, HNSW ind
@@ -568,37 +577,44 @@ The RuVector Format (RVF) packages an entire trained model — weights, HNSW ind
568
577
docker run --rm -v $(pwd):/out ruvnet/wifi-densepose:latest --export-rvf /out/model.rvf
569
578
```
570
579
571
-
See [ADR-023](docs/adr/ADR-023-trained-densepose-model-ruvector-pipeline.md).
580
+
Built on the [rvf](https://github.com/ruvnet/ruvector/tree/main/crates/rvf) crate family (rvf-types, rvf-wire, rvf-manifest, rvf-index, rvf-quant, rvf-crypto, rvf-runtime). See [ADR-023](docs/adr/ADR-023-trained-densepose-model-ruvector-pipeline.md).
572
581
573
582
</details>
574
583
575
584
<details>
576
585
<summary><aid="training--fine-tuning"></a><strong>🧬 Training & Fine-Tuning</strong> — MM-Fi/Wi-Pose pre-training, SONA adaptation</summary>
577
586
578
-
The training pipeline implements 8 phases in pure Rust (7,832 lines, zero external ML dependencies). It trains a graph transformer with cross-attention to map CSI feature matrices to 17 COCO body keypoints and DensePose UV coordinates — following the approach from the CMU "DensePose From WiFi" paper (arXiv:2301.00250).
587
+
The training pipeline implements 8 phases in pure Rust (7,832 lines, zero external ML dependencies). It trains a graph transformer with cross-attention to map CSI feature matrices to 17 COCO body keypoints and DensePose UV coordinates — following the approach from the CMU "DensePose From WiFi" paper ([arXiv:2301.00250](https://arxiv.org/abs/2301.00250)). RuVector crates provide the core building blocks: [ruvector-attention](https://github.com/ruvnet/ruvector/tree/main/crates/ruvector-attention) for cross-attention layers, [ruvector-mincut](https://github.com/ruvnet/ruvector/tree/main/crates/ruvector-mincut) for multi-person matching, and [ruvector-temporal-tensor](https://github.com/ruvnet/ruvector/tree/main/crates/ruvector-temporal-tensor) for CSI buffer compression.
**Best-epoch snapshotting**: the trainer saves the best validation loss weights and restores them before checkpoint/export — prevents exporting overfit final-epoch parameters.
599
+
| Phase | Module | What It Does | RuVector Crate |
**SONA (Self-Optimizing Neural Architecture)** — the continuous adaptation system:
611
+
612
+
| Component | What It Does | Why It Matters |
613
+
|-----------|-------------|----------------|
614
+
|**Micro-LoRA (rank-4)**| Trains small A×B weight deltas instead of full weights | 100x fewer parameters to update → runs on ESP32 |
615
+
|**EWC++ (Fisher matrix)**| Penalizes changes to important weights from previous environments | Prevents catastrophic forgetting when moving between rooms |
616
+
|**EnvironmentDetector**| Monitors CSI feature drift with 3-sigma threshold | Auto-triggers adaptation when the model is moved to a new space |
617
+
|**Best-epoch snapshot**| Saves best validation loss weights, restores before export | Prevents shipping overfit final-epoch parameters |
The full RuVector ecosystem includes 90+ crates. See [github.com/ruvnet/ruvector](https://github.com/ruvnet/ruvector) for the complete library, and [`vendor/ruvector/`](vendor/ruvector/) for the vendored source in this project.
0 commit comments