git-subtree-dir: vendor/ruvector git-subtree-split: b64c21726f2bb37286d9ee36a7869fef60cc6900
1.3 KiB
1.3 KiB
ADR-CE-015: Adapt Without Losing Control
Status: Accepted Date: 2026-01-22 Parent: ADR-014 Coherence Engine Architecture
Context
Static systems become stale. Adaptive systems can drift or be gamed. The coherence engine needs to:
- Learn from experience
- Improve over time
- Maintain governance and control
Decision
Adapt without losing control - persistent tracking enables learning within governance.
Adaptation mechanisms:
- Threshold autotuning: SONA proposes, humans approve
- Learned restriction maps: GNN training with EWC++ (no forgetting)
- ReasoningBank patterns: Store successful approaches
- Deterministic replay: Verify adaptations against history
Control mechanisms:
- Policy bundles require signatures: No unauthorized changes
- Witness chain is immutable: Cannot hide past decisions
- Lineage tracking: Every adaptation has provenance
- Rollback support: Can revert to previous policy
Consequences
Benefits
- System improves with experience
- Governance maintained throughout
- Can audit all adaptations
Risks
- Adaptation speed limited by approval process
- Learning quality depends on trace quality
References
- ADR-014: Coherence Engine Architecture
- ADR-CE-007: Threshold Autotuning