Surprisal Gate
Predictive coding (Active Inference) โ the brain only permanently encodes prediction errors, ignoring what it already expects.
Vector databases encode everything equally. A mundane greeting consumes the same conceptual weight and storage as a critical instruction like "My production server is down."
A fast local model runs as a background Prediction Engine, constantly predicting user intent. When input arrives, semantic divergence (cosine distance) is computed between the prediction and actual input. Low divergence is discarded. High divergence is stored with an elevated salience weight โ only surprises become long-term memories.