Back-flow of distinguishability is a model-agnostic witness of training memory in neural networks, quantifying non-Markovianity by measuring the increased distinguishability of model outcomes after sequential interventions. It provides a principled diagnostic for understanding how optimizer and data states influence training dynamics.
Back-flow of distinguishability is a new way to measure how much "memory" a neural network's training process has, showing that training isn't always a simple step-by-step process. It helps researchers understand how past training decisions, like optimizer settings or data batches, continue to influence future learning.
BFD, training memory witness, non-Markovianity witness
Was this definition helpful?