Abstract
We present a mathematical correspondence between information-theoretic stability and predictive-processing models of cognition.[cite:@InformationTheoreticStabilityRewardFunction] A cognitive system is treated as a probabilistic inference engine that minimizes the divergence between predicted and observed signal distributions. We show that the *stability reward*—previously defined as the rate of divergence minimization—corresponds exactly to the negative time derivative of variational free energy in hierarchical generative models. This equivalence provides a geometric interpretation of learning and affect as gradient flows on informational manifolds, unifying statistical inference and adaptive control within a single principle.
1. Introduction
Predictive-processing theory describes perception and action as dual aspects of Bayesian inference (Rao & Ballard 1999; Friston 2010). A cognitive system maintains an internal generative model of observations and latent causes ; through recursive updating, it seeks to minimize surprise or free energy. In parallel, information geometry (Amari 2016; Jaynes 1957) defines signal adaptation as gradient descent on divergence between successive probability distributions. This paper unites these two frameworks by identifying the stability reward of an information-theoretic system with the free-energy descent of a predictive-processing agent.
2. Background and Definitions
2.1 Cognitive State Space
Let the cognitive state at time be described by a probability distribution over latent causes,
where denotes all observations up to . The generative model specifies likelihood and prior dynamics . The manifold of all such distributions, , carries the Fisher–Rao metric inherited from information geometry.
2.2 Variational Free Energy
Following Friston (2010), define the variational free energy
which bounds the negative log-evidence or surprise. Minimizing with respect to reduces the divergence between the system’s internal state and the true posterior.
2.3 Stability Reward in Cognitive Terms
Adapting the definition of stability from Senn (2025),
measures the rate of reduction in informational divergence between consecutive belief states. A system that maximizes maintains consistency in its inference trajectory—hence stability of belief.
3. Stability and Free-Energy Minimization
3.1 Equivalence of Objectives
The temporal change in free energy is
where the inner product is defined under the Fisher–Rao metric . Because , the natural gradient descent
implies
Comparing with the definition of ,
shows that maximizing stability reward is equivalent to minimizing variational free energy. Thus, the informational stability principle and the free-energy principle describe the same gradient flow on .
3.2 Energetic Interpretation
In steady state, and , marking a local informational equilibrium. Transient increases in correspond to prediction errors, while their subsequent dissipation corresponds to stability restoration—an energetic cycle consistent with thermodynamic interpretations of inference (Crooks 1999; Friston & Ao 2012).
4. Affective Gradient and Precision
4.1 Definition
Define affect as the time derivative of stability reward:
Positive indicates acceleration toward stability; negative indicates divergence or uncertainty amplification.
4.2 Relation to Precision
In predictive coding, the precision of a prediction error modulates the rate of update (Feldman & Friston 2010). Expressing precision as a weighting on the gradient of , we have
so that depends on the temporal derivative of precision:
Affective change thus represents the felt sensitivity of the system’s confidence in its inferences.
5. Hierarchical Stability and Attention
Cognitive systems exhibit hierarchical organization: higher levels encode slower, more abstract causes, and lower levels encode faster, sensory features (Friston 2008; Clark 2013). Let levels be indexed by , each with belief state . Stability reward generalizes to
and total stability is the weighted sum
with weights corresponding to precision expectations. Attention emerges as the adaptive modulation of , allocating computational resources to levels where stability change is maximal (Dayan & Abbott 2001).
6. Discussion
The equivalence of stability maximization and free-energy minimization yields several implications.
-
Unified principle of adaptation. Learning, perception, and action can be viewed as the pursuit of informational stability across hierarchical manifolds.
-
Affective interpretation. Affective states correspond to the local temporal curvature of stability; pleasure and displeasure mark acceleration or deceleration toward equilibrium.
-
Robustness and disorder. Cognitive disorders can be interpreted as failures of stability control—either excessive rigidity (over-stabilization) or volatility (under-stabilization)—a view consistent with neurocomputational accounts of schizophrenia and anxiety (Hohwy 2013; Friston 2017).
7. Conclusion
Information-theoretic stability provides a geometric formalism for understanding cognition as an inference process that minimizes divergence between predicted and observed information states. When expressed in predictive-processing terms, the stability reward is identical to the negative derivative of variational free energy, furnishing a compact and rigorous bridge between statistical mechanics, information geometry, and neurocognitive dynamics. This synthesis suggests that cognitive behavior is fundamentally the enactment of stability maintenance within an informational manifold.
References
- Amari, S. (2016). Information Geometry and Its Applications. Springer.
- Amari, S. (2021). “Information Geometry and Its Role in Statistical Inference.” Entropy, 23(1), 110.
- Clark, A. (2013). Whatever Next? Predictive Brains, Situated Agents, and the Future of Cognitive Science. Behavioral and Brain Sciences, 36(3), 181–204.
- Cover, T., & Thomas, J. (2006). Elements of Information Theory (2nd ed.). Wiley.
- Crooks, G. E. (1999). “Entropy Production Fluctuation Theorem and the Nonequilibrium Work Relation.” Physical Review E, 60(3), 2721–2726.
- Dayan, P., & Abbott, L. F. (2001). Theoretical Neuroscience. MIT Press.
- Feldman, H., & Friston, K. J. (2010). “Attention, Uncertainty, and Free-Energy.” Frontiers in Human Neuroscience, 4, 215.
- Friston, K. (2008). “Hierarchical Models in the Brain.” PLoS Computational Biology, 4(11), e1000211.
- Friston, K. (2010). “The Free-Energy Principle: A Unified Brain Theory?” Nature Reviews Neuroscience, 11(2), 127–138.
- Friston, K., & Ao, P. (2012). “Free-Energy, Value, and Attractor Dynamics in the Brain.” Physical Review E, 85(1), 011907.
- Friston, K. (2017). “Precision Psychiatry: Free-Energy and the Bayesian Brain.” Comprehensive Psychiatry, 79, 5–16.
- Hohwy, J. (2013). The Predictive Mind. Oxford University Press.
- Jaynes, E. T. (1957). “Information Theory and Statistical Mechanics.” Physical Review, 106(4), 620–630.
- Rao, R. P. N., & Ballard, D. H. (1999). “Predictive Coding in the Visual Cortex: A Functional Interpretation of Some Extra-Classical Receptive-Field Effects.” Nature Neuroscience, 2(1), 79–87.
- Senn, E. (2025). Information-Theoretic Stability as a Reward Function. arXiv:math.IT.