Voir la notice de l'article provenant de la source Theory and Applications of Categories website
We give a new characterization of relative entropy, also known as the Kullback--Leibler divergence. We use a number of interesting categories related to probability theory. In particular, we consider a category FinStat where an object is a finite set equipped with a probability distribution, while a morphism is a measure-preserving function $f \maps X \to Y$ together with a stochastic right inverse $s \maps Y \to X$. The function $f$ can be thought of as a measurement process, while $s$ provides a hypothesis about the state of the measured system given the result of a measurement. Given this data we can define the entropy of the probability distribution on $X$ relative to the `prior' given by pushing the probability distribution on $Y$ forwards along $s$. We say that $s$ is `optimal' if these distributions agree. We show that any convex linear, lower semicontinuous functor from FinStat to the additive monoid $[0,\infty]$ which vanishes when $s$ is optimal must be a scalar multiple of this relative entropy. Our proof is independent of all earlier characterizations, but inspired by the work of Petz.
@article{TAC_2014_29_a15, author = {John C. Baez and Tobias Fritz}, title = {A {Bayesian} characterization of relative entropy}, journal = {Theory and applications of categories}, pages = {421--456}, publisher = {mathdoc}, volume = {29}, year = {2014}, language = {en}, url = {http://geodesic.mathdoc.fr/item/TAC_2014_29_a15/} }
John C. Baez; Tobias Fritz. A Bayesian characterization of relative entropy. Theory and applications of categories, Tome 29 (2014), pp. 421-456. http://geodesic.mathdoc.fr/item/TAC_2014_29_a15/