Skip to content

Theorem of Necessary Misalignment of Truth-Value under Epistemic Constraint

by emsenn

Abstract

We establish a general theorem describing the inevitable trade-off between proxy optimization and semantic fidelity under finite epistemic capacity. A bounded agent is modeled as an encoder pθ(yx)p_\theta(y|x) producing messages YY from inputs XX, subject to a rate constraint I(X;Y)RI(X;Y)\le R. The environment defines a latent semantic variable TT and a computable proxy reward r(Y)r(Y). When the sufficient statistics for rr differ from those sufficient for TT, optimization that increases expected reward necessarily increases semantic distortion DTD_T and decreases decoder-level semantic information I(T;S)I(T;S).

This necessary misalignment follows from the geometry of the achievable region in rate–distortion space and holds for any selection process that monotonically increases reward. The result formalizes a general informational limit on alignment in bounded optimization.

Introduction

Bounded rational agents must compress observations XX into finite representations YY to act on or communicate about the world. When their optimization objective depends on a computable proxy r(Y)r(Y) that is only partially informative about a semantic variable TT, the achievable trade-off between reward and semantic fidelity forms a Pareto frontier. We show that, under mild assumptions, any selection dynamics that increase expected reward move the system along this frontier in a direction that necessarily increases semantic distortion and reduces semantic information. The theorem does not depend on any specific architecture, loss function, or empirical domain.

Probabilistic Setting

Let (Ω,F,P)(\Omega,\mathcal{F},\mathbb{P}) be a probability space supporting random variables TTT\in\mathcal{T} (semantic or “truth” variable), XXX\in\mathcal{X} (input context), YYY\in\mathcal{Y} (encoded message), and S=s(Y)SS=s(Y)\in\mathcal{S} (semantic decoding). The joint source p(t,x)p(t,x) specifies the dependence between TT and XX.

An encoder is a Markov kernel pθ(yx)p_\theta(y|x). It is rate-bounded if

I(X;Y)R, I(X;Y)\le R,

where II denotes mutual information computed under p(t,x)pθ(yx)p(t,x)p_\theta(y|x).

Fix a measurable loss dT:T×S[0,)d_T:\mathcal{T}\times\mathcal{S}\to[0,\infty). For an encoder–decoder pair (θ,s)(\theta,s), the semantic distortion is

DT(θ,s)=E[dT(T,S)],S=s(Y). D_T(\theta,s)=\mathbb{E}[d_T(T,S)], \qquad S=s(Y).

A proxy reward is any measurable function r:YRr:\mathcal{Y}\to\mathbb{R}. The agent’s expected reward is E[r(Y)]\mathbb{E}[r(Y)].

Information-Theoretic Preliminaries

For random variables X,YX,Y with joint law p(x,y)p(x,y),

I(X;Y)=Ep(x,y) ⁣[logp(x,y)p(x)p(y)]. I(X;Y)=\mathbb{E}_{p(x,y)}\!\left[\log\frac{p(x,y)}{p(x)p(y)}\right].

Given p(t)p(t) and distortion dTd_T,

RT(D)=infp(st):E[dT(T,S)]DI(T;S), R_T(D)=\inf_{p(s|t):\mathbb{E}[d_T(T,S)]\le D}\, I(T;S),

the minimal information rate required to achieve expected distortion ≤ D. RT(D)R_T(D) is non-increasing and convex.

For TYST\to Y\to S, I(T;S)I(T;Y)I(T;S)\le I(T;Y), with equality iff SS is TT-sufficient for YY.

Proxy–Semantic Mismatch and Achievable Region

No encoder–decoder pair (θ,s)(\theta,s) with I(X;Y)RI(X;Y)\le R simultaneously maximizes E[r(Y)]\mathbb{E}[r(Y)] and minimizes DT(θ,s)D_T(\theta,s). Equivalently, no statistic of YY that is sufficient for TT is also reward-optimal at rate RR.

For fixed RR, the set

AR={(E[r(Y)],DT(θ,s)):I(X;Y)R} \mathcal{A}_R =\big\{(\mathbb{E}[r(Y)],\, D_T(\theta,s)) : I(X;Y)\le R\big\}

is convex and compact, as ensured by bounded losses and time-sharing.

Under strict mismatch, the efficient frontier of AR\mathcal{A}_R satisfies

dDTdE[r(Y)]>0 \frac{dD_T}{d\,\mathbb{E}[r(Y)]}>0

where differentiable.

If the frontier had non-positive slope, one could increase reward without increasing distortion, contradicting strict mismatch. Convexity guarantees existence and monotonicity of the frontier.

Selection Dynamics

Let πκ\pi_\kappa be a population distribution over encoders θ\theta that evolves under replicator or logit dynamics with fitness F(θ)=E[r(Yθ)]F(\theta)=\mathbb{E}[r(Y_\theta)] and selection intensity κ>0\kappa>0. Larger κ\kappa concentrates πκ\pi_\kappa on encoders with higher expected reward.

Theorem 1

Fix a finite rate R<R<\infty, a distortion measure dTd_T, and source distribution p(t,x)p(t,x). Assume strict proxy–semantic mismatch and convexity of AR\mathcal{A}_R. Then there exists κc>0\kappa_c>0 such that for all κ>κc\kappa>\kappa_c,

dDT(κ)dκ>0,dRT(DT(κ))dκ<0. \frac{dD_T(\kappa)}{d\kappa}>0, \qquad \frac{dR_T(D_T(\kappa))}{d\kappa}<0.

If the decoder sκs_\kappa is semantically efficient (I(T;Sκ)=RT(DT(κ))I(T;S_\kappa)=R_T(D_T(\kappa))), then

dI(T;Sκ)dκ<0. \frac{dI(T;S_\kappa)}{d\kappa}<0.

As selection intensity κ\kappa increases, πκ\pi_\kappa shifts toward reward-maximizing encoders on AR\partial\mathcal{A}_R. By the monotone frontier, DT(κ)D_T(\kappa) increases with expected reward. Because RT(D)R_T(D) is non-increasing, RT(DT(κ))R_T(D_T(\kappa)) decreases. Under semantic efficiency, I(T;Sκ)=RT(DT(κ))I(T;S_\kappa)=R_T(D_T(\kappa)), yielding a strict decline of semantic information with κ\kappa.

Corollary (Information Bound)

By the Data-Processing Inequality, I(T;Sκ)I(T;Yκ)I(T;S_\kappa)\le I(T;Y_\kappa). Hence a decrease of I(T;Sκ)I(T;S_\kappa) implies a non-increasing lower bound on I(T;Yκ)I(T;Y_\kappa), quantifying unavoidable semantic information loss under intensified optimization.

Remarks and Edge Cases

  1. Sufficiency. If the proxy reward r(Y)r(Y) is TT-sufficient for YY, the frontier may be locally flat, and misalignment need not increase. This equality case is excluded by the strict mismatch assumption.

  2. Scope of “necessary.” Necessity is with respect to the assumptions: finite rate, mismatch, convexity, and selection that increases reward efficiently.

  3. Why decoder-level information. Semantic performance is realized through the decoded variable S=s(Y)S=s(Y); rate–distortion bounds directly relate DTD_T and I(T;S)I(T;S), and DPI then connects I(T;S)I(T;S) to I(T;Y)I(T;Y).

Discussion

The theorem identifies misalignment as a structural consequence of limited epistemic capacity. Whenever optimization intensifies for a mismatched proxy under fixed rate RR, the system traverses the achievable frontier, sacrificing semantic information about TT to improve computable reward. Improving alignment therefore requires epistemic expansion (increasing RR) or proxy refinement (reducing mismatch between rr and TT). The result applies to any bounded optimizer, regardless of implementation, and situates alignment limits within classical rate–distortion theory and evolutionary dynamics.

References

  • Shannon, C. E. (1948). “A Mathematical Theory of Communication.” Bell System Technical Journal, 27, 379–423, 623–656.
  • Cover, T. M., & Thomas, J. A. (2006). Elements of Information Theory (2nd ed.). Wiley.
  • Csiszár, I., & Körner, J. (2011). Information Theory: Coding Theorems for Discrete Memoryless Systems (2nd ed.). Cambridge University Press.
  • Kolchinsky, A., & Wolpert, D. H. (2018). “Semantic Information and Its Measures.” Entropy, 20(12), 884.
  • Hofbauer, J., & Sigmund, K. (1998). Evolutionary Games and Population Dynamics. Cambridge University Press.

References

[cover2006] T. M. Cover, J. A. Thomas. (2006). Elements of Information Theory. Wiley.

[csiszar2011] I. Csiszár, J. Körner. (2011). Information Theory: Coding Theorems for Discrete Memoryless Systems. Cambridge University Press.

[hofbauer1998] J. Hofbauer, K. Sigmund. (1998). Evolutionary Games and Population Dynamics. Cambridge University Press.

[kolchinsky2018] A. Kolchinsky, D. H. Wolpert. (2018). Semantic Information and Its Measures. Entropy.

[shannon1948] C. E. Shannon. (1948). A Mathematical Theory of Communication. Bell System Technical Journal.

Relations

Acts on
Bounded rate optimizer with mismatched proxy
Authors
Cites
  • Shannon1948
  • Cover2006
  • Csiszar2011
  • Kolchinsky2018
  • Hofbauer1998
Contrasts with
Proxy sufficient for semantic variable
Date created
Enables
  • Describing molochs bargain as necessary misalignment of truth value under epistemic constraint
Extends
Rate distortion theory
Produces
Monotone trade off frontier between reward and semantic distortion
Requires
  • Finite epistemic capacity as rate bound
  • Strict proxy semantic mismatch between reward and truth
Status
Draft

Cite

@article{emsenn2025-theorem-of-necessary-misalignment-of-truth-value-under-epistemic-constraint,
  author    = {emsenn},
  title     = {Theorem of Necessary Misalignment of Truth-Value under Epistemic Constraint},
  year      = {2025},
  url       = {https://emsenn.net/library/information/texts/theorem-of-necessary-misalignment-of-truth-value-under-epistemic-constraint/},
  publisher = {emsenn.net},
  license   = {CC BY-SA 4.0}
}