Pinsker-Marginal Bound for f-Divergence
- Pinsker-Marginal Bound is a set of inequalities that offers an optimal upper bound on f-divergences, notably KL divergence, by incorporating total-variation distance and explicit density-ratio constraints.
- It is established using convex analysis and extremal three-atom distributions, which provide tight controls in regimes of high concentration or bounded densities.
- The bound generalizes classical and reverse Pinsker inequalities, enabling precise computations in applications such as hypothesis testing, quantization, and distribution synthesis.
The Pinsker-Marginal Bound is an optimal upper bound on -divergences, most notably Kullback–Leibler (KL) divergence, in terms of the total-variation (TV) distance under additional pointwise constraints on the likelihood ratio. This family of inequalities generalizes and sharpens the classical Pinsker and “reverse Pinsker” bounds, incorporating information about the essential infimum and supremum of the Radon–Nikodým derivative between the pair of probability measures, leading to tight controls in high-concentration or bounded-density regimes. The Pinsker-Marginal bounds also facilitate the computation of worst-case -divergence for given TV distance and density envelope, with explicit tightness and extremal constructions. They are central in information theory, hypothesis testing, distribution synthesis, and quantization theory.
1. Formal Definition and Principal Results
The general Pinsker-Marginal inequality, as established in Binette (Binette, 2018), upper bounds an arbitrary -divergence between and constrained by TV distance and essential bounds on the Radon–Nikodým derivative: $\mathcal{A}(\delta,m,M) = \Bigl\{(P,Q): P\ll Q,\; \essinf\tfrac{dP}{dQ}=m,\; \esssup\tfrac{dP}{dQ}=M,\; TV(P,Q)=\delta\Bigr\}$ Let be convex with , and define . The optimal bound is
If or , both sides are set to zero. This covers the full parametrized set of admissible pairs, and achieves equality on simple three-atom distributions.
In the case (KL divergence), it yields
with , .
2. Variational Method and Tightness
The underpinning argument is based on convex analysis: Jensen’s inequality applied to the likelihood ratio on . The decomposition partitions the domain into and , applies convexity on each, and uses the constraint on TV distance and mean values to achieve sharp two-point upper bounds. The extremal distributions that attain these bounds are three-point measures: with , , and determined to achieve . Direct computation verifies attainment of the supremum.
Global (unconstrained ) Pinsker-Marginal bounds are maximized at the endpoint . At this point, the supremum coincides with the enveloped Jensen functional (Simic 2009), with the Pinsker-Marginal bound strictly sharper (Binette, 2018).
3. Classical Reverse Pinsker and Pinsker Relations
The Pinsker-Marginal bounds generalize the best-known reverse Pinsker inequalities. For unconstrained measures,
is classical Pinsker. Reverse Pinsker inequalities provide a (typically non-sharp) upper bound on in terms of . In the binary case, refined reverse Pinsker inequalities scale as , with controlled by minimal values of (Sason, 2015).
Pinsker-Marginal bounds sharpen these classical relationships by adding density-ratio constraints; for finite alphabets, they yield quadratic improvements over prior reverse Pinsker results (such as those of Verdú and Csiszár–Talata) in both coefficient and asymptotic behavior (Sason, 2015, Binette, 2018). For (TV), they reproduce the identity .
4. Extensions: Rényi Divergences and -Divergence
Pinsker-Marginal inequalities specialize cleanly to Rényi divergences of arbitrary order (including ), with explicit functions of , , , and minimal atom probabilities. For -divergence (), the Pinsker-Marginal bound becomes
outperforming prior two-sided bounds and providing sharp characterization of outlier-weighted divergences.
Sason (Sason, 2015) presents further upper bounds for general -divergence and KL that account for infimum and supremum of the density ratio, minimizing the gap in worst-case scenarios, with precise scaling for finite-alphabet and small- regimes.
5. Special Cases and Applications
A table summarizes key specializations of the Pinsker-Marginal inequality:
| Pinsker–Marginal Bound | Interpretation | |
|---|---|---|
| KL divergence—optimal in bounded-likelihood | ||
| TV—identity | ||
| —linear in density bounds |
Applications include:
- Quantization and source coding, where Pinsker-Marginal bounds give KL–TV tradeoffs for design of codes under bounded density hypotheses.
- Statistical estimation and hypothesis testing, for non-asymptotic control over error probabilities and minimax risk analysis.
- Distribution synthesis and quantization error control, particularly in Bayesian nonparametric settings where worst-case densities are bounded (Binette, 2018).
6. Comparison to Related Work and Literature
Pinsker-Marginal bounds strictly improve and generalize results due to Simic (2009) for global bounds, and specialize to limit cases articulated by Verdú (2014), Sason–Verdú (2016), and Vajda (1972) by taking or . Binette (Binette, 2018) establishes optimality in the context of -divergence, encompassing both unbounded and bounded regimes, as well as multi-atom constructions for tightness.
A comparison with reverse Pinsker inequalities as in Sason (Sason, 2015) reveals that the Pinsker-Marginal bound is exact when the likelihood ratio takes at most two values, and that for finite alphabets the improvement in coefficient for the reverse Pinsker is a factor of two or more in the quadratic regime. The extension to Rényi divergence orders clarifies the regime in which the TV–KL or TV–Rényi relationships are linear versus quadratic in the distance parameter.
7. Illustrative Extremal Construction and Interpretive Remarks
The extremal example central to the Pinsker-Marginal theory uses a three-atom distribution where probability mass is concentrated optimally in two “worst-case” intervals, realizing saturation of Jensen's inequality. These extremal constructions model worst-case (maximal-divergence) situations for quantization and channel synthesis scenarios. The tightness and universality of the Pinsker-Marginal bounds underscore their role as canonical sharp inequalities in both theoretical analysis and practical metric control across information-theoretic disciplines (Binette, 2018).
References
- "A Note on Reverse Pinsker Inequalities" (Binette, 2018)
- "On Reverse Pinsker Inequalities" (Sason, 2015)