Petz–Augustin Information Overview
- Petz–Augustin Information is a generalization of mutual information defined through a variational problem involving Rényi divergence, extending to both classical and quantum contexts.
- It features a unique minimizer, the Augustin mean, which is computed via fixed-point and iterative methods that converge geometrically to optimize channel coding error exponents.
- Its analytic structure—with continuity, concavity, and real-analytic behavior—underpins strong-converse results and efficient error exponent analysis in information theory.
The Petz–Augustin information, also referred to as Augustin information or Augustin–Csiszár mutual information in classical contexts, is a generalization of the classical mutual information parametrized by a Rényi order . It is defined as a solution to a variational problem involving the expected Rényi divergence between families of output distributions (or states in the quantum context) and an auxiliary measure or density. This quantity arises naturally in error exponent analysis, optimal channel coding, and as the core of several generalizations of information-theoretic quantities in both classical and quantum settings.
1. Definitions and Formulations
Let denote a channel with finite input and output alphabets and a probability mass function on . The order- Petz–Augustin information is defined by the infimum: where is the order- Rényi divergence. In the quantum setting, replacing with density matrices , with a density operator , and the sum with the trace, this generalizes to
For , this reduces to the usual mutual information . The unique minimizer in the infimum, when it exists, is called the Augustin mean or (Nakiboglu, 2018, Cheng et al., 2021).
2. Variational Representations and Properties
Several equivalent variational representations exist:
- Classical case: .
- Quantum case: (You et al., 2021, Chu et al., 10 Jan 2026, Chu et al., 10 Feb 2025).
Key properties:
- The Augustin mean is the unique minimizer and also the unique fixed-point of a nonlinear operator ("Augustin operator") defined as an average of order-α tilted versions of with respect to (Nakiboglu, 2018, Cheng et al., 2021).
- The mapping is uniformly equicontinuous and jointly continuous in , and concave in for (Cheng et al., 2018).
- The function is real-analytic in for fixed and (Nakiboglu, 2018).
- The scaled auxiliary function , with , is concave in and jointly continuous in .
3. Fixed-Point and Iterative Computation Methods
3.1 Classical and Quantum Fixed-Point Iteration
For discrete alphabets, the minimizer satisfies the fixed-point equation: This iteration converges to the unique minimizer for at a geometric rate in Hilbert’s projective metric (Tsai et al., 2024).
3.2 Quantum Iterative Methods
For quantum channels, algorithms follow similar fixed-point structures: with per-iteration and initialization complexities and respectively, where is the number of inputs and the state dimension (Chu et al., 10 Feb 2025). Convergence is guaranteed at a linear rate in the Thompson metric.
3.3 Alternating Optimization for Augustin–Csiszár Mutual Information
An Alternating Optimization (AO) algorithm alternates between optimizing candidate output distributions and reverse channels, generalizing the Arimoto–Blahut algorithm. Each AO step involves basic renormalizations, and convergence is global under mild conditions (Kamatsuka et al., 2024).
3.4 Mirror Descent and Polyak Step
Entropic mirror descent using negative entropy as mirror map provides convergence guarantees under weak conditions even when conventional Lipschitz or Hessian bounds do not apply. A Polyak-style adaptive step size is crucial for practical efficiency and provable convergence to -optimality, extending to quantum settings (You et al., 2021, Chu et al., 10 Jan 2026).
4. Analytic Structure and Error Exponents
Analytic results show that and its scaled auxiliary function possess continuity and concavity properties in the parameter and prior. These properties are instrumental for establishing minimax identities, entropic dualities, and Fenchel duality principles in error-exponent analysis:
- The strong-converse exponent in quantum channel coding obeys Sion’s minimax identity, permitting it to be attained by constant-composition codes (Cheng et al., 2018).
- Duality theorems connect source-channel tradeoffs with side information in operational exponents.
- The Augustin mean and center are essential for the van Erven–Harremoës lower bounds and cost-constrained capacities (Nakiboglu, 2018).
5. Convergence Rates and Complexity
The convergence rate and complexity of iterative algorithms for Petz–Augustin information are now rigorously established.
| Algorithm | Convergence Rate | Per-Iteration Cost | Applicable Regime |
|---|---|---|---|
| Classical fixed-point (Tsai et al., 2024) | in , | ||
| Quantum fixed-point (Chu et al., 10 Feb 2025) | in | ||
| Mirror descent (quantum) (You et al., 2021, Chu et al., 10 Jan 2026) | , -optimality |
The fixed-point iterations are shown to be contractive in natural projective metrics (Hilbert’s and Thompson’s), thus enabling geometric convergence and scalability to high dimensions.
6. Operational and Coding-Theoretic Relevance
Petz–Augustin information governs the optimal error exponents in channel coding beyond the Shannon capacity (coding at rates above capacity for the strong converse, or calculating sphere-packing exponents). The “Augustin capacity”—the maximized Augustin information over all input distributions—characterizes the random coding exponent and meets strong-converse benchmarks for classical-quantum channels (Cheng et al., 2018, Chu et al., 10 Jan 2026).
The unique Augustin mean serves as the operational center in information radius inequalities, cost-constrained capacities, and in dual formulations of rate-distortion and channel coding with side information. The variational and interpolation methods underpinning analytic results provide the functional foundations for modern error exponent analysis in both classical and quantum information theory.
7. Practical Implementation and Extensions
Modern first-order algorithms—including fixed-point, alternating minimization, and mirror descent with Polyak step size—are implementable using only matrix powers, traces, and elementary normalization, making them well-suited for large-scale and high-dimensional quantum channels. The methods are robust to initialization and avoid the stalling issues characteristic of singularities encountered in projected gradient approaches (You et al., 2021, Chu et al., 10 Feb 2025, Chu et al., 10 Jan 2026, Kamatsuka et al., 2024).
Extensions include resource-theoretical measures, hypothesis testing, rate-distortion functions, and other quantum-information convex programs, provided local gradient boundedness is maintained. Recent work generalizes the Blahut–Arimoto mirror-descent interpretation to the full order- regime, establishing non-asymptotic and explicit complexity bounds for error-envelope computation (Chu et al., 10 Jan 2026).
References:
- (Nakiboglu, 2018) B. Nakiboğlu, "The Augustin Capacity and Center"
- (Cheng et al., 2018) H. Cheng, M. Mosonyi, "Properties of Noncommutative Renyi and Augustin Information"
- (Cheng et al., 2021) Y. Nakiboglu, H. Cheng, "On the Existence of the Augustin Mean"
- (You et al., 2021) K. Li et al., "Minimizing Quantum Renyi Divergences via Mirror Descent with Polyak Step Size"
- (Kamatsuka et al., 2024) D. Kamatsuka, M. Kazama, T. Yoshida, "Alternating Optimization Approach for Computing α-Mutual Information and α-Capacity"
- (Tsai et al., 2024) K. Kato et al., "Linear Convergence in Hilbert's Projective Metric for Computing Augustin Information and a Rényi Information Measure"
- (Chu et al., 10 Feb 2025) D. Kamatsuka et al., "A Linearly Convergent Algorithm for Computing the Petz-Augustin Information"
- (Chu et al., 10 Jan 2026) D. Kamatsuka et al., "Algorithms for Computing the Petz-Augustin Capacity"