Papers
Topics
Authors
Recent
Search
2000 character limit reached

Lossy Common Information in Source Coding

Updated 31 January 2026
  • Lossy common information is a measure that quantifies the minimal rate of a common message in Gray–Wyner networks under prescribed distortion constraints.
  • It extends Wyner’s and Gács–Körner’s frameworks by incorporating fidelity requirements, offering insights into rate trade-offs for both discrete and Gaussian sources.
  • Practical implementations leverage explicit constructions like polar codes and learnable neural codecs, enabling distributed representation learning in modern signal processing.

Lossy common information generalizes classical information-theoretic characterizations of shared structure among correlated sources to contexts involving fidelity constraints, specifically within the Gray–Wyner network. It quantifies the minimal required rate of a common message that, combined with optimal private side-channels, enables reconstruction of the sources under prescribed distortion levels. This operationalizes the concept of “commonality” in lossy multiterminal source coding and allows for rigorous analyses of rate trade-offs in both discrete and continuous (notably, Gaussian) settings. The framework subsumes both Wyner’s and Gács–Körner’s common information as extremes, delineates plateau regions where common information is distortion-invariant, and now extends to learnable architectures for distributed representation learning in signal processing and machine learning.

1. Fundamental Definitions and Gray–Wyner Network Model

The Gray–Wyner network models two or more correlated sources X1,X2,,XNX_1, X_2, \ldots, X_N, which are compressed by an encoder into a common message %%%%1%%%% and private messages SiS_i, with rates R0R_0 and RiR_i respectively. Each decoder reconstructs its respective target using the common and private messages, subject to per-letter distortion constraints DiD_i. The achievable rate region RGW(D1,D2)\mathcal{R}_{GW}(D_1, D_2) is characterized by the existence of an auxiliary variable UU and reconstructions (X^,Y^)(\hat X, \hat Y) such that

R0I(X,Y;U),R1I(X;X^U),R2I(Y;Y^U),R_0 \geq I(X, Y; U),\quad R_1 \geq I(X; \hat X|U),\quad R_2 \geq I(Y; \hat Y|U),

with E[dX(X,X^)]D1\mathbb{E}[d_X(X, \hat X)] \leq D_1, E[dY(Y,Y^)]D2\mathbb{E}[d_Y(Y, \hat Y)] \leq D_2 (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026). The sum-rate minimization R0+R1+R2R_0 + R_1 + R_2 is fundamentally linked to the joint rate-distortion function RXY(D1,D2)R_{XY}(D_1, D_2).

Wyner’s lossy common information CW(X,Y;D1,D2)C_W(X, Y; D_1, D_2) is defined as the minimum possible common rate R0R_0 such that the total coding rate equals the joint rate-distortion bound: CW(X,Y;D1,D2)=inf{R0:R1,R2  with  (R0,R1,R2)RGW(D1,D2),  R0+R1+R2=RXY(D1,D2)}C_W(X, Y; D_1, D_2) = \inf\, \{ R_0 : \exists\, R_1, R_2\; \text{with}\; (R_0, R_1, R_2) \in \mathcal{R}_{GW}(D_1, D_2),\; R_0+R_1+R_2 = R_{XY}(D_1, D_2) \} This infimum is achieved under the Markov constraints (X,Y)(X^,Y^)U(X, Y) - (\hat X, \hat Y) - U and X^UY^\hat X \leftrightarrow U \leftrightarrow \hat Y with (X^,Y^)(\hat X, \hat Y) optimal for RXY(D1,D2)R_{XY}(D_1, D_2) (Viswanatha et al., 2014, Andrade, 6 Jul 2025, Andrade et al., 29 Jan 2026).

2. Lossy Extensions: Wyner and Gács–Körner Notions

The two dominant notions—Wyner’s and Gács–Körner’s—are extended to lossy settings via distinct operational criteria in the Gray–Wyner region:

  • Wyner’s Lossy Common Information: Corresponds to the operating point achieving minimum sum transmit rate. The single-letter characterization is:

CW(X,Y;D1,D2)=infPI(X,Y;U)C_W(X, Y; D_1, D_2) = \inf_{P}\, I(X, Y ; U)

where PP is as above (Viswanatha et al., 2014, Xu et al., 2013, Andrade et al., 29 Jan 2026).

  • Gács–Körner’s Lossy Common Information: Maximizes the extractable common rate when each source is encoded at its individual rate-distortion bound. The characterization is:

CGK(X,Y;D1,D2)=supQI(X,Y;V)C_{GK}(X, Y; D_1, D_2) = \sup_{Q}\, I(X, Y; V)

subject to P(X^X)P(\hat X | X) (resp. P(Y^Y)P(\hat Y | Y)) achieving RX(D1)R_X(D_1) (resp. RY(D2)R_Y(D_2)), and appropriate Markov constraints (Viswanatha et al., 2014, Andrade, 6 Jul 2025, Andrade et al., 29 Jan 2026).

The relationship between these quantities and the mutual information of the reconstructed variables (Z^1,Z^2)(\hat Z_1, \hat Z_2) is bounded as: CGK(X1,X2;D1,D2)I(Z^1;Z^2)CW(X1,X2;D1,D2)C_{GK}(X_1, X_2; D_1, D_2) \leq I(\hat Z_1 ; \hat Z_2) \leq C_W(X_1, X_2; D_1, D_2) with strict equality only when a "perfect common part" WW exists, separating all mutual dependence (Andrade, 6 Jul 2025).

3. Rate-Distortion Characterization and Plateaus

The solution to the optimization for CWC_W can exhibit a plateau: for distortions (D1,D2)(D_1, D_2) within a nontrivial region, the lossy common information is constant and coincides with the lossless (zero-distortion) Wyner common information. That is,

CW(X,Y;D1,D2)=CW(X,Y)C_W(X, Y; D_1, D_2) = C_W(X, Y)

so long as (D1,D2)(D_1, D_2) are sufficiently small (the so-called "Wyner plateau") (Xu et al., 2013, Shi et al., 2016, Charalambous et al., 2019). Outside this region, CW(X,Y;D1,D2)C_W(X, Y; D_1, D_2) generally increases with distortion or can be zero if the sources are effectively uncorrelated at the required resolution.

For multivariate Gaussian sources, this plateau is explicit: on Di1ρD_i \leq 1 - \rho, CW(X,Y;D1,D2)=12log1+ρ1ρC_W(X, Y; D_1, D_2) = \frac{1}{2}\log\frac{1+\rho}{1-\rho} (for correlation ρ\rho) (Xu et al., 2013, Charalambous et al., 2019, Shi et al., 2016). The explicit canonical-variable construction and weak-realization theory provide a complete parametrization of conditional-independence-inducing latent variables WW, and a closed-form expression for the minimal common rate in the quadratic-Gaussian case (Charalambous et al., 2019).

4. Operational and Structural Properties

Lossy common information precisely characterizes the boundary between efficient joint compression and source-specific refinements. The transmit rate Rt=R0+R1+R2R_t = R_0 + R_1 + R_2 is minimized at the Wyner operating point, while the receive rate Rr=2R0+R1+R2R_r = 2R_0 + R_1 + R_2 is minimized at the Gács–Körner point. The transmit–receive trade-off is continuous across the Gray–Wyner region; CWC_W and CGKC_{GK} represent its extremes (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026).

Key theorems establish:

  • Convexity and monotonicity of the common information as a function of "excess rate";
  • The operational significance of the Pangloss plane and its intersection with the Gray–Wyner region as yielding CWC_W;
  • The necessity of certain Markov factorizations among (X,Y,X^,Y^,U)(X, Y, \hat X, \hat Y, U) for achievability (Viswanatha et al., 2014, Andrade, 6 Jul 2025).

For lossless sources, CGK(X,Y)I(X;Y)CW(X,Y)C_{GK}(X, Y) \leq I(X; Y) \leq C_W(X, Y), with equality when all shared information can be deterministically separated (Andrade, 6 Jul 2025).

5. Explicit Constructions and Computation

Polar codes (for discrete) and polar lattices (for Gaussians) allow explicit extraction of Wyner’s lossy common information (Shi et al., 2016). The strategy for DSBS is to polar-quantize under the joint test channel, extract the common part as a high-entropy block, and compress private deviations. In the Gaussian case, the problem reduces to optimal quantization of a single latent WW; the common information plateaus for distortion levels below 1ρ1-\rho.

An explicit Gaussian algorithm follows:

  1. Canonicalization via Hotelling SVD.
  2. Parameter extraction: D=diag(d1,,dn)D = \operatorname{diag}(d_1,\ldots,d_n).
  3. Check Din(1d1)D_i \leq n(1-d_1).
  4. Compute CW=12jlog1+dj1djC_W = \frac{1}{2}\sum_j \log\frac{1 + d_j}{1 - d_j} (Charalambous et al., 2019).

The discrete Gaussian approximation and explicit coding constructions are proven to be achievable to within vanishing error (Shi et al., 2016).

6. Learnable Networks and Applications

Recent advances operationalize Gray–Wyner theory via learnable neural codecs for multitask computer vision problems (Andrade et al., 29 Jan 2026). These architectures instantiate three-channel (common and private) codes with structured neural transforms and entropy models. The Lagrangian-relaxed loss jointly optimizes rate allocation and distortion, automatically discovering the optimal splitting of common and private rates as predicted by theory. Empirical results verify that the learned codes attain the predicted rate savings on transmit–receive frontiers, with shared channels saturating theoretical bounds in strong-dependence regimes. Noteworthy effects include:

  • Dominantly shared codes when input PMFs coincide,
  • Zero shared rate for independent tasks,
  • Adaptive bit allocation for mixed dependence.

Lossy common information interrelates with multiple research axes:

  • Limited common randomness: The minimum common-randomness rate for constrained distortion, single-letter achievable region, and its optimization as a convex program (Saldi et al., 2014).
  • Mutual information bounds: I(Z^1;Z^2)I(\hat Z_1; \hat Z_2) forms a tight sandwich between lossy Wyner and Gács–Körner CIs for all achievable reconstructions (Andrade, 6 Jul 2025).
  • Generalizations: Extensions to NN-tuples, arbitrary alphabets, and output distribution constraints, with the unified perspective of the Gray–Wyner rate region (Xu et al., 2013).
  • Unified transmit/receive trade-off: The locus of achievable (R0,R1,R2)(R_0, R_1, R_2) traces contours on the Gray–Wyner surface, interpolating between fully-shared and fully-private extreme points (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026).

Table: Summary of Characterizations

Notion Definition Markov Constraint
Lossy Wyner CI CWC_W infI(X,Y;U)\inf I(X,Y;U) (X,Y)(X^,Y^)U,X^UY^(X,Y)-(\hat X,\hat Y)-U,\,\hat X-U-\hat Y
Lossy Gács-Körner CI supI(X,Y;V)\sup I(X,Y;V) YXV,XYV,XX^V,YY^VY-X-V,\,X-Y-V,\,X-\hat X-V,\,Y-\hat Y-V
Mutual Info Bound KI(Z^1;Z^2)CK \leq I(\hat Z_1;\hat Z_2) \leq C N/A

Wyner’s and Gács–Körner’s notions represent fundamental bounds in multiterminal source coding and are critical for understanding redundancy, sequential refinability, and practical codec design, in both classical and modern machine learning systems. Their generalizations to arbitrary sources, distortion regimes, and learnable representations continue to inform theoretical analysis and applied algorithm development across several disciplines.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Lossy Common Information.