The Rate-Distortion-Perception-Classification Tradeoff: Joint Source Coding and Modulation via Inverse-Domain GANs
Abstract: The joint source-channel coding (JSCC) framework leverages deep learning to learn from data the best codes for source and channel coding. When the output signal, rather than being binary, is directly mapped onto the IQ domain (complex-valued), we call the resulting framework joint source coding and modulation (JSCM). We consider a JSCM scenario and show the existence of a strict tradeoff between channel rate, distortion, perception, and classification accuracy, a tradeoff that we name RDPC. We then propose two image compression methods to navigate that tradeoff: the RDPCO algorithm which, under simple assumptions, directly solves the optimization problem characterizing the tradeoff, and an algorithm based on an inverse-domain generative adversarial network (ID-GAN), which is more general and achieves extreme compression. Simulation results corroborate the theoretical findings, showing that both algorithms exhibit the RDPC tradeoff. They also demonstrate that the proposed ID-GAN algorithm effectively balances image distortion, perception, and classification accuracy, and significantly outperforms traditional separation-based methods and recent deep JSCM architectures in terms of one or more of these metrics.
- “Elements of Information Theory” Wiley & Sons, 1991
- E. Bourtsoulatze, D.B. Kurka and D. Gündüz “Deep joint source-channel coding for wireless image transmission” In IEEE Trans. Cog. Comms. Network. 39.1, 2019, pp. 89–100
- “The perception-distortion tradeoff” In CVPR, 2018, pp. 6228–6237
- “Rethinking Lossy Compression: The rate-distortion-perception tradeoff” In ICML, 2019, pp. 675–685
- D. Liu, H. Zhang and Z. Xiong “On the classification-distortion-perception tradeoff” In NeurIPS, 2019, pp. 1–10
- “Generative adversarial nets” In NeurIPS, 2014, pp. 1–9
- “In-domain GAN inversion for real image editing” In ECCV, 2020, pp. 592–608
- “Generative adversarial networks for extreme learned image compression” In ICCV, 2019, pp. 221–231
- M. Jankowski, D. Gündüz and K. Mikolajczyk “Wireless image retrieval at the edge” In IEEE J. Selected Areas in Communications 39.1, 2020, pp. 89–100
- “DeepJSCC-f: Deep joint source-channel coding of images with feedback” In IEEE J. Selected Areas in Inf. Th. 1.1, 2020, pp. 178–193
- “Bandwidth-agile image transmission with deep joint source-channel coding” In IEEE Trans. Wireless Comm. 20.12, 2021, pp. 8081–8095
- “Deep Joint Source-Channel Coding for CSI Feedback: An End-to-End Approach” In IEEE J. Selected Areas in Communications 41.1, 2023, pp. 260–273
- M. Yang, C. Bian and H.-S. Kim “Deep Joint Source Channel Coding for Wireless Image Transmission with OFDM” In IEEE Int. Conf. Comms., 2021, pp. 1–6
- N. Farsad, M. Rao and A. Goldsmith “Deep Learning for Joint Source-Channel Coding of Text” In ICASSP, 2018, pp. 2326–2330
- “Deep Learning Enabled Semantic Communication Systems” In IEEE T-SP 69, 2021, pp. 2663–2675
- “Semantic Communication Systems for Speech Transmission” In IEEE J. Selected Areas in Communications 39.8, 2021, pp. 2434–2444
- “Cooperative Task-Oriented Communication for Multi-Modal Data with Transmission Control” In IEEE Int. Conf. Comms. Workshops, 2023, pp. 1635–1640
- Q.Yang Z.Zhang, S. He and al. “Semantic communication approach for multi-task image transmission” In IEEE VTC, 2022, pp. 1–2
- D.E. Rumelhart, G.E. Hinton and R.J. Williams “Learning representations by back-propagating errors” In Nature 323.6088, 1986, pp. 533–536
- “Wireless Image Transmission Using Deep Source Channel Coding With Attention Modules” In IEEE Trans. Circuits Sys. for Video Tech. 32.4, 2022, pp. 2315–2328
- T. Karras, S. Laine and T. Aila “A style-based generator architecture for generative adversarial networks” In CVPR, 2019, pp. 4401–4410
- “Generative Joint Source-Channel Coding for Semantic Image Transmission” In IEEE J. Selected Areas in Communications 41.8, 2023, pp. 2645–2657
- “Analyzing and improving the image quality of StyleGAN” In CVPR, 2020, pp. 8110–8119
- “The unreasonable effectiveness of deep features as a perceptual metric” In CVPR, 2018, pp. 586–595
- “On perceptual lossy compression: The cost of perceptual reconstruction and an optimal training framework” In ICML, 2021, pp. 11682–11692
- “Information theory and statistics: A tutorial” In Found. and Trends in Communications and Information Theory 1.4, 2004, pp. 417–528
- “Rényi divergence and Kullback-Leibler divergence” In IEEE T-IT 60.7, 2014, pp. 3797–3820
- M. Arjovsky, S. Chintala and L. Bottou “Wasserstein generative adversarial networks” In ICML, 2017, pp. 214–223
- “Improved training of Wasserstein GANs” In NeurIPS, 2017, pp. 1–11
- A. Jacot, F. Gabriel and C. Hongler “Neural tangent kernel: Convergence and generalization in neural networks” In NeurIPS, 2018, pp. 1–10
- “Wide neural networks of any depth evolve as linear models under gradient descent” In NeurIPS, 2019, pp. 1–10
- “The distance between two random vectors with given dispersion matrices” In Linear algebra and its applications 48, 1982, pp. 257–263
- “The Fréchet distance between multivariate normal distributions” In J. Multivariate Analysis 12, 1982, pp. 450–455
- R.O. Duda, P.E. Hart and D.G. Stork “Pattern Classification” Wiley, 2001
- “Convex Optimization” Cambridge University Press, 2004 URL: https://web.stanford.edu/~boyd/cvxbook/
- “Gradient-based learning applied to document recognition” In Proc IEEE 86.11, 1998, pp. 2278–2324
- R. Gallager “Low-density parity-check codes” In IRE Trans. Inf. Th. 8.1, 1962, pp. 21–28
- “GANs trained by a two time-scale update rule converge to a local Nash equilibrium” In NeurIPS, 2017, pp. 1–12
- C. Villani “Optimal Transport: Old and New” Springer, 2009
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.