Learning to See Through Dazzle
Abstract: Machine vision is susceptible to laser dazzle, where intense laser light can blind and distort its perception of the environment through oversaturation or permanent damage to sensor pixels. Here we employ a wavefront-coded phase mask to diffuse the energy of laser light and introduce a sandwich generative adversarial network (SGAN) to restore images from complex image degradations, such as varying laser-induced image saturation, mask-induced image blurring, unknown lighting conditions, and various noise corruptions. The SGAN architecture combines discriminative and generative methods by wrapping two GANs around a learnable image deconvolution module. In addition, we make use of Fourier feature representations to reduce the spectral bias of neural networks and improve its learning of high-frequency image details. End-to-end training includes the realistic physics-based synthesis of a large set of training data from publicly available images. We trained the SGAN to suppress the peak laser irradiance as high as $106$ times the sensor saturation threshold - the point at which camera sensors may experience damage without the mask. The trained model was evaluated on both a synthetic data set and data collected from the laboratory. The proposed image restoration model quantitatively and qualitatively outperforms state-of-the-art methods for a wide range of scene contents, laser powers, incident laser angles, ambient illumination strengths, and noise characteristics.
- O. Steinvall, “The potential role of laser in combating uav: part 2; laser as a countermeasure and weapon,” in Technologies for Optical Countermeasures XVIII and High-Power Lasers: Technology and Systems, Platforms, Effects V, vol. 11867. SPIE, 2021.
- O. Steinvall, “Laser dazzling: an overview,” in Technologies for Optical Countermeasures XIX, vol. 12738. SPIE, 2023, pp. 17–31.
- G. D. Lewis, A. Borghgraef, and M. Vandewal, “The disruptive impact of dynamic laser dazzling on template matching algorithms applied to thermal infrared imagery,” in Technologies for Optical Countermeasures XIX, vol. 12738. SPIE, 2023, p. 1273803.
- R. Duan, X. Mao, A. K. Qin, Y. Chen, S. Ye, Y. He, and Y. Yang, “Adversarial laser beam: Effective physical-world attack to dnns in a blink,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 16 062–16 071.
- K. Kim, J. Kim, S. Song, J.-H. Choi, C. Joo, and J.-S. Lee, “Engineering pupil function for optical adversarial attacks,” Optics Express, vol. 30, no. 5, pp. 6500–6518, 2022.
- Y. Sun, Y. Huang, and X. Wei, “Embodied adversarial attack: A dynamic robust physical attack in autonomous driving,” arXiv preprint arXiv:2312.09554, 2023.
- F. Quercioli, “Beyond laser safety glasses: augmented reality in optics laboratories,” Applied optics, vol. 56, no. 4, pp. 1148–1150, 2017.
- G. Owczarek, M. Wodzyński, J. Szkudlarek, and M. Jachowicz, “Virtual reality (vr) for laser safety training,” in 2021 IEEE 2nd International Conference on Human-Machine Systems (ICHMS). IEEE, 2021, pp. 1–3.
- J. Deniel and S. Thommet, “Occupational eye protection using augmented reality: a proof of concept,” Radioprotection, vol. 57, no. 2, pp. 165–173, 2022.
- K. Li, A. Choudhuri, S. Schmidt, T. Lang, R. Bacher, I. Hartl, W. Leemans, and F. Steinicke, “Mixed reality for laser safety at advanced optics laboratories,” in International Laser Safety Conference, no. PUBDB-2023-07345. Control-System, 2023.
- “Laser incidents,” www.faa.gov/about/initiatives/lasers/laws.
- “Laser hazard,” www.osha.gov/laser-hazards.
- D. Malayanur, V. N. Mysore et al., “Laser safety eyewear,” CosmoDerma, vol. 2, 2022.
- www.ilda.com/camera-sensor-damage.htm.
- G. J. Ruane, A. T. Watnik, and G. A. Swartzlander, “Reducing the risk of laser damage in a focal plane array using linear pupil-plane phase elements,” Applied optics, vol. 54, no. 2, pp. 210–218, 2015.
- G. Ritt, “Laser safety calculations for imaging sensors,” Sensors, vol. 19, no. 17, 2019.
- F. Théberge, M. Auclair, J.-F. Daigle, and D. Pudo, “Damage thresholds of silicon-based cameras for in-band and out-of-band laser expositions,” Applied Optics, vol. 61, no. 10, pp. 2473–2482, 2022.
- G. Ritt, B. Schwarz, and B. Eberle, “Preventing image information loss of imaging sensors in case of laser dazzle,” Optical Engineering, vol. 58, no. 1, p. 013109, 2019.
- G. Ritt and B. Eberle, “Use of complementary wavelength bands for laser dazzle protection,” Optical Engineering, vol. 59, no. 1, p. 015106, 2020.
- G. J. Ruane and G. A. Swartzlander, “Optical vortex coronagraphy with an elliptical aperture,” Applied Optics, vol. 52, no. 2, pp. 171–176, 2013.
- G. J. Ruane, P. Kanburapa, J. Han, and G. A. Swartzlander, “Vortex-phase filtering technique for extracting spatial information from unresolved sources,” Applied Optics, vol. 53, no. 20, pp. 4503–4508, 2014.
- A. T. Watnik, S. Divitt, and M. Hart, “Separation of coherent and incoherent light using image plane vortex phase masks,” JOSA A, vol. 40, no. 7, pp. 1372–1381, 2023.
- G. Swartzlander Jr, B. Justus, A. Huston, A. Campillo, and C. Law, “Characteristics of a low f-number broadband visible thermal optical limiter,” International Journal of Nonlinear Optical Physics, vol. 2, no. 04, pp. 577–611, 1993.
- D. Dini, M. J. Calvete, and M. Hanack, “Nonlinear optical materials for the smart filtering of optical radiation,” Chemical reviews, vol. 116, no. 22, pp. 13 043–13 233, 2016.
- J. Caillieaudeaux, O. Muller, M. Guerchoux, C. Bruder, L. Merlat, A.-S. Schuller, and C. Delaite, “Thermoset polymers as host for optical limiting,” Journal of Applied Polymer Science, vol. 141, no. 3, p. e54810, 2024.
- L. Wang, “Self-activating liquid crystal devices for smart laser protection,” Liquid Crystals, vol. 43, no. 13-15, pp. 2062–2078, 2016.
- R. Zhang, Z. Zhang, J. Han, L. Yang, J. Li, Z. Song, T. Wang, and J. Zhu, “Advanced liquid crystal-based switchable optical devices for light protection applications: principles and strategies,” Light: Science & Applications, vol. 12, no. 1, p. 11, 2023.
- A. Howes, Z. Zhu, D. Curie, J. R. Avila, V. D. Wheeler, R. F. Haglund, and J. G. Valentine, “Optical limiting based on huygens’ metasurfaces,” Nano Letters, vol. 20, no. 6, pp. 4638–4644, 2020.
- N. Bonod, P. Brianceau, J. Daurios, S. Grosjean, N. Roquin, J.-F. Gleyze, L. Lamaignère, and J. Neauport, “Linear-to-circular polarization conversion with full-silica meta-optics to reduce nonlinear effects in high-energy lasers,” Nature Communications, vol. 14, no. 1, p. 5383, 2023.
- G. Lewis, C. Santos, and M. Vandewal, “Mitigation of laser dazzle effects on a mid-wave infrared thermal imager by reducing the integration time of the focal plane array,” in Technologies for Optical Countermeasures XVI, vol. 11161. International Society for Optics and Photonics, 2019, p. 1116108.
- R. H. Schleijpen, S. Binsbergen, A. Vosteen, K. de Groot-Trouw, D. Meuken, and A. M. Van Eijk, “Smoke as protection against high energy laser effects,” in Technologies for Optical Countermeasures XVIII and High-Power Lasers: Technology and Systems, Platforms, Effects V, vol. 11867. SPIE, 2021, p. 1186704.
- A. T. Watnik, G. J. Ruane, and G. A. Swartzlander, “Incoherent imaging in the presence of unwanted laser radiation: vortex and axicon wavefront coding,” Optical Engineering, vol. 55, no. 12, p. 123102, 2016.
- J. H. Wirth, A. T. Watnik, and G. A. Swartzlander, “Experimental observations of a laser suppression imaging system using pupil-plane phase elements,” Applied optics, vol. 56, no. 33, pp. 9205–9211, 2017.
- J. H. Wirth, A. T. Watnik, and G. A. Swartzlander, “Optimized pupil-plane phase masks for high-contrast imaging,” Applied optics, vol. 57, no. 20, pp. 5688–5693, 2018.
- J. H. Wirth, A. T. Watnik, and G. A. Swartzlander, “Computational imaging for reducing peak irradiance on focal planes,” in Computational Imaging III, vol. 10669. International Society for Optics and Photonics, 2018, p. 106690U.
- J. H. Wirth, A. T. Watnik, and G. A. Swartzlander, “Half-ring point spread functions,” Optics letters, vol. 45, no. 8, pp. 2179–2182, 2020.
- K. Novak and A. T. Watnik, “Compact vortex wavefront coding camera,” in Computational Imaging V, vol. 11396. International Society for Optics and Photonics, 2020, p. 113960O.
- K. Novak and A. T. Watnik, “Imaging through deconvolution with a spatially variant point spread function,” in Computational Imaging VI, vol. 11731. International Society for Optics and Photonics, 2021, p. 1173105.
- X. Peng, P. R. Srivastava, and G. A. Swartzlander, “Cnn-based real-time image restoration in laser suppression imaging,” in Imaging and Sensing Congress. Optica Publishing Group, 2021, pp. JTh6A–10.
- A. Ghosh and G. Swartzlander, “Shivanet: Shift variant image deconvolution using deep learning,” in 2023 IEEE Western New York Image and Signal Processing Workshop (WNYISPW). IEEE, 2023, pp. 1–4.
- A. Ghosh, “Shift variant image deconvolution using deep learning,” Ph.D. dissertation, Rochester Institute of Technology, 2023.
- G. Barbastathis, A. Ozcan, and G. Situ, “On the use of deep learning for computational imaging,” Optica, vol. 6, no. 8, pp. 921–943, 2019.
- Y. Peng, Q. Fu, F. Heide, and W. Heidrich, “The diffractive achromat full spectrum computational imaging with diffractive optics,” ACM Transactions on Graphics (TOG), vol. 35, no. 4, pp. 1–11, 2016.
- X. Dun, H. Ikoma, G. Wetzstein, Z. Wang, X. Cheng, and Y. Peng, “Learned rotationally symmetric diffractive achromat for full-spectrum computational imaging,” Optica, vol. 7, no. 8, pp. 913–922, 2020.
- R. Raskar, A. Agrawal, and J. Tumblin, “Coded exposure photography: motion deblurring using fluttered shutter,” in Acm Siggraph 2006 Papers, 2006, pp. 795–804.
- V. Sitzmann, S. Diamond, Y. Peng, X. Dun, S. Boyd, W. Heidrich, F. Heide, and G. Wetzstein, “End-to-end optimization of optics and image processing for achromatic extended depth of field and super-resolution imaging,” ACM Transactions on Graphics (TOG), vol. 37, no. 4, pp. 1–13, 2018.
- Y. Wu, V. Boominathan, H. Chen, A. Sankaranarayanan, and A. Veeraraghavan, “Phasecam3d—learning phase masks for passive single view depth estimation,” in 2019 IEEE International Conference on Computational Photography (ICCP). IEEE, 2019, pp. 1–12.
- S. Tan, Y. Wu, S.-I. Yu, and A. Veeraraghavan, “Codedstereo: Learned phase masks for large depth-of-field stereo,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 7170–7179.
- E.-V. Talvala, A. Adams, M. Horowitz, and M. Levoy, “Veiling glare in high dynamic range imaging,” ACM Transactions on Graphics (TOG), vol. 26, no. 3, pp. 37–es, 2007.
- R. Raskar, A. Agrawal, C. A. Wilson, and A. Veeraraghavan, “Glare aware photography: 4d ray sampling for reducing glare effects of camera lenses,” in ACM SIGGRAPH 2008 papers, 2008, pp. 1–10.
- M. Rouf, R. Mantiuk, W. Heidrich, M. Trentacoste, and C. Lau, “Glare encoding of high dynamic range images,” in CVPR 2011. IEEE, 2011, pp. 289–296.
- Q. Sun, E. Tseng, Q. Fu, W. Heidrich, and F. Heide, “Learning rank-1 diffractive optics for single-shot high dynamic range imaging,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1386–1396.
- C. A. Metzler, H. Ikoma, Y. Peng, and G. Wetzstein, “Deep optics for single-shot high-dynamic-range imaging,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 1375–1385.
- A. Veeraraghavan, R. Raskar, A. Agrawal, A. Mohan, and J. Tumblin, “Dappled photography: Mask enhanced cameras for heterodyned light fields and coded aperture refocusing,” ACM Transactions on Graphics (TOG), vol. 26, no. 3, p. 69, 2007.
- C. Hinojosa, J. C. Niebles, and H. Arguello, “Learning privacy-preserving optics for human pose estimation,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 2573–2582.
- Z. Tasneem, G. Milione, Y.-H. Tsai, X. Yu, A. Veeraraghavan, M. Chandraker, and F. Pittaluga, “Learning phase mask for privacy-preserving passive depth estimation,” in Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part VII. Springer, 2022, pp. 504–521.
- D. S. Jeon, S.-H. Baek, S. Yi, Q. Fu, X. Dun, W. Heidrich, and M. H. Kim, “Compact snapshot hyperspectral imaging with diffracted rotation,” ACM Transactions on Graphics (TOG), vol. 38, no. 4, pp. 1–13, 2019.
- S.-H. Baek, H. Ikoma, D. S. Jeon, Y. Li, W. Heidrich, G. Wetzstein, and M. H. Kim, “Single-shot hyperspectral-depth imaging with learned diffractive optics,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2651–2660.
- A. Kotwal, A. Levin, and I. Gkioulekas, “Interferometric transmission probing with coded mutual intensity,” ACM Transactions on Graphics (TOG), vol. 39, no. 4, pp. 74–1, 2020.
- X. Peng, G. J. Ruane, M. B. Quadrelli, and G. A. Swartzlander, “Randomized apertures: high resolution imaging in far field,” Optics express, vol. 25, no. 15, pp. 18 296–18 313, 2017.
- X. Peng, G. J. Ruane, and G. A. Swartzlander Jr, “Randomized aperture imaging,” arXiv preprint arXiv:1601.00033, 2016.
- X. Lin, Y. Liu, J. Wu, and Q. Dai, “Spatial-spectral encoded compressive hyperspectral imaging,” ACM Transactions on Graphics (TOG), vol. 33, no. 6, pp. 1–11, 2014.
- E. Vargas, J. N. Martel, G. Wetzstein, and H. Arguello, “Time-multiplexed coded aperture imaging: Learned coded aperture and pixel exposures for compressive imaging systems,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2692–2702.
- Y. Peng, S. Choi, N. Padmanaban, and G. Wetzstein, “Neural holography with camera-in-the-loop training,” ACM Transactions on Graphics (TOG), vol. 39, no. 6, pp. 1–14, 2020.
- S. Choi, M. Gopakumar, Y. Peng, J. Kim, and G. Wetzstein, “Neural 3d holography: learning accurate wave propagation models for 3d holographic virtual and augmented reality displays,” ACM Transactions on Graphics (TOG), vol. 40, no. 6, pp. 1–12, 2021.
- C. Wang, Q. Fu, X. Dun, and W. Heidrich, “Megapixel adaptive optics: towards correcting large-scale distortions in computational cameras,” ACM Transactions on Graphics (TOG), vol. 37, no. 4, pp. 1–12, 2018.
- Q. Sun, J. Zhang, X. Dun, B. Ghanem, Y. Peng, and W. Heidrich, “End-to-end learned, optically coded super-resolution spad camera,” ACM Transactions on Graphics (TOG), vol. 39, no. 2, pp. 1–14, 2020.
- V. Boominathan, J. T. Robinson, L. Waller, and A. Veeraraghavan, “Recent advances in lensless imaging,” Optica, vol. 9, no. 1, pp. 1–16, 2022.
- D. Krishnan and R. Fergus, “Fast image deconvolution using hyper-laplacian priors,” Advances in neural information processing systems, vol. 22, 2009.
- M. Grosse, G. Wetzstein, A. Grundhöfer, and O. Bimber, “Coded aperture projection,” ACM Transactions on Graphics (TOG), vol. 29, no. 3, pp. 1–12, 2010.
- E. P. Simoncelli, B. Olshausen et al., “Natural image statistics and neural representation,” Annual review of neuroscience, vol. 24, no. 1, pp. 1193–1216, 2001.
- A. Levin, R. Fergus, F. Durand, and W. T. Freeman, “Image and depth from a conventional camera with a coded aperture,” ACM transactions on graphics (TOG), vol. 26, no. 3, pp. 70–es, 2007.
- S. Osher, M. Burger, D. Goldfarb, J. Xu, and W. Yin, “An iterative regularization method for total variation-based image restoration,” Multiscale Modeling & Simulation, vol. 4, no. 2, pp. 460–489, 2005.
- U. Schmidt, C. Rother, S. Nowozin, J. Jancsary, and S. Roth, “Discriminative non-blind deblurring,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 604–611.
- U. Schmidt and S. Roth, “Shrinkage fields for effective image restoration,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 2774–2781.
- D. Geman and C. Yang, “Nonlinear image recovery with half-quadratic regularization,” IEEE transactions on Image Processing, vol. 4, no. 7, pp. 932–946, 1995.
- A. Levin and Y. Weiss, “User assisted separation of reflections from a single image using a sparsity prior,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 9, pp. 1647–1654, 2007.
- E. Esser, X. Zhang, and T. F. Chan, “A general framework for a class of first order primal-dual algorithms for convex optimization in imaging science,” SIAM Journal on Imaging Sciences, vol. 3, no. 4, pp. 1015–1046, 2010.
- S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein et al., “Distributed optimization and statistical learning via the alternating direction method of multipliers,” Foundations and Trends® in Machine learning, vol. 3, no. 1, pp. 1–122, 2011.
- N. Antipa, G. Kuo, R. Heckel, B. Mildenhall, E. Bostan, R. Ng, and L. Waller, “Diffusercam: lensless single-exposure 3d imaging,” Optica, vol. 5, no. 1, pp. 1–9, 2018.
- K. Monakhova, K. Yanny, N. Aggarwal, and L. Waller, “Spectral diffusercam: lensless snapshot hyperspectral imaging with a spectral filter array,” Optica, vol. 7, no. 10, pp. 1298–1307, 2020.
- V. Boominathan, J. K. Adams, J. T. Robinson, and A. Veeraraghavan, “Phlatcam: Designed phase-mask based thin lensless camera,” IEEE transactions on pattern analysis and machine intelligence, vol. 42, no. 7, pp. 1618–1629, 2020.
- Y. Zheng and M. S. Asif, “Joint image and depth estimation with mask-based lensless cameras,” IEEE Transactions on Computational Imaging, vol. 6, pp. 1167–1178, 2020.
- C. J. Schuler, H. Christopher Burger, S. Harmeling, and B. Scholkopf, “A machine learning approach for non-blind image deconvolution,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2013, pp. 1067–1074.
- H. Son and S. Lee, “Fast non-blind deconvolution via regularized residual networks with long/short skip-connections,” in 2017 IEEE International Conference on Computational Photography (ICCP). IEEE, 2017, pp. 1–10.
- K. Yanny, K. Monakhova, R. W. Shuai, and L. Waller, “Deep learning for fast spatially varying deconvolution,” Optica, vol. 9, no. 1, pp. 96–99, 2022.
- J. Dong, S. Roth, and B. Schiele, “Deep wiener deconvolution: Wiener meets deep learning for image deblurring,” Advances in Neural Information Processing Systems, vol. 33, pp. 1048–1059, 2020.
- Z. Shi, Y. Bahat, S.-H. Baek, Q. Fu, H. Amata, X. Li, P. Chakravarthula, W. Heidrich, and F. Heide, “Seeing through obstructions with diffractive cloaking,” ACM Transactions on Graphics (TOG), vol. 41, no. 4, pp. 1–15, 2022.
- J. Zhang, J. Pan, W.-S. Lai, R. W. Lau, and M.-H. Yang, “Learning fully convolutional networks for iterative non-blind deconvolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3817–3825.
- J. Kruse, C. Rother, and U. Schmidt, “Learning to push the limits of efficient fft-based image deconvolution,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 4586–4594.
- K. Monakhova, J. Yurtsever, G. Kuo, N. Antipa, K. Yanny, and L. Waller, “Learned reconstructions for practical mask-based lensless imaging,” Optics express, vol. 27, no. 20, pp. 28 075–28 090, 2019.
- H. Chen, J. Gu, O. Gallo, M.-Y. Liu, A. Veeraraghavan, and J. Kautz, “Reblur2deblur: Deblurring videos via self-supervised learning,” in 2018 IEEE International Conference on Computational Photography (ICCP). IEEE, 2018, pp. 1–9.
- D. Ren, K. Zhang, Q. Wang, Q. Hu, and W. Zuo, “Neural blind deconvolution using deep priors,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 3341–3350.
- K. Monakhova, V. Tran, G. Kuo, and L. Waller, “Untrained networks for compressive lensless photography,” Optics Express, vol. 29, no. 13, pp. 20 913–20 929, 2021.
- S. Harmeling, S. Sra, M. Hirsch, and B. Schölkopf, “Multiframe blind deconvolution, super-resolution, and saturation correction via incremental em,” in 2010 IEEE International Conference on Image Processing. IEEE, 2010, pp. 3313–3316.
- O. Whyte, J. Sivic, and A. Zisserman, “Deblurring shaken and partially saturated images,” International journal of computer vision, vol. 110, no. 2, pp. 185–201, 2014.
- L. Chen, J. Zhang, J. Pan, S. Lin, F. Fang, and J. S. Ren, “Learning a non-blind deblurring network for night blurry images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10 542–10 550.
- J. Pan, Z. Lin, Z. Su, and M.-H. Yang, “Robust kernel estimation with outliers handling for image deblurring,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2800–2808.
- L. Chen, J. Zhang, S. Lin, F. Fang, and J. S. Ren, “Blind deblurring for saturated images,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 6308–6316.
- K. Zhang, W. Ren, W. Luo, W.-S. Lai, B. Stenger, M.-H. Yang, and H. Li, “Deep image deblurring: A survey,” International Journal of Computer Vision, pp. 1–28, 2022.
- O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention. Springer, 2015, pp. 234–241.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- X. Tao, H. Gao, X. Shen, J. Wang, and J. Jia, “Scale-recurrent network for deep image deblurring,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8174–8182.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- J. Liang, J. Cao, G. Sun, K. Zhang, L. Van Gool, and R. Timofte, “Swinir: Image restoration using swin transformer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 1833–1844.
- Z. Wang, X. Cun, J. Bao, W. Zhou, J. Liu, and H. Li, “Uformer: A general u-shaped transformer for image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 17 683–17 693.
- S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, and M.-H. Yang, “Restormer: Efficient transformer for high-resolution image restoration,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5728–5739.
- H. Chen, Y. Wang, T. Guo, C. Xu, Y. Deng, Z. Liu, S. Ma, C. Xu, C. Xu, and W. Gao, “Pre-trained image processing transformer,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 12 299–12 310.
- Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 10 012–10 022.
- Z. Shen, M. Zhang, H. Zhao, S. Yi, and H. Li, “Efficient attention: Attention with linear complexities,” in WACV, 2021.
- J. Lee-Thorp, J. Ainslie, I. Eckstein, and S. Ontanon, “Fnet: Mixing tokens with fourier transforms,” arXiv preprint arXiv:2105.03824, 2021.
- H. Liu, Z. Dai, D. So, and Q. V. Le, “Pay attention to mlps,” Advances in Neural Information Processing Systems, vol. 34, pp. 9204–9215, 2021.
- Z. Tu, H. Talebi, H. Zhang, F. Yang, P. Milanfar, A. Bovik, and Y. Li, “Maxim: Multi-axis mlp for image processing,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 5769–5780.
- O. Rippel, J. Snoek, and R. P. Adams, “Spectral representations for convolutional neural networks,” Advances in neural information processing systems, vol. 28, 2015.
- L. Chi, B. Jiang, and Y. Mu, “Fast fourier convolution,” Advances in Neural Information Processing Systems, vol. 33, pp. 4479–4488, 2020.
- M. Tancik, P. Srinivasan, B. Mildenhall, S. Fridovich-Keil, N. Raghavan, U. Singhal, R. Ramamoorthi, J. Barron, and R. Ng, “Fourier features let networks learn high frequency functions in low dimensional domains,” Advances in Neural Information Processing Systems, vol. 33, pp. 7537–7547, 2020.
- J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and super-resolution,” in European conference on computer vision. Springer, 2016, pp. 694–711.
- K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in International Conference on Learning Representations, 2015.
- T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, A. Tao, J. Kautz, and B. Catanzaro, “High-resolution image synthesis and semantic manipulation with conditional gans,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8798–8807.
- S.-J. Cho, S.-W. Ji, J.-P. Hong, S.-W. Jung, and S.-J. Ko, “Rethinking coarse-to-fine approach in single image deblurring,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 4641–4650.
- F.-J. Tsai, Y.-T. Peng, Y.-Y. Lin, C.-C. Tsai, and C.-W. Lin, “Stripformer: Strip transformer for fast image deblurring,” in ECCV, 2022.
- X. Mao, Y. Liu, F. Liu, Q. Li, W. Shen, and Y. Wang, “Intriguing findings of frequency selection for image deblurring,” in Proceedings of the 37th AAAI Conference on Artificial Intelligence, 2023.
- H. Zhang, Y. Dai, H. Li, and P. Koniusz, “Deep stacked hierarchical multi-patch network for image deblurring,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5978–5986.
- S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, M.-H. Yang, and L. Shao, “Multi-stage progressive image restoration,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 14 821–14 831.
- R. Feng, C. Li, H. Chen, S. Li, C. C. Loy, and J. Gu, “Removing diffraction image artifacts in under-display camera via dynamic skip connection network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 662–671.
- R. Liu, J. Lehman, P. Molino, F. Petroski Such, E. Frank, A. Sergeev, and J. Yosinski, “An intriguing failing of convolutional neural networks and the coordconv solution,” Advances in neural information processing systems, vol. 31, 2018.
- C. H. Lin, C.-C. Chang, Y.-S. Chen, D.-C. Juan, W. Wei, and H.-T. Chen, “Coco-gan: Generation by parts via conditional coordinating,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 4512–4521.
- C. H. Lin, H.-Y. Lee, Y.-C. Cheng, S. Tulyakov, and M.-H. Yang, “Infinitygan: Towards infinite-pixel image synthesis,” in International Conference on Learning Representations, 2021.
- D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” stat, vol. 1050, p. 1, 2014.
- D. P. Kingma and P. Dhariwal, “Glow: Generative flow with invertible 1x1 convolutions,” Advances in neural information processing systems, vol. 31, 2018.
- I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol. 27, 2014.
- J. Sohl-Dickstein, E. Weiss, N. Maheswaranathan, and S. Ganguli, “Deep unsupervised learning using nonequilibrium thermodynamics,” in International Conference on Machine Learning. PMLR, 2015, pp. 2256–2265.
- R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer, “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 10 684–10 695.
- Z. Xiao, K. Kreis, and A. Vahdat, “Tackling the generative learning trilemma with denoising diffusion gans,” in International Conference on Learning Representations, 2021.
- P. Dhariwal and A. Nichol, “Diffusion models beat gans on image synthesis,” Advances in neural information processing systems, vol. 34, pp. 8780–8794, 2021.
- J. Ho and T. Salimans, “Classifier-free diffusion guidance,” in NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications, 2021.
- H. Zheng, W. Nie, A. Vahdat, K. Azizzadenesheli, and A. Anandkumar, “Fast sampling of diffusion models via operator learning,” in International Conference on Machine Learning. PMLR, 2023, pp. 42 390–42 402.
- C. Meng, R. Rombach, R. Gao, D. Kingma, S. Ermon, J. Ho, and T. Salimans, “On distillation of guided diffusion models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 14 297–14 306.
- M. Kang, J.-Y. Zhu, R. Zhang, J. Park, E. Shechtman, S. Paris, and T. Park, “Scaling up gans for text-to-image synthesis,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 10 124–10 134.
- A. Sauer, T. Karras, S. Laine, A. Geiger, and T. Aila, “Stylegan-t: Unlocking the power of gans for fast large-scale text-to-image synthesis,” in International conference on machine learning. PMLR, 2023.
- P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1125–1134.
- J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2223–2232.
- T. Gruber, F. Julca-Aguilar, M. Bijelic, and F. Heide, “Gated2depth: Real-time dense lidar from gated images,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 1506–1516.
- Y. Shao, L. Li, W. Ren, C. Gao, and N. Sang, “Domain adaptation for image dehazing,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 2808–2817.
- C. Yang, Y. Shen, Z. Zhang, Y. Xu, J. Zhu, Z. Wu, and B. Zhou, “One-shot generative domain adaptation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 7733–7742.
- J. Chen, J. Chen, H. Chao, and M. Yang, “Image blind denoising with generative adversarial network based noise modeling,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3155–3164.
- O. Kupyn, V. Budzan, M. Mykhailych, D. Mishkin, and J. Matas, “Deblurgan: Blind motion deblurring using conditional adversarial networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 8183–8192.
- K. Zhang, W. Luo, Y. Zhong, L. Ma, B. Stenger, W. Liu, and H. Li, “Deblurring by realistic blurring,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 2737–2746.
- J. Pan, J. Dong, Y. Liu, J. Zhang, J. Ren, J. Tang, Y.-W. Tai, and M.-H. Yang, “Physics-based generative adversarial models for image restoration and beyond,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 7, pp. 2449–2462, 2020.
- J. Yu, Z. Lin, J. Yang, X. Shen, X. Lu, and T. S. Huang, “Generative image inpainting with contextual attention,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5505–5514.
- J. Yu, Z. Lin, J. Yang, X. Shen, X. Lu, and T. S. Huang, “Free-form image inpainting with gated convolution,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 4471–4480.
- S. Zhao, J. Cui, Y. Sheng, Y. Dong, X. Liang, I. Eric, C. Chang, and Y. Xu, “Large scale image completion via co-modulated generative adversarial networks,” in International Conference on Learning Representations, 2020.
- P. Teterwak, A. Sarna, D. Krishnan, A. Maschinot, D. Belanger, C. Liu, and W. T. Freeman, “Boundless: Generative adversarial networks for image extension,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 10 521–10 530.
- Y.-C. Cheng, C. H. Lin, H.-Y. Lee, J. Ren, S. Tulyakov, and M.-H. Yang, “Inout: Diverse image outpainting via gan inversion,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 11 431–11 440.
- Y. Peng, Q. Sun, X. Dun, G. Wetzstein, W. Heidrich, and F. Heide, “Learned large field-of-view imaging with thin-plate optics,” ACM Transactions on Graphics (TOG), vol. 38, no. 6, pp. 219–1, 2019.
- J. D. Rego, K. Kulkarni, and S. Jayasuriya, “Robust lensless image reconstruction via psf estimation,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 403–412.
- X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, and P. Abbeel, “Infogan: Interpretable representation learning by information maximizing generative adversarial nets,” Advances in neural information processing systems, vol. 29, 2016.
- T. Karras, S. Laine, and T. Aila, “A style-based generator architecture for generative adversarial networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4401–4410.
- T. Karras, M. Aittala, S. Laine, E. Härkönen, J. Hellsten, J. Lehtinen, and T. Aila, “Alias-free generative adversarial networks,” Advances in Neural Information Processing Systems, vol. 34, pp. 852–863, 2021.
- A. Sauer, K. Schwarz, and A. Geiger, “Stylegan-xl: Scaling stylegan to large diverse datasets,” in ACM SIGGRAPH 2022 conference proceedings, 2022, pp. 1–10.
- H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena, “Self-attention generative adversarial networks,” in International conference on machine learning. PMLR, 2019, pp. 7354–7363.
- I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C. Courville, “Improved training of wasserstein gans,” Advances in neural information processing systems, vol. 30, 2017.
- T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral normalization for generative adversarial networks,” in International Conference on Learning Representations, 2018.
- M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adversarial networks,” in International conference on machine learning. PMLR, 2017, pp. 214–223.
- P. Charbonnier, L. Blanc-Feraud, G. Aubert, and M. Barlaud, “Two deterministic half-quadratic regularization algorithms for computed imaging,” in Proceedings of 1st International Conference on Image Processing, vol. 2. IEEE, 1994, pp. 168–172.
- M. Hu and J. Tan, “Adaptive osculatory rational interpolation for image processing,” Journal of Computational and Applied Mathematics, vol. 195, no. 1-2, pp. 46–53, 2006.
- J. Wang, X. Li, and J. Yang, “Stacked conditional generative adversarial networks for jointly learning shadow detection and shadow removal,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 1788–1797.
- Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
- Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multiscale structural similarity for image quality assessment,” in The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, vol. 2. Ieee, 2003, pp. 1398–1402.
- R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 586–595.
- K. Ding, K. Ma, S. Wang, and E. P. Simoncelli, “Image quality assessment: Unifying structure and texture similarity,” IEEE transactions on pattern analysis and machine intelligence, 2020.
- https://unsplash.com/license.
- J. Liang, R. He, and T. Tan, “A comprehensive survey on test-time adaptation under distribution shifts,” arXiv preprint arXiv:2303.15361, 2023.
- K. Kim, S. Lee, and S. Cho, “Mssnet: Multi-scale-stage network for single image deblurring,” in European Conference on Computer Vision. Springer, 2022, pp. 524–539.
- E. Tseng, A. Mosleh, F. Mannan, K. St-Arnaud, A. Sharma, Y. Peng, A. Braun, D. Nowrouzezahrai, J.-F. Lalonde, and F. Heide, “Differentiable compound optics and processing pipeline optimization for end-to-end camera design,” ACM Transactions on Graphics (TOG), vol. 40, no. 2, pp. 1–19, 2021.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.