Papers
Topics
Authors
Recent
Search
2000 character limit reached

Joint Learning of Blind Super-Resolution and Crack Segmentation for Realistic Degraded Images

Published 24 Feb 2023 in cs.CV, cs.AI, and eess.IV | (2302.12491v3)

Abstract: This paper proposes crack segmentation augmented by super resolution (SR) with deep neural networks. In the proposed method, a SR network is jointly trained with a binary segmentation network in an end-to-end manner. This joint learning allows the SR network to be optimized for improving segmentation results. For realistic scenarios, the SR network is extended from non-blind to blind for processing a low-resolution image degraded by unknown blurs. The joint network is improved by our proposed two extra paths that further encourage the mutual optimization between SR and segmentation. Comparative experiments with State of The Art (SoTA) segmentation methods demonstrate the superiority of our joint learning, and various ablation studies prove the effects of our contributions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (70)
  1. F. Elghaish, S. Talebi, E. Abdellatef, S. T. Matarneh, M. R. Hosseini, S. Wu, M. Mayouf, A. Hajirasouli et al., “Developing a new deep learning cnn model to detect and classify highway cracks,” Journal of Engineering, Design and Technology, 2021.
  2. E. Shelhamer, J. Long, and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 4, pp. 640–651, 2017.
  3. K. He, G. Gkioxari, P. Dollár, and R. B. Girshick, “Mask R-CNN,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 42, no. 2, pp. 386–397, 2020.
  4. A. Kirillov, K. He, R. B. Girshick, C. Rother, and P. Dollár, “Panoptic segmentation,” in CVPR, 2019.
  5. S. Stent, R. Gherardi, B. Stenger, K. Soga, and R. Cipolla, “An image-based system for change detection on tunnel linings,” in MVA, 2013.
  6. Y. Fei, K. C. P. Wang, A. Zhang, C. Chen, J. Q. Li, Y. Liu, G. Yang, and B. Li, “Pixel-level cracking detection on 3d asphalt pavement images through deep-learning- based cracknet-v,” IEEE Trans. Intell. Transp. Syst., vol. 21, no. 1, pp. 273–284, 2020.
  7. Y. Kondo and N. Ukita, “Crack segmentation for low-resolution images using joint learning with super-resolution,” in MVA, 2021.
  8. H. Bae, K. Jang, and Y.-K. An, “Deep super resolution crack network (srcnet) for improving computer vision–based automated crack detectability in in situ bridges,” Structural Health Monitoring, vol. 20, no. 4, pp. 1428–1442, 2021.
  9. L. Wang, D. Li, Y. Zhu, L. Tian, and Y. Shan, “Dual super-resolution learning for semantic segmentation,” in CVPR, 2020, https://github.com/Dootmaan/DSRL.
  10. S. Minaee, Y. Boykov, F. Porikli, A. Plaza, N. Kehtarnavaz, and D. Terzopoulos, “Image segmentation using deep learning: A survey,” arXiv, 2020.
  11. D. Dais, İhsan Engin Bal, E. Smyrou, and V. Sarhosis, “Automatic crack classification and segmentation on masonry surfaces using convolutional neural networks and transfer learning,” Automation in Construction, vol. 125, p. 103606, 2021.
  12. T. Lin, P. Goyal, R. B. Girshick, K. He, and P. Dollár, “Focal loss for dense object detection,” in ICCV, 2017.
  13. Z. Li, K. Kamnitsas, and B. Glocker, “Analyzing overfitting under class imbalance in neural networks for image segmentation,” IEEE Trans. Medical Imaging, vol. 40, no. 3, pp. 1065–1077, 2021.
  14. M. S. Hossain, J. M. Betts, and A. P. Paplinski, “Dual focal loss to address class imbalance in semantic segmentation,” Neurocomputing, vol. 462, pp. 69–87, 2021.
  15. S. R. Bulò, G. Neuhold, and P. Kontschieder, “Loss max-pooling for semantic image segmentation,” in CVPR, 2017.
  16. L. Gong, Y. Zhang, Y. Zhang, Y. Yang, and W. Xu, “Erroneous pixel prediction for semantic image segmentation,” Comput. Vis. Media, vol. 8, no. 1, pp. 165–175, 2022.
  17. F. Milletari, N. Navab, and S. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 3DV, 2016.
  18. C. H. Sudre, W. Li, T. Vercauteren, S. Ourselin, and M. J. Cardoso, “Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations,” in MICCAI, 2017.
  19. S. A. Taghanaki, Y. Zheng, S. K. Zhou, B. Georgescu, P. Sharma, D. Xu, D. Comaniciu, and G. Hamarneh, “Combo loss: Handling input and output imbalance in multi-organ segmentation,” Comput Med Imaging Graph, vol. 75, pp. 24–33, 2019.
  20. D. Karimi and S. E. Salcudean, “Reducing the hausdorff distance in medical image segmentation with convolutional neural networks,” IEEE Trans. Medical Imaging, vol. 39, no. 2, pp. 499–513, 2020.
  21. H. Kervadec, J. Bouchtiba, C. Desrosiers, E. Granger, J. Dolz, and I. B. Ayed, “Boundary loss for highly unbalanced segmentation,” Medical Image Anal., vol. 67, p. 101851, 2021.
  22. K. Zhang, Y. Zhang, and H.-D. Cheng, “Crackgan: Pavement crack detection using partially accurate ground truths based on generative adversarial learning,” TITS, vol. 22, no. 2, pp. 1306–1319, 2020.
  23. A. Rezaie, R. Achanta, M. Godio, and K. Beyer, “Comparison of crack segmentation using digital image correlation measurements and deep learning,” Construction and Building Materials, vol. 261, no. 20, p. 120474, 2020.
  24. H. Chen, Y. Su, and W. He, “Automatic crack segmentation using deep high-resolution representation learning,” Applied Optics, vol. 60, no. 21, pp. 6080–6090, 2021.
  25. Y. Liu, J. Yao, X. Lu, R. Xie, and L. Li, “Deepcrack: A deep hierarchical feature learning architecture for crack segmentation,” Neurocomputing, vol. 338, pp. 139–153, 2019.
  26. H. Liu, X. Miao, C. Mertz, C. Xu, and H. Kong, “Crackformer: Transformer network for fine-grained crack detection,” in ICCV, 2021.
  27. L. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, no. 4, pp. 834–848, 2018.
  28. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in MICCAI, 2015.
  29. F. Yang, L. Zhang, S. Yu, D. Prokhorov, X. Mei, and H. Ling, “Feature pyramid and hierarchical boosting network for pavement crack detection,” TITS, vol. 21, no. 4, pp. 1525–1535, 2020.
  30. Z. Wang, J. Chen, and S. C. H. Hoi, “Deep learning for image super-resolution: A survey,” TPAMI, vol. 43, no. 10, pp. 3365–3387, 2021.
  31. K. Zhang, L. V. Gool, and R. Timofte, “Deep unfolding network for image super-resolution,” in CVPR, 2020.
  32. J. Lee, J. Park, K. Lee, J. Min, G. Kim, B. Lee, B. Ku, D. K. Han, and H. Ko, “FBRNN: feedback recurrent neural network for extreme image super-resolution,” in CVPR Workshop, 2020.
  33. L. Lu, W. Li, X. Tao, J. Lu, and J. Jia, “MASA-SR: matching acceleration and spatial adaptation for reference-based image super-resolution,” in CVPR, 2021.
  34. J. W. Soh, S. Cho, and N. I. Cho, “Meta-transfer learning for zero-shot super-resolution,” in CVPR, 2020.
  35. K. Zhang, J. L. L. V. Gool, and R. Timofte, “Designing a practical degradation model for deep blind image super-resolution,” in ICCV, 2021.
  36. S. A. Hussein, T. Tirer, and R. Giryes, “Correction filter for single image super-resolution: Robustifying off-the-shelf deep super-resolvers,” in CVPR, 2020.
  37. J. Gu, H. Lu, W. Zuo, and C. Dong, “Blind super-resolution with iterative kernel correction,” in CVPR, 2019.
  38. L. Wang, Y. Wang, X. Dong, Q. Xu, J. Yang, W. An, and Y. Guo, “Unsupervised degradation representation learning for blind super-resolution,” in CVPR, 2021.
  39. Y. Guo, J. Chen, J. Wang, Q. Chen, J. Cao, Z. Deng, Y. Xu, and M. Tan, “Closed-loop matters: Dual regression networks for single image super-resolution,” in CVPR, 2020.
  40. S. Y. Kim, H. Sim, and M. Kim, “Koalanet: Blind super-resolution using kernel-oriented adaptive local adjustment,” in CVPR, 2021.
  41. T. Yoshida, Y. Kondo, T. Maeda, K. Akita, and N. Ukita, “Kernelized back-projection networks for blind super resolution,” arXiv, 2023.
  42. Z. Hao, Y. Liu, H. Qin, J. Yan, X. Li, and X. Hu, “Scale-aware face detection,” in CVPR, 2017.
  43. Z. Guo, G. Wu, X. Song, W. Yuan, Q. Chen, H. Zhang, X. Shi, M. Xu, Y. Xu, R. Shibasaki, and X. Shao, “Super-resolution integrated building semantic segmentation for multi-source remote sensing imagery,” IEEE Access, vol. 7, pp. 99 381–99 397, 2019.
  44. D. Choi, J. H. Choi, J. W. Choi, and B. C. Song, “Sharpness enhancement and super-resolution of around-view monitor images,” IEEE Trans. Intell. Transp. Syst., vol. 19, no. 8, pp. 2650–2662, 2018.
  45. M. Singh, S. Nagpal, R. Singh, and M. Vatsa, “Dual directed capsule network for very low resolution image recognition,” in ICCV, 2019.
  46. Y. Zhang, Y. Bai, M. Ding, S. Xu, and B. Ghanem, “Kgsnet: Key-point-guided super-resolution network for pedestrian detection in the wild,” IEEE Trans. Neural Networks Learn. Syst., vol. 32, no. 5, pp. 2251–2265, 2021.
  47. K. Akita, M. Haris, and N. Ukita, “Region-dependent scale proposals for super-resolution in object detection,” in IPAS, 2020.
  48. M. Haris, G. Shakhnarovich, and N. Ukita, “Task-driven super resolution: Object detection in low-resolution images,” in ICONIP, 2021.
  49. ——, “Deep back-projection networks for super-resolution,” in CVPR, 2018.
  50. H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia, “Pyramid scene parsing network,” in CVPR, 2017.
  51. Y. Yuan, X. Chen, and J. Wang, “Object-contextual representations for semantic segmentation,” in ECCV, 2020, https://github.com/openseg-group/openseg.pytorch/blob/master/MODEL_ZOO.md.
  52. J. Ma, J. Chen, M. Ng, R. Huang, Y. Li, C. Li, X. Yang, and A. L. Martel, “Loss odyssey in medical image segmentation,” Medical Image Anal., vol. 71, p. 102035, 2021.
  53. S. Ryou, S. Jeong, and P. Perona, “Anchor loss: Modulating loss scale based on prediction difficulty,” in ICCV, 2019.
  54. DeepMind, “Surface distance metrics,” https://github.com/deepmind/surface-distance.
  55. X. Wang, K. Yu, C. Dong, and C. C. Loy, “Recovering realistic texture in image super-resolution by deep spatial feature transform,” in CVPR, 2018.
  56. E. Agustsson and R. Timofte, “Ntire 2017 challenge on single image super-resolution: Dataset and study,” in CVPRW, 2017.
  57. R. Timofte, E. Agustsson, L. Van Gool, M.-H. Yang, and L. Zhang, “Ntire 2017 challenge on single image super-resolution: Methods and results,” in CVPRW, 2017.
  58. J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in CVPR, 2009.
  59. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in ICLR, 2015.
  60. “Torchvision.models,” https://pytorch.org/vision/stable/models.html.
  61. D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in ICLR, 2015.
  62. Khanhha, “Crack segmentation,” 2020, https://github.com/khanhha/crack_segmentation.
  63. L. Zhang, F. Yang, Y. D. Zhang, and Y. J. Zhu, “Road crack detection using deep convolutional neural network,” in ICIP, 2016.
  64. M. Eisenbach, R. Stricker, D. Seichter, K. Amende, K. Debes, M. Sesselmann, D. Ebersbach, U. Stoeckert, and H.-M. Gross, “How to get pavement distress detection ready for deep learning? a systematic approach.” in IJCNN, 2017.
  65. Y. Shi, L. Cui, Z. Qi, F. Meng, and Z. Chen, “Automatic road crack detection using random structured forests,” TITS, vol. 17, no. 12, pp. 3434–3445, 2016.
  66. R. Amhazand, S. Chambon, J. Idier, and V. Baltazart, “Automatic crack detection on two-dimensional pavement images: An algorithm based on minimal path selection,” TITS, vol. 17, no. 10, pp. 2718–2729, 2016.
  67. Q. Zou, Y. Cao, Q. Li, Q. Mao, and S. Wang, “Cracktree: Automatic crack detection from pavement images,” Pattern Recognition Letters, vol. 33, no. 3, pp. 227–238, 2012.
  68. L. Yang, B. Li, W. Li, L. Zhaoming, G. Yang, and J. Xiao, “Deep concrete inspection using unmanned aerial vehicle towards cssc database,” in IROS, 2017.
  69. Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
  70. M. Haris, G. Shakhnarovich, and N. Ukita, “Space-time-aware multi-resolution video enhancement,” in CVPR, 2020.
Citations (4)

Summary

  • The paper proposes a deep neural network framework that jointly learns blind super-resolution and crack segmentation end-to-end for images with unknown degradations.
  • Experimental evaluation shows the joint learning method significantly outperforms independent SR and segmentation approaches, improving metrics like IoU by approximately 10%.
  • This joint learning framework has significant implications for automated visual inspection systems, enabling improved crack detection in real-world degraded images for infrastructure monitoring.

Joint Learning of Blind Super-Resolution and Crack Segmentation for Realistic Degraded Images

This paper addresses the complex challenge of segmenting cracks in realistically-degraded images by proposing a method that jointly learns Blind Super-Resolution (SR) and Crack Segmentation. The significant contributions of this study lie in the emphasis on practical scenarios where images undergo unknown degradations such as blurring due to motion or out-of-focus conditions, which are commonplace in automatic inspection from vehicles like drones and cars.

Key Contributions and Methodology

  1. Joint SR and Segmentation Framework: The study advances the current methodologies by integrating a deep neural network-based Super-Resolution with binary segmentation, allowing for end-to-end training. This integration extends from non-blind to blind SR, optimizing for segmentation performance even in scenarios with unknown blur.
  2. Mutual Optimization Paths: To further enhance the network's performance, additional paths are introduced that optimize SR specifically for segmentation tasks. These paths ensure a two-way beneficial relationship between SR and segmentation processes.
  3. Boundary Combo Loss: A novel Boundary Combo (BC) loss function is introduced. This loss function is designed to robustly handle class-imbalance problems by providing global constraints on the detection of fine boundaries, tailored to improve the detection of thin cracks amidst a predominance of non-crack pixels.
  4. Segmentation-aware SR-loss Weights: By incorporating weights based on segmentation awareness, some challenges such as gradient vanishing during training can be mitigated. These weights help in directing the optimization process to better support both SR and segmentation simultaneously.
  5. Blur-reflected Task Learning: The method proposes leveraging the blur characteristics estimated during SR for improving segmentation robustness. By skipping blur influences to the segmentation network, finer segmentation outputs are achieved.

Experimental Evaluation

The experimental component of this work underscores its comparative superiority to state-of-the-art segmentation methods. The joint learning framework consistently outperforms independent SR and segmentation solutions. Metrics such as IoU, PSNR, and Structural Similarity (SSIM) highlight the efficacy of the proposed method, with impressive improvements like an increase in IoU of approximately 10% over traditional non-blind SR-based methods.

Implications and Future Work

The implications of this study are substantial for automated visual inspection systems, especially in infrastructure monitoring. Real-world applications include detecting deteriorations in civil infrastructure like bridges and roads, where frequent manual inspection is impractical.

Theoretically, this paper adds to the discourse on multi-task learning, demonstrating how intricately connected visual tasks can mutually benefit from a unified learning process, especially under real-world degradation effects.

Future considerations might explore extending this framework to include video sequences, thus leveraging temporal data for improved SR and segmentation results. Moreover, addressing even more severe degradations and testing on a broader array of real-world conditions can further elevate the applicability of this research.

In summary, this paper presents a robust and practical approach to handling an intricate problem that combines advanced techniques in super-resolution and segmentation to tackle real-world challenges in automated inspection and monitoring.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.