Desnowgan: An efficient single image snow removal framework using cross-resolution lateral connection and gans

Da Wei Jaw, Shih Chia Huang*, Sy Yen Kuo

*Corresponding author for this work

Research output: Contribution to journalJournal Article peer-review

65 Scopus citations

Abstract

In this paper, we present a simple, efficient, and highly modularized network architecture for single-image snow-removal. To address the challenging snow-removal problem in terms of network interpretability and computational complexity, we employ a pyramidal hierarchical design with lateral connections across different resolutions. This design enables us to incorporate high-level semantic features with other feature maps at different scales to enrich location information and reduce computational time. In addition, a refinement stage based on recently introduced generative adversarial networks (GANs) is proposed to further improve the visual quality of the resulting snow-removed images and make a refined image and a clean image indistinguishable by a computer vision algorithm to avoid the potential perturbations of machine interpretation. Finally, atrous spatial pyramid pooling (ASPP) is adopted to probe features at multiple scales and further boost the performance. The proposed DesnowGAN (DS-GAN) performs significantly better than state-of-the-art methods quantitatively and qualitatively on the Snow100K dataset.

Original languageEnglish
Article number9119486
Pages (from-to)1342-1350
Number of pages9
JournalIEEE Transactions on Circuits and Systems for Video Technology
Volume31
Issue number4
DOIs
StatePublished - 04 2021
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 1991-2012 IEEE.

Keywords

  • deep learning
  • generative adversarial networks
  • image enhancement
  • image restoration
  • Snow Removal

Fingerprint

Dive into the research topics of 'Desnowgan: An efficient single image snow removal framework using cross-resolution lateral connection and gans'. Together they form a unique fingerprint.

Cite this