Remote sensing image dehazing (RSID) is an extremely challenging problem due to the irregular and nonuniform distribution of haze. The existing RSID methods achieve excellent performance using deep learning; however,… Click to show full abstract
Remote sensing image dehazing (RSID) is an extremely challenging problem due to the irregular and nonuniform distribution of haze. The existing RSID methods achieve excellent performance using deep learning; however, relying on paired synthetic data is limited to their generality in various haze distribution. In this letter, we present a memory-oriented generative adversarial network (MO-GAN), which tries to capture the desired hazy features in an unpaired learning manner toward single RSID. For better extracting the haze-relevant features, a novel multistage attentive-recurrent memory module is designed to guide an autoencoder neural network, which can record the various appearances of haze distribution at different stages. To well differentiate fake images from real ones, a dual region discriminator is constructed to handle spatially varying haze densities in global and local regions. Extensive experiments demonstrate that our designed MO-GAN outperforms the recent comparing approaches on the various frequently used datasets, especially in real world nonuniform haze conditions. The source code is released in https://github.com/cxtalk/MO-GAN.
               
Click one of the above tabs to view related content.