A novel image super-resolution reconstruction algorithm based on improved GANs and gradient penalty

2019 ◽  
Vol 12 (3) ◽  
pp. 400-413 ◽  
Author(s):  
Shuangshuang Liu ◽  
Xiaoling Li

Purpose Conventional image super-resolution reconstruction by the conventional deep learning architectures suffers from the problems of hard training and gradient disappearing. In order to solve such problems, the purpose of this paper is to propose a novel image super-resolution algorithm based on improved generative adversarial networks (GANs) with Wasserstein distance and gradient penalty. Design/methodology/approach The proposed algorithm first introduces the conventional GANs architecture, the Wasserstein distance and the gradient penalty for the task of image super-resolution reconstruction (SRWGANs-GP). In addition, a novel perceptual loss function is designed for the SRWGANs-GP to meet the task of image super-resolution reconstruction. The content loss is extracted from the deep model’s feature maps, and such features are introduced to calculate mean square error (MSE) for the loss calculation of generators. Findings To validate the effectiveness and feasibility of the proposed algorithm, a lot of compared experiments are applied on three common data sets, i.e. Set5, Set14 and BSD100. Experimental results have shown that the proposed SRWGANs-GP architecture has a stable error gradient and iteratively convergence. Compared with the baseline deep models, the proposed GANs models have a significant improvement on performance and efficiency for image super-resolution reconstruction. The MSE calculated by the deep model’s feature maps gives more advantages for constructing contour and texture. Originality/value Compared with the state-of-the-art algorithms, the proposed algorithm obtains a better performance on image super-resolution and better reconstruction results on contour and texture.

2021 ◽  
Vol 12 (6) ◽  
pp. 1-20
Author(s):  
Fayaz Ali Dharejo ◽  
Farah Deeba ◽  
Yuanchun Zhou ◽  
Bhagwan Das ◽  
Munsif Ali Jatoi ◽  
...  

Single Image Super-resolution (SISR) produces high-resolution images with fine spatial resolutions from a remotely sensed image with low spatial resolution. Recently, deep learning and generative adversarial networks (GANs) have made breakthroughs for the challenging task of single image super-resolution (SISR) . However, the generated image still suffers from undesirable artifacts such as the absence of texture-feature representation and high-frequency information. We propose a frequency domain-based spatio-temporal remote sensing single image super-resolution technique to reconstruct the HR image combined with generative adversarial networks (GANs) on various frequency bands (TWIST-GAN). We have introduced a new method incorporating Wavelet Transform (WT) characteristics and transferred generative adversarial network. The LR image has been split into various frequency bands by using the WT, whereas the transfer generative adversarial network predicts high-frequency components via a proposed architecture. Finally, the inverse transfer of wavelets produces a reconstructed image with super-resolution. The model is first trained on an external DIV2 K dataset and validated with the UC Merced Landsat remote sensing dataset and Set14 with each image size of 256 × 256. Following that, transferred GANs are used to process spatio-temporal remote sensing images in order to minimize computation cost differences and improve texture information. The findings are compared qualitatively and qualitatively with the current state-of-art approaches. In addition, we saved about 43% of the GPU memory during training and accelerated the execution of our simplified version by eliminating batch normalization layers.


2020 ◽  
Vol 125 ◽  
pp. 83-91 ◽  
Author(s):  
Tzu-An Song ◽  
Samadrita Roy Chowdhury ◽  
Fan Yang ◽  
Joyita Dutta

2020 ◽  
Vol 34 (04) ◽  
pp. 3121-3129 ◽  
Author(s):  
Shady Abu Hussein ◽  
Tom Tirer ◽  
Raja Giryes

In the recent years, there has been a significant improvement in the quality of samples produced by (deep) generative models such as variational auto-encoders and generative adversarial networks. However, the representation capabilities of these methods still do not capture the full distribution for complex classes of images, such as human faces. This deficiency has been clearly observed in previous works that use pre-trained generative models to solve imaging inverse problems. In this paper, we suggest to mitigate the limited representation capabilities of generators by making them image-adaptive and enforcing compliance of the restoration with the observations via back-projections. We empirically demonstrate the advantages of our proposed approach for image super-resolution and compressed sensing.


Sign in / Sign up

Export Citation Format

Share Document