scholarly journals Devnagari Handwritten Characters Image Super-Resolution based on Enhanced SRGAN

2021 ◽  
Vol 16 (1) ◽  
pp. 103-109
Author(s):  
Prasiddha Siwakoti ◽  
Sharad Kumar Ghimire

The difficulty in machine learning-based image super-resolution is to generate high-frequency component in an image without introducing any artifacts. In this paper, Devnagari handwritten characters image using a generative adversarial network with a classifier is generated in high-resolution which is also classifiable. The generator architecture is modified by removing all batch normalization layers in generator architecture with a residual in residual dense block. Batch normalization is removed because it produces unwanted artifacts in the generated images. A Devnagari handwritten characters classifier is built using CNN. The classifier is used in the network to calculate the content loss. The adversarial loss is obtained from the GAN architecture and both of the losses are added to obtain total loss. Generated HR images is validated using six different evaluation metrics among which MSE, PSNR determines pixel-wise difference and SSIM compares images perceptually. Similarly, FID is used to measure the statistical similarity between the batch of generated images and its original batch. Finally, the Gradient similarity is used to assess the quality of the generated image. From the experimental results, we obtain MSE, PSNR and SSIM as 0.0507, 12.95(dB) and 0.8172 respectively. Similarly, the FID value obtained was 27.5 with the classification accuracy of image data of 98%. The gradient similarity between the generated image and the ground truth obtained was 0.9124.

2021 ◽  
Author(s):  
Jiaoyue Li ◽  
Weifeng Liu ◽  
Kai Zhang ◽  
Baodi Liu

Remote sensing image super-resolution (SR) plays an essential role in many remote sensing applications. Recently, remote sensing image super-resolution methods based on deep learning have shown remarkable performance. However, directly utilizing the deep learning methods becomes helpless to recover the remote sensing images with a large number of complex objectives or scene. So we propose an edge-based dense connection generative adversarial network (SREDGAN), which minimizes the edge differences between the generated image and its corresponding ground truth. Experimental results on NWPU-VHR-10 and UCAS-AOD datasets demonstrate that our method improves 1.92 and 0.045 in PSNR and SSIM compared with SRGAN, respectively.


2021 ◽  
Vol 12 (6) ◽  
pp. 1-20
Author(s):  
Fayaz Ali Dharejo ◽  
Farah Deeba ◽  
Yuanchun Zhou ◽  
Bhagwan Das ◽  
Munsif Ali Jatoi ◽  
...  

Single Image Super-resolution (SISR) produces high-resolution images with fine spatial resolutions from a remotely sensed image with low spatial resolution. Recently, deep learning and generative adversarial networks (GANs) have made breakthroughs for the challenging task of single image super-resolution (SISR) . However, the generated image still suffers from undesirable artifacts such as the absence of texture-feature representation and high-frequency information. We propose a frequency domain-based spatio-temporal remote sensing single image super-resolution technique to reconstruct the HR image combined with generative adversarial networks (GANs) on various frequency bands (TWIST-GAN). We have introduced a new method incorporating Wavelet Transform (WT) characteristics and transferred generative adversarial network. The LR image has been split into various frequency bands by using the WT, whereas the transfer generative adversarial network predicts high-frequency components via a proposed architecture. Finally, the inverse transfer of wavelets produces a reconstructed image with super-resolution. The model is first trained on an external DIV2 K dataset and validated with the UC Merced Landsat remote sensing dataset and Set14 with each image size of 256 × 256. Following that, transferred GANs are used to process spatio-temporal remote sensing images in order to minimize computation cost differences and improve texture information. The findings are compared qualitatively and qualitatively with the current state-of-art approaches. In addition, we saved about 43% of the GPU memory during training and accelerated the execution of our simplified version by eliminating batch normalization layers.


2021 ◽  
Vol 36 (5) ◽  
pp. 705-712
Author(s):  
Zong-hang CHEN ◽  
◽  
Hai-long HU ◽  
Jian-min YAO ◽  
Qun YAN ◽  
...  

Symmetry ◽  
2020 ◽  
Vol 12 (3) ◽  
pp. 449 ◽  
Author(s):  
Can Li ◽  
Liejun Wang ◽  
Shuli Cheng ◽  
Naixiang Ao

In recent years, the common algorithms for image super-resolution based on deep learning have been increasingly successful, but there is still a large gap between the results generated by each algorithm and the ground-truth. Even some algorithms that are dedicated to image perception produce more textures that do not exist in the original image, and these artefacts also affect the visual perceptual quality of the image. We believe that in the existing perceptual-based image super-resolution algorithm, it is necessary to consider Super-Resolution (SR) image quality, which can restore the important structural parts of the original picture. This paper mainly improves the Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN) algorithm in the following aspects: adding a shallow network structure, adding the dual attention mechanism in the generator and the discriminator, including the second-order channel mechanism and spatial attention mechanism and optimizing perceptual loss by adding second-order covariance normalization at the end of feature extractor. The results of this paper ensure image perceptual quality while reducing image distortion and artefacts, improving the perceived similarity of images and making the images more in line with human visual perception.


Sign in / Sign up

Export Citation Format

Share Document