scholarly journals Guest Editorial: Generative Adversarial Networks for Computer Vision

2020 ◽  
Vol 128 (10-11) ◽  
pp. 2363-2365
Author(s):  
Jun-Yan Zhu ◽  
Hongsheng Li ◽  
Eli Shechtman ◽  
Ming-Yu Liu ◽  
Jan Kautz ◽  
...  
IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 14985-15006 ◽  
Author(s):  
Yang-Jie Cao ◽  
Li-Li Jia ◽  
Yong-Xia Chen ◽  
Nan Lin ◽  
Cong Yang ◽  
...  

2021 ◽  
Vol 2021 (1) ◽  
pp. 43-48
Author(s):  
Mekides Assefa Abebe

Exposure problems, due to standard camera sensor limitations, often lead to image quality degradations such as loss of details and change in color appearance. The quality degradations further hiders the performances of imaging and computer vision applications. Therefore, the reconstruction and enhancement of uderand over-exposed images is essential for various applications. Accordingly, an increasing number of conventional and deep learning reconstruction approaches have been introduced in recent years. Most conventional methods follow color imaging pipeline, which strongly emphasize on the reconstructed color and content accuracy. The deep learning (DL) approaches have conversely shown stronger capability on recovering lost details. However, the design of most DL architectures and objective functions don’t take color fidelity into consideration and, hence, the analysis of existing DL methods with respect to color and content fidelity will be pertinent. Accordingly, this work presents performance evaluation and results of recent DL based overexposure reconstruction solutions. For the evaluation, various datasets from related research domains were merged and two generative adversarial networks (GAN) based models were additionally adopted for tone mapping application scenario. Overall results show various limitations, mainly for severely over-exposed contents, and a promising potential for DL approaches, GAN, to reconstruct details and appearance.


2022 ◽  
pp. 98-110
Author(s):  
Md Fazle Rabby ◽  
Md Abdullah Al Momin ◽  
Xiali Hei

Generative adversarial networks have been a highly focused research topic in computer vision, especially in image synthesis and image-to-image translation. There are a lot of variations in generative nets, and different GANs are suitable for different applications. In this chapter, the authors investigated conditional generative adversarial networks to generate fake images, such as handwritten signatures. The authors demonstrated an implementation of conditional generative adversarial networks, which can generate fake handwritten signatures according to a condition vector tailored by humans.


2020 ◽  
Vol 1570 ◽  
pp. 012064
Author(s):  
Yunfei Li ◽  
Lidan Wang ◽  
Taixing Chen ◽  
Ziyuan Wang ◽  
Shukai Duan

2021 ◽  
Vol 54 (2) ◽  
pp. 1-38
Author(s):  
Zhengwei Wang ◽  
Qi She ◽  
Tomás E. Ward

Generative adversarial networks (GANs) have been extensively studied in the past few years. Arguably their most significant impact has been in the area of computer vision where great advances have been made in challenges such as plausible image generation, image-to-image translation, facial attribute manipulation, and similar domains. Despite the significant successes achieved to date, applying GANs to real-world problems still poses significant challenges, three of which we focus on here. These are as follows: (1) the generation of high quality images, (2) diversity of image generation, and (3) stabilizing training. Focusing on the degree to which popular GAN technologies have made progress against these challenges, we provide a detailed review of the state-of-the-art in GAN-related research in the published scientific literature. We further structure this review through a convenient taxonomy we have adopted based on variations in GAN architectures and loss functions. While several reviews for GANs have been presented to date, none have considered the status of this field based on their progress toward addressing practical challenges relevant to computer vision. Accordingly, we review and critically discuss the most popular architecture-variant, and loss-variant GANs, for tackling these challenges. Our objective is to provide an overview as well as a critical analysis of the status of GAN research in terms of relevant progress toward critical computer vision application requirements. As we do this we also discuss the most compelling applications in computer vision in which GANs have demonstrated considerable success along with some suggestions for future research directions. Codes related to the GAN-variants studied in this work is summarized on https://github.com/sheqi/GAN_Review.


Author(s):  
Judy Simon

Computer vision, also known as computational visual perception, is a branch of artificial intelligence that allows computers to interpret digital pictures and videos in a manner comparable to biological vision. It entails the development of techniques for simulating biological vision. The aim of computer vision is to extract more meaningful information from visual input than that of a biological vision. Computer vision is exploding due to the avalanche of data being produced today. Powerful generative models, such as Generative Adversarial Networks (GANs), are responsible for significant advances in the field of picture creation. The focus of this research is to concentrate on textual content descriptors in the images used by GANs to generate synthetic data from the MNIST dataset to either supplement or replace the original data while training classifiers. This can provide better performance than other traditional image enlarging procedures due to the good handling of synthetic data. It shows that training classifiers on synthetic data are as effective as training them on pure data alone, and it also reveals that, for small training data sets, supplementing the dataset by first training GANs on the data may lead to a significant increase in classifier performance.


2020 ◽  
Author(s):  
Vignesh Sampath ◽  
Iñaki Maurtua ◽  
Juan José Aguilar Martín ◽  
Aitor Gutierrez

Abstract Any computer vision application development starts off by acquiring images and data, then preprocessingand pattern recognition steps to perform a task. When the acquired image is highly imbalanced and notadequate, the desired task may not be achievable. Unfortunately, the occurrence of imbalance problems inacquired image datasets in certain complex real-world problems such as anomaly detection, emotionrecognition, medical image analysis, fraud detection, metallic surface defect detection, disaster prediction,etc., are inevitable. The performance of computer vision algorithms can significantly deteriorate when thetraining dataset is imbalanced. In recent years, Generative Adversarial Networks (GANs) have gainedimmense attention by researchers across a variety of application domains due to their capability to modelcomplex real-world image data. It is particularly important that GANs can not only be used to generatesynthetic images, but also its fascinating adversarial learning idea showed good potential in restoringbalance in imbalanced datasets.In this paper, we examine the most recent developments of GANs based techniques for addressingimbalance problems in image data. The real-world challenges and implementations of synthetic imagegeneration based on GANs are extensively covered in this survey. Our survey first introduces variousimbalance problems in computer vision tasks and its existing solutions, and then examine key conceptssuch as deep generative image models and GANs. After that, we propose taxonomy to summarize GANsbased techniques for addressing imbalance problems in computer vision tasks into three major categories:Image level imbalances in classification, object level imbalances in object detection and pixel levelimbalances in segmentation tasks. We elaborate the imbalance problems of each group, and furtherprovide GANs based solutions in each group. Readers will understand how GANs based techniques canhandle the problem of imbalances and boost performance of the computer vision algorithms.


Symmetry ◽  
2020 ◽  
Vol 12 (5) ◽  
pp. 689
Author(s):  
Qiang Li ◽  
Ling Gao

As a research field of symmetry journals, computer vision has received more and more attention. Person re-identification (re-ID) has become a research hotspot in computer vision. We focus on one-example person re-ID, where each person only has one labeled image in the dataset, and other images are unlabeled. There are two main challenges of the task, the insufficient labeled data, and the lack of labeled images cross-cameras. In dealing with the above issue, we propose a new one-example labeling scheme, which generates style-transferred images by CycleGAN (Cycle Generative Adversarial Networks) to ensure that for each person, there is one labeled image under each camera style. Then a self-learning framework is adopted, which iteratively train a CNN (Convolutional Neural Networks) model with labeled images and labeled style-transferred images, and mine the reliable images to assign a pseudo label. The experimental results prove that by integrating the camera style transferred images, we effectively expand the dataset, and the problem of low recognition rate caused by the lack of labeled pedestrian pictures across cameras is effectively solved. Notably, the rank-1 accuracy of our method outperforms the state-of-the-art method by 8.7 points on the Market-1501 dataset, and 6.3 points on the DukeMTMC-ReID dataset.


Sign in / Sign up

Export Citation Format

Share Document