scholarly journals Dual Encoder-Decoder Based Generative Adversarial Networks for Disentangled Facial Representation Learning

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 130159-130171 ◽  
Author(s):  
Cong Hu ◽  
Zhenhua Feng ◽  
Xiaojun Wu ◽  
Josef Kittler
2021 ◽  
Vol 7 (2) ◽  
pp. 755-758
Author(s):  
Daniel Wulff ◽  
Mohamad Mehdi ◽  
Floris Ernst ◽  
Jannis Hagenah

Abstract Data augmentation is a common method to make deep learning assessible on limited data sets. However, classical image augmentation methods result in highly unrealistic images on ultrasound data. Another approach is to utilize learning-based augmentation methods, e.g. based on variational autoencoders or generative adversarial networks. However, a large amount of data is necessary to train these models, which is typically not available in scenarios where data augmentation is needed. One solution for this problem could be a transfer of augmentation models between different medical imaging data sets. In this work, we present a qualitative study of the cross data set generalization performance of different learning-based augmentation methods for ultrasound image data. We could show that knowledge transfer is possible in ultrasound image augmentation and that the augmentation partially results in semantically meaningful transfers of structures, e.g. vessels, across domains.


Author(s):  
P. J. Soto ◽  
J. D. Bermudez ◽  
P. N. Happ ◽  
R. Q. Feitosa

<p><strong>Abstract.</strong> This work aims at investigating unsupervised and semi-supervised representation learning methods based on generative adversarial networks for remote sensing scene classification. The work introduces a novel approach, which consists in a semi-supervised extension of a prior unsupervised method, known as MARTA-GAN. The proposed approach was compared experimentally with two baselines upon two public datasets, <i>UC-MERCED</i> and <i>NWPU-RESISC45</i>. The experiments assessed the performance of each approach under different amounts of labeled data. The impact of fine-tuning was also investigated. The proposed method delivered in our analysis the best overall accuracy under scarce labeled samples, both in terms of absolute value and in terms of variability across multiple runs.</p>


2020 ◽  
Vol 8 ◽  
Author(s):  
Sirajul Salekin ◽  
Milad Mostavi ◽  
Yu-Chiao Chiu ◽  
Yidong Chen ◽  
Jianqiu Zhang ◽  
...  

Author(s):  
Shicheng Cui ◽  
Qianmu Li ◽  
Shu-Ching Chen

Abstract The analytics on graph-structured data in cyber spaces has advanced many human-centered computing technologies. However, if only utilizing the structural properties, we might be prohibited from unraveling unknown social relations of nodes especially in the structureless networked systems. Up-to-date ways to unfold latent relationships from graph-structured data are network representation learning (NRL) techniques, but it is difficult for most existing ones to deal with the network-structureless situations due to the fact that they largely depend on the observed connections. With the ever-broader spectrum of human-centered networked systems, large quantities of textual information have been generated and collected from social and physical spaces, which may provide the clues of hidden social relations. In order to discover latent social relations from the accompanied text resources, this paper attempts to bridge the gap between text data and graph-structured data so that the textual information can be encoded to substitute for those incomplete structural information. Generative adversarial networks (GANs) are employed in the cross-modal framework to make the transformed data indistinguishable in graph-domain space and also capable of depicting structure-aware relationships with network homophily. Experiments conducted on three text-based network benchmarks demonstrate that our approach can reveal more realistic social relations from text-domain information compared against the state-of-the-art baselines.


Sign in / Sign up

Export Citation Format

Share Document