Pixel-Level Character Motion Style Transfer using Conditional Adversarial Networks

Author(s):  
Dong Hu ◽  
Shu-Juan Peng ◽  
Xin Liu
Sensors ◽  
2020 ◽  
Vol 20 (9) ◽  
pp. 2605 ◽  
Author(s):  
Rafael Anicet Zanini ◽  
Esther Luna Colombini

This paper proposes two new data augmentation approaches based on Deep Convolutional Generative Adversarial Networks (DCGANs) and Style Transfer for augmenting Parkinson’s Disease (PD) electromyography (EMG) signals. The experimental results indicate that the proposed models can adapt to different frequencies and amplitudes of tremor, simulating each patient’s tremor patterns and extending them to different sets of movement protocols. Therefore, one could use these models for extending the existing patient dataset and generating tremor simulations for validating treatment approaches on different movement scenarios.


2019 ◽  
Vol 1314 ◽  
pp. 012191
Author(s):  
Jie Wu ◽  
Jin Duan ◽  
Jinqiang Yu ◽  
Haodong Shi ◽  
Yingchao Li

Author(s):  
Chien-Yu Lu ◽  
Min-Xin Xue ◽  
Chia-Che Chang ◽  
Che-Rung Lee ◽  
Li Su

Style transfer of polyphonic music recordings is a challenging task when considering the modeling of diverse, imaginative, and reasonable music pieces in the style different from their original one. To achieve this, learning stable multi-modal representations for both domain-variant (i.e., style) and domaininvariant (i.e., content) information of music in an unsupervised manner is critical. In this paper, we propose an unsupervised music style transfer method without the need for parallel data. Besides, to characterize the multi-modal distribution of music pieces, we employ the Multi-modal Unsupervised Image-to-Image Translation (MUNIT) framework in the proposed system. This allows one to generate diverse outputs from the learned latent distributions representing contents and styles. Moreover, to better capture the granularity of sound, such as the perceptual dimensions of timbre and the nuance in instrument-specific performance, cognitively plausible features including mel-frequency cepstral coefficients (MFCC), spectral difference, and spectral envelope, are combined with the widely-used mel-spectrogram into a timbreenhanced multi-channel input representation. The Relativistic average Generative Adversarial Networks (RaGAN) is also utilized to achieve fast convergence and high stability. We conduct experiments on bilateral style transfer tasks among three different genres, namely piano solo, guitar solo, and string quartet. Results demonstrate the advantages of the proposed method in music style transfer with improved sound quality and in allowing users to manipulate the output.


2020 ◽  
Vol 27 (3) ◽  
pp. 54-65
Author(s):  
Yaochen Li ◽  
Xiao Wu ◽  
Danhui Lu ◽  
Ling Li ◽  
Yuehu Liu ◽  
...  

2021 ◽  
Vol 2021 ◽  
pp. 1-17
Author(s):  
Huaijun Wang ◽  
Dandan Du ◽  
Junhuai Li ◽  
Wenchao Ji ◽  
Lei Yu

Motion capture technology plays an important role in the production field of film and television, animation, etc. In order to reduce the cost of data acquisition and improve the reuse rate of motion capture data and the effect of movement style migration, the synthesis technology of motion capture data in human movement has become a research hotspot in this field. In this paper, kinematic constraints (KC) and cyclic consistency (CC) network are employed to study the methods of kinematic style migration. Firstly, cycle-consistent adversarial network (CCycleGAN) is constructed, and the motion style migration network based on convolutional self-encoder is used as a generator to establish the cyclic consistent constraint between the generated motion and the content motion, so as to improve the action consistency between the generated motion and the content motion and eliminate the lag phenomenon of the generated motion. Then, kinematic constraints are introduced to normalize the movement generation, so as to solve the problems such as jitter and sliding step in the movement style migration results. Experimental results show that the generated motion of the cyclic consistent style transfer method with kinematic constraints is more similar to the style of style motion, which improves the effect of motion style transfer.


Symmetry ◽  
2018 ◽  
Vol 10 (7) ◽  
pp. 294 ◽  
Author(s):  
Xiaowei Xue ◽  
Chunxue Wu ◽  
Ze Sun ◽  
Yan Wu ◽  
Neal Xiong

A 3D city model is critical for the construction of a digital city. One of the methods of building a 3D city model is tilt photogrammetry. In this method, oblique photography is crucial for generating the model because the visual quality of photography directly impacts the model’s visual effect. Yet, sometimes, oblique photography does not have good visual quality due to a bad season or defective photographic equipment. For example, for oblique photography taken in winter, vegetation is brown. If this photography is employed to generate the 3D model, the result would be bad visually. Yet, common methods for vegetation greening in oblique photography rely on the assistance of the infrared band, which is not available sometimes. Thus, a method for vegetation greening in winter oblique photography without the infrared band is required, which is proposed in this paper. The method was inspired by the work on CycleGAN (Cycle-consistence Adversarial Networks). In brief, the problem of turning vegetation green in winter oblique photography is considered as a style transfer problem. Summer oblique photography generally has green vegetation. By applying CycleGAN, winter oblique photography can be transferred to summer oblique photography, and the vegetation can turn green. Yet, due to the existence of “checkerboard artifacts”, the original result cannot be applied for real production. To reduce artifacts, the generator of CycleGAN is modified. As the final results suggest, the proposed method unlocks the bottleneck of vegetation greening when the infrared band is not available and artifacts are reduced.


2020 ◽  
Vol 2020 ◽  
pp. 1-12
Author(s):  
Hyunhee Lee ◽  
Jaechoon Jo ◽  
Heuiseok Lim

Due to institutional and privacy issues, medical imaging researches are confronted with serious data scarcity. Image synthesis using generative adversarial networks provides a generic solution to the lack of medical imaging data. We synthesize high-quality brain tumor-segmented MR images, which consists of two tasks: synthesis and segmentation. We performed experiments with two different generative networks, the first using the ResNet model, which has significant advantages of style transfer, and the second, the U-Net model, one of the most powerful models for segmentation. We compare the performance of each model and propose a more robust model for synthesizing brain tumor-segmented MR images. Although ResNet produced better-quality images than did U-Net for the same samples, it used a great deal of memory and took much longer to train. U-Net, meanwhile, segmented the brain tumors more accurately than did ResNet.


2020 ◽  
Vol 39 (4) ◽  
Author(s):  
Kfir Aberman ◽  
Yijia Weng ◽  
Dani Lischinski ◽  
Daniel Cohen-Or ◽  
Baoquan Chen
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document