Non-Parallel Text Style Transfer using Self-Attentional Discriminator as Supervisor

Author(s):  
Kuan Feng ◽  
Yanmin Zhu ◽  
Jiadi Yu
Keyword(s):  
Author(s):  
Di Yin ◽  
Shujian Huang ◽  
Xin-Yu Dai ◽  
Jiajun Chen

Text style transfer aims to rephrase a given sentence into a different style without changing its original content. Since parallel corpora (i.e. sentence pairs with the same content but different styles) are usually unavailable, most previous works solely guide the transfer process with distributional information, i.e. using style-related classifiers or language models, which neglect the correspondence of instances, leading to poor transfer performance, especially for the content preservation. In this paper, we propose making partial comparisons to explicitly model the content and style correspondence of instances, respectively. To train the partial comparators, we propose methods to extract partial-parallel training instances automatically from the non-parallel data, and to further enhance the training process by using data augmentation. We perform experiments that compare our method to other existing approaches on two review datasets. Both automatic and manual evaluations show that our approach can significantly improve the performance of existing adversarial methods, and outperforms most state-of-the-art models. Our code and data will be available on Github.


2019 ◽  
Author(s):  
Vineet John ◽  
Lili Mou ◽  
Hareesh Bahuleyan ◽  
Olga Vechtomova

2021 ◽  
Vol 12 (3) ◽  
pp. 1-16
Author(s):  
Yukai Shi ◽  
Sen Zhang ◽  
Chenxing Zhou ◽  
Xiaodan Liang ◽  
Xiaojun Yang ◽  
...  

Non-parallel text style transfer has attracted increasing research interests in recent years. Despite successes in transferring the style based on the encoder-decoder framework, current approaches still lack the ability to preserve the content and even logic of original sentences, mainly due to the large unconstrained model space or too simplified assumptions on latent embedding space. Since language itself is an intelligent product of humans with certain grammars and has a limited rule-based model space by its nature, relieving this problem requires reconciling the model capacity of deep neural networks with the intrinsic model constraints from human linguistic rules. To this end, we propose a method called Graph Transformer–based Auto-Encoder, which models a sentence as a linguistic graph and performs feature extraction and style transfer at the graph level, to maximally retain the content and the linguistic structure of original sentences. Quantitative experiment results on three non-parallel text style transfer tasks show that our model outperforms state-of-the-art methods in content preservation, while achieving comparable performance on transfer accuracy and sentence naturalness.


2019 ◽  
Author(s):  
Utsav Krishnan ◽  
Akshal Sharma ◽  
Pratik Chattopadhyay

2020 ◽  
Vol 499 (3) ◽  
pp. 4054-4067
Author(s):  
Steven Cunnington ◽  
Stefano Camera ◽  
Alkistis Pourtsidou

ABSTRACT Potential evidence for primordial non-Gaussianity (PNG) is expected to lie in the largest scales mapped by cosmological surveys. Forthcoming 21 cm intensity mapping experiments will aim to probe these scales by surveying neutral hydrogen (H i) within galaxies. However, foreground signals dominate the 21 cm emission, meaning foreground cleaning is required to recover the cosmological signal. The effect this has is to damp the H i power spectrum on the largest scales, especially along the line of sight. Whilst there is agreement that this contamination is potentially problematic for probing PNG, it is yet to be fully explored and quantified. In this work, we carry out the first forecasts on fNL that incorporate simulated foreground maps that are removed using techniques employed in real data. Using an Monte Carlo Markov Chain analysis on an SKA1-MID-like survey, we demonstrate that foreground cleaned data recovers biased values [$f_{\rm NL}= -102.1_{-7.96}^{+8.39}$ (68 per cent CL)] on our fNL = 0 fiducial input. Introducing a model with fixed parameters for the foreground contamination allows us to recover unbiased results ($f_{\rm NL}= -2.94_{-11.9}^{+11.4}$). However, it is not clear that we will have sufficient understanding of foreground contamination to allow for such rigid models. Treating the main parameter $k_\parallel ^\text{FG}$ in our foreground model as a nuisance parameter and marginalizing over it, still recovers unbiased results but at the expense of larger errors ($f_{\rm NL}= 0.75^{+40.2}_{-44.5}$), which can only be reduced by imposing the Planck 2018 prior. Our results show that significant progress on understanding and controlling foreground removal effects is necessary for studying PNG with H i intensity mapping.


Sign in / Sign up

Export Citation Format

Share Document