Symmetrical Adversarial Training Network: A Novel Model for Text Generation

Author(s):  
Yongzhen Gao ◽  
ChongJun Wang
2019 ◽  
Author(s):  
Pei Ke ◽  
Fei Huang ◽  
Minlie Huang ◽  
Xiaoyan Zhu

Author(s):  
Jianing Li ◽  
Yanyan Lan ◽  
Jiafeng Guo ◽  
Jun Xu ◽  
Xueqi Cheng

Neural language models based on recurrent neural networks (RNNLM) have significantly improved the performance for text generation, yet the quality of generated text represented by Turing Test pass rate is still far from satisfying. Some researchers propose to use adversarial training or reinforcement learning to promote the quality, however, such methods usually introduce great challenges in the training and parameter tuning processes. Through our analysis, we find the problem of RNNLM comes from the usage of maximum likelihood estimation (MLE) as the objective function, which requires the generated distribution to precisely recover the true distribution. Such requirement favors high generation diversity which restricted the generation quality. This is not suitable when the overall quality is low, since high generation diversity usually indicates lot of errors rather than diverse good samples. In this paper, we propose to achieve differentiated distribution recovery, DDR for short. The key idea is to make the optimal generation probability proportional to the β-th power of the true probability, where β > 1. In this way, the generation quality can be greatly improved by sacrificing diversity from noises and rare patterns. Experiments on synthetic data and two public text datasets show that our DDR method achieves more flexible quality-diversity trade-off and higher Turing Test pass rate, as compared with baseline methods including RNNLM, SeqGAN and LeakGAN.


2021 ◽  
Vol 9 ◽  
pp. 586-604
Author(s):  
Abbas Ghaddar ◽  
Philippe Langlais ◽  
Ahmad Rashid ◽  
Mehdi Rezagholizadeh

Abstract In this work, we examine the ability of NER models to use contextual information when predicting the type of an ambiguous entity. We introduce NRB, a new testbed carefully designed to diagnose Name Regularity Bias of NER models. Our results indicate that all state-of-the-art models we tested show such a bias; BERT fine-tuned models significantly outperforming feature-based (LSTM-CRF) ones on NRB, despite having comparable (sometimes lower) performance on standard benchmarks. To mitigate this bias, we propose a novel model-agnostic training method that adds learnable adversarial noise to some entity mentions, thus enforcing models to focus more strongly on the contextual signal, leading to significant gains on NRB. Combining it with two other training strategies, data augmentation and parameter freezing, leads to further gains.


2020 ◽  
Vol 34 (05) ◽  
pp. 9466-9473
Author(s):  
Haiyan Yin ◽  
Dingcheng Li ◽  
Xu Li ◽  
Ping Li

Training generative models that can generate high-quality text with sufficient diversity is an important open problem for Natural Language Generation (NLG) community. Recently, generative adversarial models have been applied extensively on text generation tasks, where the adversarially trained generators alleviate the exposure bias experienced by conventional maximum likelihood approaches and result in promising generation quality. However, due to the notorious defect of mode collapse for adversarial training, the adversarially trained generators face a quality-diversity trade-off, i.e., the generator models tend to sacrifice generation diversity severely for increasing generation quality. In this paper, we propose a novel approach which aims to improve the performance of adversarial text generation via efficiently decelerating mode collapse of the adversarial training. To this end, we introduce a cooperative training paradigm, where a language model is cooperatively trained with the generator and we utilize the language model to efficiently shape the data distribution of the generator against mode collapse. Moreover, instead of engaging the cooperative update for the generator in a principled way, we formulate a meta learning mechanism, where the cooperative update to the generator serves as a high level meta task, with an intuition of ensuring the parameters of the generator after the adversarial update would stay resistant against mode collapse. In the experiment, we demonstrate our proposed approach can efficiently slow down the pace of mode collapse for the adversarial text generators. Overall, our proposed method is able to outperform the baseline approaches with significant margins in terms of both generation quality and diversity in the testified domains.


2020 ◽  
Vol 34 (05) ◽  
pp. 9450-9457
Author(s):  
Xiaoyuan Yi ◽  
Ruoyu Li ◽  
Cheng Yang ◽  
Wenhao Li ◽  
Maosong Sun

As an essential step towards computer creativity, automatic poetry generation has gained increasing attention these years. Though recent neural models make prominent progress in some criteria of poetry quality, generated poems still suffer from the problem of poor diversity. Related literature researches show that different factors, such as life experience, historical background, etc., would influence composition styles of poets, which considerably contributes to the high diversity of human-authored poetry. Inspired by this, we propose MixPoet, a novel model that absorbs multiple factors to create various styles and promote diversity. Based on a semi-supervised variational autoencoder, our model disentangles the latent space into some subspaces, with each conditioned on one influence factor by adversarial training. In this way, the model learns a controllable latent variable to capture and mix generalized factor-related properties. Different factor mixtures lead to diverse styles and hence further differentiate generated poems from each other. Experiment results on Chinese poetry demonstrate that MixPoet improves both diversity and quality against three state-of-the-art models.


2005 ◽  
Vol 173 (4S) ◽  
pp. 172-172
Author(s):  
Masatoshi Eto ◽  
Masahiko Harano ◽  
Katsunori Tatsugami ◽  
Hirofumi Koga ◽  
Seiji Naito

2012 ◽  
Author(s):  
Richard A. Chechile ◽  
Lara N. Sloboda ◽  
Erin L. Warren ◽  
Daniel H. Barch ◽  
Jessica R. Chamberland
Keyword(s):  

CounterText ◽  
2015 ◽  
Vol 1 (3) ◽  
pp. 348-365 ◽  
Author(s):  
Mario Aquilina

What if the post-literary also meant that which operates in a literary space (almost) devoid of language as we know it: for instance, a space in which language simply frames the literary or poetic rather than ‘containing’ it? What if the countertextual also meant the (en)countering of literary text with non-textual elements, such as mathematical concepts, or with texts that we would not normally think of as literary, such as computer code? This article addresses these issues in relation to Nick Montfort's #!, a 2014 print collection of poems that presents readers with the output of computer programs as well as the programs themselves, which are designed to operate on principles of text generation regulated by specific constraints. More specifically, it focuses on two works in the collection, ‘Round’ and ‘All the Names of God’, which are read in relation to the notions of the ‘computational sublime’ and the ‘event’.


Diabetes ◽  
2018 ◽  
Vol 67 (Supplement 1) ◽  
pp. 104-OR
Author(s):  
ADRIANA VIEIRA DE ABREU ◽  
RAHUL AGRAWAL ◽  
PARKER HOWE ◽  
SIMON J. FISHER
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document