Seq2Emoji: A hybrid sequence generation model for short text emoji prediction

2021 ◽  
Vol 214 ◽  
pp. 106727
Author(s):  
Dunlu Peng ◽  
Huimin Zhao
2019 ◽  
Vol 7 ◽  
pp. 661-676 ◽  
Author(s):  
Jiatao Gu ◽  
Qi Liu ◽  
Kyunghyun Cho

Conventional neural autoregressive decoding commonly assumes a fixed left-to-right generation order, which may be sub-optimal. In this work, we propose a novel decoding algorithm— InDIGO—which supports flexible sequence generation in arbitrary orders through insertion operations. We extend Transformer, a state-of-the-art sequence generation model, to efficiently implement the proposed approach, enabling it to be trained with either a pre-defined generation order or adaptive orders obtained from beam-search. Experiments on four real-world tasks, including word order recovery, machine translation, image caption, and code generation, demonstrate that our algorithm can generate sequences following arbitrary orders, while achieving competitive or even better performance compared with the conventional left-to-right generation. The generated sequences show that InDIGO adopts adaptive generation orders based on input information.


Author(s):  
Jun Gao ◽  
Wei Bi ◽  
Xiaojiang Liu ◽  
Junhui Li ◽  
Shuming Shi

Neural generative models have become popular and achieved promising performance on short-text conversation tasks. They are generally trained to build a 1-to-1 mapping from the input post to its output response. However, a given post is often associated with multiple replies simultaneously in real applications. Previous research on this task mainly focuses on improving the relevance and informativeness of the top one generated response for each post. Very few works study generating multiple accurate and diverse responses for the same post. In this paper, we propose a novel response generation model, which considers a set of responses jointly and generates multiple diverse responses simultaneously. A reinforcement learning algorithm is designed to solve our model. Experiments on two short-text conversation tasks validate that the multiple responses generated by our model obtain higher quality and larger diversity compared with various state-ofthe-art generative models.


2018 ◽  
Vol 26 (8) ◽  
pp. 1359-1368 ◽  
Author(s):  
Kai Yu ◽  
Zijian Zhao ◽  
Xueyang Wu ◽  
Hongtao Lin ◽  
Xuan Liu

Author(s):  
Ming Yuan ◽  
Jiangtao Ren

The goal of computer-aided diagnosis is to predict patient’s diseases based on patient’s clinical data. The development of deep learning technology provides new help for clinical diagnosis. In this paper, we propose a new sequence generation model for multi-disease diagnosis prediction based on numerical feature transformation. Our model simultaneously uses patient’s laboratory test results and clinical text as input to diagnose and predict the disease that the patient may have. According to medical knowledge, our model can transform numerical features into descriptive text features, thereby enriching the semantic information of clinical texts. Besides, our model uses attention-based sequence generation methods to achieve the diagnosis of multiple diseases and better utilizes the correlation information between multiple diseases. We evaluate our model’s performance on a dataset of respiratory diseases from the real world, and experimental results show that our model’s accuracy reaches 42.75%, and the [Formula: see text] score reaches 65.65%, which is better than many other methods. It is suitable for the accurate diagnosis of multiple diseases.


2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Lu Zhou ◽  
Shuangqiao Liu ◽  
Caiyan Li ◽  
Yuemeng Sun ◽  
Yizhuo Zhang ◽  
...  

Background. The modernization of traditional Chinese medicine (TCM) demands systematic data mining using medical records. However, this process is hindered by the fact that many TCM symptoms have the same meaning but different literal expressions (i.e., TCM synonymous symptoms). This problem can be solved by using natural language processing algorithms to construct a high-quality TCM symptom normalization model for normalizing TCM synonymous symptoms to unified literal expressions. Methods. Four types of TCM symptom normalization models, based on natural language processing, were constructed to find a high-quality one: (1) a text sequence generation model based on a bidirectional long short-term memory (Bi-LSTM) neural network with an encoder-decoder structure; (2) a text classification model based on a Bi-LSTM neural network and sigmoid function; (3) a text sequence generation model based on bidirectional encoder representation from transformers (BERT) with sequence-to-sequence training method of unified language model (BERT-UniLM); (4) a text classification model based on BERT and sigmoid function (BERT-Classification). The performance of the models was compared using four metrics: accuracy, recall, precision, and F1-score. Results. The BERT-Classification model outperformed the models based on Bi-LSTM and BERT-UniLM with respect to the four metrics. Conclusions. The BERT-Classification model has superior performance in normalizing expressions of TCM synonymous symptoms.


2020 ◽  
Vol 382 ◽  
pp. 188-195 ◽  
Author(s):  
Weizhi Liao ◽  
Yu Wang ◽  
Yanchao Yin ◽  
Xiaobing Zhang ◽  
Pan Ma

Sign in / Sign up

Export Citation Format

Share Document