Improvements in multi-document abstractive summarization using multi sentence compression with Word Graph and node alignment

2021 ◽  
pp. 116154
Author(s):  
Raksha Agarwal ◽  
Niladri Chatterjee
2021 ◽  
pp. 106996
Author(s):  
Xiaoyan Cai ◽  
Kaile Shi ◽  
Yuehan Jiang ◽  
Libin Yang ◽  
Sen Liu

2021 ◽  
Author(s):  
Weizhi Liao ◽  
Yaheng Ma ◽  
Yanchao Yin ◽  
Guanglei Ye ◽  
Dongzhou Zuo

IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 29253-29263 ◽  
Author(s):  
Aniqa Dilawari ◽  
Muhammad Usman Ghani Khan

2021 ◽  
Author(s):  
Tham Vo

Abstract In abstractive summarization task, most of proposed models adopt the deep recurrent neural network (RNN)-based encoder-decoder architecture to learn and generate meaningful summary for a given input document. However, most of recent RNN-based models always suffer the challenges related to the involvement of much capturing high-frequency/reparative phrases in long documents during the training process which leads to the outcome of trivial and generic summaries are generated. Moreover, the lack of thorough analysis on the sequential and long-range dependency relationships between words within different contexts while learning the textual representation also make the generated summaries unnatural and incoherent. To deal with these challenges, in this paper we proposed a novel semantic-enhanced generative adversarial network (GAN)-based approach for abstractive text summarization task, called as: SGAN4AbSum. We use an adversarial training strategy for our text summarization model in which train the generator and discriminator to simultaneously handle the summary generation and distinguishing the generated summary with the ground-truth one. The input of generator is the jointed rich-semantic and global structural latent representations of training documents which are achieved by applying a combined BERT and graph convolutional network (GCN) textual embedding mechanism. Extensive experiments in benchmark datasets demonstrate the effectiveness of our proposed SGAN4AbSum which achieve the competitive ROUGE-based scores in comparing with state-of-the-art abstractive text summarization baselines.


2016 ◽  
Vol 78 (8) ◽  
Author(s):  
Suraya Alias ◽  
Siti Khaotijah Mohammad ◽  
Gan Keng Hoon ◽  
Tan Tien Ping

A text summary extracts serves as a condensed representation of a written input source where important and salient information is kept. However, the condensed representation itself suffer in lack of semantic and coherence if the summary was produced in verbatim using the input itself. Sentence Compression is a technique where unimportant details from a sentence are eliminated by preserving the sentence’s grammar pattern. In this study, we conducted an analysis on our developed Malay Text Corpus to discover the rules and pattern on how human summarizer compresses and eliminates unimportant constituent to construct a summary. A Pattern-Growth based model named Frequent Eliminated Pattern (FASPe) is introduced to represent the text using a set of sequence adjacent words that is frequently being eliminated across the document collection. From the rules obtained, some heuristic knowledge in Sentence Compression is presented with confidence value as high as 85% - that can be used for further reference in the area of Text Summarization for Malay language.


Sign in / Sign up

Export Citation Format

Share Document