topic labeling
Recently Published Documents


TOTAL DOCUMENTS

25
(FIVE YEARS 14)

H-INDEX

3
(FIVE YEARS 0)

2022 ◽  
Vol 7 ◽  
pp. e831
Author(s):  
Xudong Jia ◽  
Li Wang

Text classification is a fundamental task in many applications such as topic labeling, sentiment analysis, and spam detection. The text syntactic relationship and word sequence are important and useful for text classification. How to model and incorporate them to improve performance is one key challenge. Inspired by human behavior in understanding text. In this paper, we combine the syntactic relationship, sequence structure, and semantics for text representation, and propose an attention-enhanced capsule network-based text classification model. Specifically, we use graph convolutional neural networks to encode syntactic dependency trees, build multi-head attention to encode dependencies relationship in text sequence, merge with semantic information by capsule network at last. Extensive experiments on five datasets demonstrate that our approach can effectively improve the performance of text classification compared with state-of-the-art methods. The result also shows capsule network, graph convolutional neural network, and multi-headed attention has integration effects on text classification tasks.


2021 ◽  
Vol 463 ◽  
pp. 596-608
Author(s):  
Dongbin He ◽  
Yanzhao Ren ◽  
Abdul Mateen Khattak ◽  
Xinliang Liu ◽  
Sha Tao ◽  
...  
Keyword(s):  

2021 ◽  
Author(s):  
Dongbin He ◽  
Yanzhao Ren ◽  
Abdul Mateen Khattak ◽  
Xinliang Liu ◽  
Sha Tao ◽  
...  

Symmetry ◽  
2021 ◽  
Vol 13 (5) ◽  
pp. 837
Author(s):  
Olzhas Kozbagarov ◽  
Rustam Mussabayev ◽  
Nenad Mladenovic

This article presents a new conceptual approach for the interpretative topic modeling problem. It uses sentences as basic units of analysis, instead of words or n-grams, which are commonly used in the standard approaches.The proposed approach’s specifics are using sentence probability evaluations within the text corpus and clustering of sentence embeddings. The topic model estimates discrete distributions of sentence occurrences within topics and discrete distributions of topic occurrence within the text. Our approach provides the possibility of explicit interpretation of topics since sentences, unlike words, are more informative and have complete grammatical and semantic constructions inside. The method for automatic topic labeling is also provided. Contextual embeddings based on the BERT model are used to obtain corresponding sentence embeddings for their subsequent analysis. Moreover, our approach allows big data processing and shows the possibility of utilizing the combination of internal and external knowledge sources in the process of topic modeling. The internal knowledge source is represented by the text corpus itself and often it is a single knowledge source in the traditional topic modeling approaches. The external knowledge source is represented by the BERT, a machine learning model which was preliminarily trained on a huge amount of textual data and is used for generating the context-dependent sentence embeddings.


2021 ◽  
Vol 8 (2) ◽  
pp. 205316802110222
Author(s):  
Hannah Béchara ◽  
Alexander Herzog ◽  
Slava Jankin ◽  
Peter John

Topic models are widely used in natural language processing, allowing researchers to estimate the underlying themes in a collection of documents. Most topic models require the additional step of attaching meaningful labels to estimated topics, a process that is not scalable, suffers from human bias, and is difficult to replicate. We present a transfer topic labeling method that seeks to remedy these problems, using domain-specific codebooks as the knowledge base to automatically label estimated topics. We demonstrate our approach with a large-scale topic model analysis of the complete corpus of UK House of Commons speeches from 1935 to 2014, using the coding instructions of the Comparative Agendas Project to label topics. We evaluated our results using human expert coding and compared our approach with more current state-of-the-art neural methods. Our approach was simple to implement, compared favorably to expert judgments, and outperformed the neural networks model for a majority of the topics we estimated.


Author(s):  
Yi Yang ◽  
Hongan Wang ◽  
Jiaqi Zhu ◽  
Wandong Shi ◽  
Wenli Guo ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document