scholarly journals Building Chatbots from Forum Data: Model Selection Using Question Answering Metrics

Author(s):  
Martin Boyanov ◽  
◽  
Ivan Koychev ◽  
Preslav Nakov ◽  
Alessandro Moschitti ◽  
...  
Author(s):  
Bezza Hafidi ◽  
Nourddine Azzaoui

International audience Recently, Azari et al (2006) showed that (AIC) criterion and its corrected versions cannot be directly applied to model selection for longitudinal data with correlated errors. They proposed two model selection criteria, AICc and RICc, by applying likelihood and residual likelihood approaches. These two criteria are estimators of the Kullback-Leibler's divergence distance which is asymmetric. In this work, we apply the likelihood and residual likelihood approaches to propose two new criteria, suitable for small samples longitudinal data, based on the Kullback's symmetric divergence. Their performance relative to others criteria is examined in a large simulation study


2006 ◽  
Vol 50 (11) ◽  
pp. 3053-3066 ◽  
Author(s):  
Rahman Azari ◽  
Lexin Li ◽  
Chih-Ling Tsai

Information ◽  
2020 ◽  
Vol 11 (5) ◽  
pp. 241
Author(s):  
Tedo Vrbanec ◽  
Ana Meštrović

Paraphrase detection is important for a number of applications, including plagiarism detection, authorship attribution, question answering, text summarization, text mining in general, etc. In this paper, we give a performance overview of various types of corpus-based models, especially deep learning (DL) models, with the task of paraphrase detection. We report the results of eight models (LSI, TF-IDF, Word2Vec, Doc2Vec, GloVe, FastText, ELMO, and USE) evaluated on three different public available corpora: Microsoft Research Paraphrase Corpus, Clough and Stevenson and Webis Crowd Paraphrase Corpus 2011. Through a great number of experiments, we decided on the most appropriate approaches for text pre-processing: hyper-parameters, sub-model selection—where they exist (e.g., Skipgram vs. CBOW), distance measures, and semantic similarity/paraphrase detection threshold. Our findings and those of other researchers who have used deep learning models show that DL models are very competitive with traditional state-of-the-art approaches and have potential that should be further developed.


Sign in / Sign up

Export Citation Format

Share Document