scholarly journals PRM94 - ALIGNING TEXT MINING AND MACHINE LEARNING ALGORITHMS WITH BEST PRACTICES FOR STUDY SELECTION IN SYSTEMATIC LITERATURE REVIEWS

2018 ◽  
Vol 21 ◽  
pp. S371
Author(s):  
E Popoff ◽  
JP Jansen ◽  
M Besada ◽  
S Cope ◽  
S Kanters
2020 ◽  
Vol 9 (1) ◽  
Author(s):  
E. Popoff ◽  
M. Besada ◽  
J. P. Jansen ◽  
S. Cope ◽  
S. Kanters

Abstract Background Despite existing research on text mining and machine learning for title and abstract screening, the role of machine learning within systematic literature reviews (SLRs) for health technology assessment (HTA) remains unclear given lack of extensive testing and of guidance from HTA agencies. We sought to address two knowledge gaps: to extend ML algorithms to provide a reason for exclusion—to align with current practices—and to determine optimal parameter settings for feature-set generation and ML algorithms. Methods We used abstract and full-text selection data from five large SLRs (n = 3089 to 12,769 abstracts) across a variety of disease areas. Each SLR was split into training and test sets. We developed a multi-step algorithm to categorize each citation into the following categories: included; excluded for each PICOS criterion; or unclassified. We used a bag-of-words approach for feature-set generation and compared machine learning algorithms using support vector machines (SVMs), naïve Bayes (NB), and bagged classification and regression trees (CART) for classification. We also compared alternative training set strategies: using full data versus downsampling (i.e., reducing excludes to balance includes/excludes because machine learning algorithms perform better with balanced data), and using inclusion/exclusion decisions from abstract versus full-text screening. Performance comparisons were in terms of specificity, sensitivity, accuracy, and matching the reason for exclusion. Results The best-fitting model (optimized sensitivity and specificity) was based on the SVM algorithm using training data based on full-text decisions, downsampling, and excluding words occurring fewer than five times. The sensitivity and specificity of this model ranged from 94 to 100%, and 54 to 89%, respectively, across the five SLRs. On average, 75% of excluded citations were excluded with a reason and 83% of these citations matched the reviewers’ original reason for exclusion. Sensitivity significantly improved when both downsampling and abstract decisions were used. Conclusions ML algorithms can improve the efficiency of the SLR process and the proposed algorithms could reduce the workload of a second reviewer by identifying exclusions with a relevant PICOS reason, thus aligning with HTA guidance. Downsampling can be used to improve study selection, and improvements using full-text exclusions have implications for a learn-as-you-go approach.


2017 ◽  
Vol 28 ◽  
pp. v518
Author(s):  
H-L. Wong ◽  
T. Luechtefeld ◽  
A. Prawira ◽  
Z. Patterson ◽  
J. Workman ◽  
...  

2021 ◽  
Author(s):  
Burak Kolukisa ◽  
Bilge Kagan Dedeturk ◽  
Beyhan Adanur Dedeturk ◽  
Abdulkadir Gulsen ◽  
Gokhan Bakal

Author(s):  
Durmuş Özkan Şahin ◽  
Erdal Kılıç

In this study, the authors give both theoretical and experimental information about text mining, which is one of the natural language processing topics. Three different text mining problems such as news classification, sentiment analysis, and author recognition are discussed for Turkish. They aim to reduce the running time and increase the performance of machine learning algorithms. Four different machine learning algorithms and two different feature selection metrics are used to solve these text classification problems. Classification algorithms are random forest (RF), logistic regression (LR), naive bayes (NB), and sequential minimal optimization (SMO). Chi-square and information gain metrics are used as the feature selection method. The highest classification performance achieved in this study is 0.895 according to the F-measure metric. This result is obtained by using the SMO classifier and information gain metric for news classification. This study is important in terms of comparing the performances of classification algorithms and feature selection methods.


2017 ◽  
Author(s):  
Woo-Young Ahn ◽  
Paul Hendricks ◽  
Nathaniel Haines

AbstractThe easyml (easy machine learning) package lowers the barrier to entry to machine learning and is ideal for undergraduate/graduate students, and practitioners who want to quickly apply machine learning algorithms to their research without having to worry about the best practices of implementing each algorithm. The package provides standardized recipes for regression and classification algorithms in R and Python and implements them in a functional, modular, and extensible framework. This package currently implements recipes for several common machine learning algorithms (e.g., penalized linear models, random forests, and support vector machines) and provides a unified interface to each one. Importantly, users can run and evaluate each machine learning algorithm with a single line of coding. Each recipe is robust, implements best practices specific to each algorithm, and generates a report with details about the model, its performance, as well as journal-quality visualizations. The package’s functional, modular, and extensible framework also allows researchers and more advanced users to easily implement new recipes for other algorithms.


2020 ◽  
Author(s):  
Ross Jacobucci ◽  
Andrew K. Littlefield ◽  
Alex J. Millner ◽  
Evan Kleiman ◽  
Douglas Steinley

Machine learning is being utilized at an increasing rate in clinical psychology. Applying machine learning comes with a number of challenges, both in deciding which algorithms to test, and how to evaluate the predictive performance. We focus on this last component, demonstrating across both a simulation and empirical example that the method researchers choose to evaluate prediction with machine learning can have large consequences for the substantive conclusions. More specifically, we demonstrate that one method of evaluation that has been used repeatedly in clinical research, the optimism corrected bootstrap, can result in extremely biased results when paired with specific machine learning algorithms. We conclude with providing recommendations for researchers and a discussion of additional best practices.


Sign in / Sign up

Export Citation Format

Share Document