Constructing VEGGIE: Machine Learning for Context-Sensitive Graph Grammars

Author(s):  
Keven Ates ◽  
Kang Zhang
2012 ◽  
Vol 23 (7) ◽  
pp. 1635-1655 ◽  
Author(s):  
Yang ZOU ◽  
Jian LÜ ◽  
Chun CAO ◽  
Hao HU ◽  
Wei SONG ◽  
...  

PLoS ONE ◽  
2015 ◽  
Vol 10 (11) ◽  
pp. e0142776
Author(s):  
Yi Wang ◽  
XiaoQin Zeng ◽  
Han Ding

2021 ◽  
Vol 26 (1) ◽  
pp. 47-57
Author(s):  
Paul Menounga Mbilong ◽  
Asmae Berhich ◽  
Imane Jebli ◽  
Asmae El Kassiri ◽  
Fatima-Zahra Belouadha

Coronavirus 2019 (COVID-19) has reached the stage of an international epidemic with a major socioeconomic negative impact. Considering the weakness of the healthy structure and the limited availability of test kits, particularly in emerging countries, predicting the spread of COVID-19 is expected to help decision-makers to improve health management and contribute to alleviating the related risks. In this article, we studied the effectiveness of machine learning techniques using Morocco as a case-study. We studied the performance of six multi-step models derived from both Machine Learning and Deep Learning regards multiple scenarios by combining different time lags and three COVID-19 datasets(periods): confinement, deconfinement, and hybrid datasets. The results prove the efficiency of Deep Learning models and identify the best combinations of these models and the time lags enabling good predictions of new cases. The results also show that the prediction of the spread of COVID-19 is a context sensitive problem.


2009 ◽  
pp. 135-147 ◽  
Author(s):  
Yang Zou ◽  
Jian Lü ◽  
Xiaoqin Zeng ◽  
Xiaoxing Ma ◽  
Qiliang Yang

2021 ◽  
Author(s):  
Zahra Shakeri Hossein Abad ◽  
Joon Lee ◽  
Gregory P. Butler ◽  
Wendy Thompson

BACKGROUND Crowdsourcing services such as Amazon Mechanical Turk (AMT) allow researchers to use the collective intelligence of a wide range of online users for labour-intensive tasks. Since the manual verification of the quality of the collected results is difficult due to the large volume of data and the quick turnaround time of the process, many questions remain to be explored regarding the reliability of these resources for developing digital public health systems. OBJECTIVE The main objective of this study is to explore and evaluate the application of crowdsourcing, in general, and AMT, in specific, for developing digital public health surveillance systems. METHODS We collected 296,166 crowd-generated labels for 98,722 tweets, labelled by 610 AMT workers, to develop machine learning (ML) models for detecting behaviours related to physical activity, sedentary behaviour, and sleep quality (PASS) among Twitter users. To infer the ground truth labels and explore the quality of these labels, we studied four statistical consensus methods that are agnostic of task features and only focus on worker labelling behaviour. Moreover, to model the meta-information associated with each labelling task and leverage the potentials of context-sensitive data in the truth inference process, we developed seven ML models, including traditional classifiers (offline and active), a deep-learning-based classification model, and a hybrid convolutional neural network (CNN) model. RESULTS While most of the crowdsourcing-based studies in public health have often equated majority vote with quality, the results of our study using a truth set of 9,000 manually labelled tweets show that consensus-based inference models mask underlying uncertainty in the data and overlook the importance of task meta-information. Our evaluations across three PASS datasets show that truth inference is a context-sensitive process, and none of the studied methods in this paper was consistently superior to others in predicting the truth label. We also found that the performance of the ML models trained on crowd-labelled data is sensitive to the quality of these labels, and poor-quality labels lead to incorrect assessment of these models. Finally, we provide a set of practical recommendations to improve the quality and reliability of crowdsourced data. CONCLUSIONS Findings indicate the importance of the quality of crowd-generated labels in developing machine learning models designed for decision-making purposes, such as public health surveillance decisions. A combination of inference models outlined and analyzed in this work could be used to quantitatively measure and improve the quality of crowd-generated labels for training ML models. CLINICALTRIAL Not Applicable


Sign in / Sign up

Export Citation Format

Share Document