scholarly journals Secure Decision Forest Evaluation

Author(s):  
Slim Bettaieb ◽  
Loic Bidoux ◽  
Olivier Blazy ◽  
Baptiste Cottier ◽  
David Pointcheval
Keyword(s):  
2020 ◽  
Vol 96 ◽  
pp. 101930
Author(s):  
Zhitao Guan ◽  
Xianwen Sun ◽  
Lingyun Shi ◽  
Longfei Wu ◽  
Xiaojiang Du

Author(s):  
Md Nasim Adnan ◽  
Ryan H.L. Ip ◽  
Michael Bewong ◽  
Md Zahidul Islam
Keyword(s):  

2021 ◽  
Vol 2021 ◽  
pp. 1-8
Author(s):  
Khalid Twarish Alhamazani ◽  
Jalawi Alshudukhi ◽  
Saud Aljaloud ◽  
Solomon Abebaw

Chronic kidney disease (CKD) is a global health issue with a high rate of morbidity and mortality and a high rate of disease progression. Because there are no visible symptoms in the early stages of CKD, patients frequently go unnoticed. The early detection of CKD allows patients to receive timely treatment, slowing the disease’s progression. Due to its rapid recognition performance and accuracy, machine learning models can effectively assist physicians in achieving this goal. We propose a machine learning methodology for the CKD diagnosis in this paper. This information was completely anonymized. As a reference, the CRISP-DM® model (Cross industry standard process for data mining) was used. The data were processed in its entirety in the cloud on the Azure platform, where the sample data was unbalanced. Then the processes for exploration and analysis were carried out. According to what we have learned, the data were balanced using the SMOTE technique. Four matching algorithms were used after the data balancing was completed successfully. Artificial intelligence (AI) (logistic regression, decision forest, neural network, and jungle of decisions). The decision forest outperformed the other machine learning models with a score of 92%, indicating that the approach used in this study provides a good baseline for solutions in the production.


2019 ◽  
Vol 6 (1) ◽  
Author(s):  
Gautam Pal ◽  
Xianbin Hong ◽  
Zhuo Wang ◽  
Hongyi Wu ◽  
Gangmin Li ◽  
...  

Abstract Introduction This paper presents a lifelong learning framework which constantly adapts with changing data patterns over time through incremental learning approach. In many big data systems, iterative re-training high dimensional data from scratch is computationally infeasible since constant data stream ingestion on top of a historical data pool increases the training time exponentially. Therefore, the need arises on how to retain past learning and fast update the model incrementally based on the new data. Also, the current machine learning approaches do the model prediction without providing a comprehensive root cause analysis. To resolve these limitations, our framework lays foundations on an ensemble process between stream data with historical batch data for an incremental lifelong learning (LML) model. Case description A cancer patient’s pathological tests like blood, DNA, urine or tissue analysis provide a unique signature based on the DNA combinations. Our analysis allows personalized and targeted medications and achieves a therapeutic response. Model is evaluated through data from The National Cancer Institute’s Genomic Data Commons unified data repository. The aim is to prescribe personalized medicine based on the thousands of genotype and phenotype parameters for each patient. Discussion and evaluation The model uses a dimension reduction method to reduce training time at an online sliding window setting. We identify the Gleason score as a determining factor for cancer possibility and substantiate our claim through Lilliefors and Kolmogorov–Smirnov test. We present clustering and Random Decision Forest results. The model’s prediction accuracy is compared with standard machine learning algorithms for numeric and categorical fields. Conclusion We propose an ensemble framework of stream and batch data for incremental lifelong learning. The framework successively applies first streaming clustering technique and then Random Decision Forest Regressor/Classifier to isolate anomalous patient data and provides reasoning through root cause analysis by feature correlations with an aim to improve the overall survival rate. While the stream clustering technique creates groups of patient profiles, RDF further drills down into each group for comparison and reasoning for useful actionable insights. The proposed MALA architecture retains the past learned knowledge and transfer to future learning and iteratively becomes more knowledgeable over time.


2019 ◽  
Vol 15 (5) ◽  
pp. 673-680
Author(s):  
Budi Padmaja ◽  
Venkata Rama Prasad Vaddella ◽  
Kota Venkata Naga Sunitha

Sign in / Sign up

Export Citation Format

Share Document