scholarly journals An interpretable model for incident heart failure prediction with uncertainty estimation

2020 ◽  
Vol 41 (Supplement_2) ◽  
Author(s):  
Y Li ◽  
S Rao ◽  
A Hassaine ◽  
R Ramakrishnan ◽  
Y Zhu ◽  
...  

Abstract Background Forecasting incident heart failure is a critical demand for prevention. Recent research suggested the superior performance of deep learning models on the prediction tasks using electronic health records. However, even with a relatively accurate predictive performance, the major impediments to the wider use of deep learning models for clinical decision making are the difficulties of assigning a level of confidence to model predictions and the interpretability of predictions. Purpose We aimed to develop a deep learning framework for more accurate incident heart failure prediction, with provision of measures of uncertainty and interpretability. Methods We used a longitudinal linked electronic health records dataset, Clinical Practice Research Datalink, involving 788,880 patients, 8.3% of whom had an incident heart failure diagnosis. To embed the uncertainty estimation mechanism into the deep learning models, we developed a probabilistic framework based on a novel transformer deep learning model: deep Bayesian Gaussian processes (DBGP). We investigated the performance of incident heart failure prediction and uncertainty estimation for the model and validated it using an external held-out dataset. Diagnoses, medications, and age for each encounter were included as predictors. By comparing the uncertainty, we investigated the possibility of identifying the correct predictions from wrong ones to avoid potential misclassification. Using model distillation meant to mimic a well-trained complex model with simple models, we investigated the importance of associations between diagnoses, medications and heart failure with an interpretable linear regression component learned from DBGP. Results The DBGP achieved high precision with 0.941 as AUROC for external validation. More importantly, it showed the uncertainty information could distinguish the correct predictions from wrong ones, with significant difference (p-value with 500 samples) between distribution of uncertainties for negative predictions (3.21e-69 between true negative and false negative), and positive predictions (3.39e-22 between true positive and false positive). Utilising the distilled model, we can specify the contribution of each diagnosis and medication to heart failure prediction. For instance, Losartan/Fosinopril, Bisoprolol and Left bundle-branch block showed strong association to heart failure incidence with coefficient 0.11 (95% CI: 0.10, 0.12), 0.09 (0.08, 0.11) and 0.09 (0.07, 0.11) respectively; Peritoneal adhesions, Trochanteric bursitis and Galactorrhea showed strong disassociations with coefficient −0.07 (−0.09, −0.05), −0.07 (−0.09, −0.04) and −0.06 (−0.08, −0.04) individually. Conclusions Our novel probabilistic deep learning framework adds a measure of uncertainty the prediction and helps to mitigate misclassification. Model distillation provides an opportunity to interpret deep learning models and offers a data-driven perspective for risk factor analysis. Funding Acknowledgement Type of funding source: Public Institution(s). Main funding source(s): Oxford Martin School,University of Oxford; NIHR Oxford Biomedical Research Centre, University of Oxford

2020 ◽  
Vol 41 (Supplement_2) ◽  
Author(s):  
S Rao ◽  
Y Li ◽  
R Ramakrishnan ◽  
A Hassaine ◽  
D Canoy ◽  
...  

Abstract Background/Introduction Predicting incident heart failure has been challenging. Deep learning models when applied to rich electronic health records (EHR) offer some theoretical advantages. However, empirical evidence for their superior performance is limited and they remain commonly uninterpretable, hampering their wider use in medical practice. Purpose We developed a deep learning framework for more accurate and yet interpretable prediction of incident heart failure. Methods We used longitudinally linked EHR from practices across England, involving 100,071 patients, 13% of whom had been diagnosed with incident heart failure during follow-up. We investigated the predictive performance of a novel transformer deep learning model, “Transformer for Heart Failure” (BEHRT-HF), and validated it using both an external held-out dataset and an internal five-fold cross-validation mechanism using area under receiver operating characteristic (AUROC) and area under the precision recall curve (AUPRC). Predictor groups included all outpatient and inpatient diagnoses within their temporal context, medications, age, and calendar year for each encounter. By treating diagnoses as anchors, we alternatively removed different modalities (ablation study) to understand the importance of individual modalities to the performance of incident heart failure prediction. Using perturbation-based techniques, we investigated the importance of associations between selected predictors and heart failure to improve model interpretability. Results BEHRT-HF achieved high accuracy with AUROC 0.932 and AUPRC 0.695 for external validation, and AUROC 0.933 (95% CI: 0.928, 0.938) and AUPRC 0.700 (95% CI: 0.682, 0.718) for internal validation. Compared to the state-of-the-art recurrent deep learning model, RETAIN-EX, BEHRT-HF outperformed it by 0.079 and 0.030 in terms of AUPRC and AUROC. Ablation study showed that medications were strong predictors, and calendar year was more important than age. Utilising perturbation, we identified and ranked the intensity of associations between diagnoses and heart failure. For instance, the method showed that established risk factors including myocardial infarction, atrial fibrillation and flutter, and hypertension all strongly associated with the heart failure prediction. Additionally, when population was stratified into different age groups, incident occurrence of a given disease had generally a higher contribution to heart failure prediction in younger ages than when diagnosed later in life. Conclusions Our state-of-the-art deep learning framework outperforms the predictive performance of existing models whilst enabling a data-driven way of exploring the relative contribution of a range of risk factors in the context of other temporal information. Funding Acknowledgement Type of funding source: Private grant(s) and/or Sponsorship. Main funding source(s): National Institute for Health Research, Oxford Martin School, Oxford Biomedical Research Centre


2021 ◽  
Vol 42 (Supplement_1) ◽  
Author(s):  
M Lewis ◽  
J Figueroa

Abstract   Recent health reforms have created incentives for cardiologists and accountable care organizations to participate in value-based care models for heart failure (HF). Accurate risk stratification of HF patients is critical to efficiently deploy interventions aimed at reducing preventable utilization. The goal of this paper was to compare deep learning approaches with traditional logistic regression (LR) to predict preventable utilization among HF patients. We conducted a prognostic study using data on 93,260 HF patients continuously enrolled for 2-years in a large U.S. commercial insurer to develop and validate prediction models for three outcomes of interest: preventable hospitalizations, preventable emergency department (ED) visits, and preventable costs. Patients were split into training, validation, and testing samples. Outcomes were modeled using traditional and enhanced LR and compared to gradient boosting model and deep learning models using sequential and non-sequential inputs. Evaluation metrics included precision (positive predictive value) at k, cost capture, and Area Under the Receiver operating characteristic (AUROC). Deep learning models consistently outperformed LR for all three outcomes with respect to the chosen evaluation metrics. Precision at 1% for preventable hospitalizations was 43% for deep learning compared to 30% for enhanced LR. Precision at 1% for preventable ED visits was 39% for deep learning compared to 33% for enhanced LR. For preventable cost, cost capture at 1% was 30% for sequential deep learning, compared to 18% for enhanced LR. The highest AUROCs for deep learning were 0.778, 0.681 and 0.727, respectively. These results offer a promising approach to identify patients for targeted interventions. FUNDunding Acknowledgement Type of funding sources: Private company. Main funding source(s): internally funded by Diagnostic Robotics Inc.


Author(s):  
Malusi Sibiya ◽  
Mbuyu Sumbwanyambe

Machine learning systems use different algorithms to detect the diseases affecting the plant leaves. Nevertheless, selecting a suitable machine learning framework differs from study to study, depending on the features and complexity of the software packages. This paper introduces a taxonomic inspection of the literature in deep learning frameworks for the detection of plant leaf diseases. The objective of this study is to identify the dominating software frameworks in the literature for modelling machine learning plant leaf disease detecting systems.


PLoS ONE ◽  
2021 ◽  
Vol 16 (1) ◽  
pp. e0245177
Author(s):  
Xing Han Lu ◽  
Aihua Liu ◽  
Shih-Chieh Fuh ◽  
Yi Lian ◽  
Liming Guo ◽  
...  

Motivation Recurrent neural networks (RNN) are powerful frameworks to model medical time series records. Recent studies showed improved accuracy of predicting future medical events (e.g., readmission, mortality) by leveraging large amount of high-dimensional data. However, very few studies have explored the ability of RNN in predicting long-term trajectories of recurrent events, which is more informative than predicting one single event in directing medical intervention. Methods In this study, we focus on heart failure (HF) which is the leading cause of death among cardiovascular diseases. We present a novel RNN framework named Deep Heart-failure Trajectory Model (DHTM) for modelling the long-term trajectories of recurrent HF. DHTM auto-regressively predicts the future HF onsets of each patient and uses the predicted HF as input to predict the HF event at the next time point. Furthermore, we propose an augmented DHTM named DHTM+C (where “C” stands for co-morbidities), which jointly predicts both the HF and a set of acute co-morbidities diagnoses. To efficiently train the DHTM+C model, we devised a novel RNN architecture to model disease progression implicated in the co-morbidities. Results Our deep learning models confers higher prediction accuracy for both the next-step HF prediction and the HF trajectory prediction compared to the baseline non-neural network models and the baseline RNN model. Compared to DHTM, DHTM+C is able to output higher probability of HF for high-risk patients, even in cases where it is only given less than 2 years of data to predict over 5 years of trajectory. We illustrated multiple non-trivial real patient examples of complex HF trajectories, indicating a promising path for creating highly accurate and scalable longitudinal deep learning models for modeling the chronic disease.


2019 ◽  
Vol 21 (10) ◽  
pp. 1197-1206 ◽  
Author(s):  
Alicia Uijl ◽  
Stefan Koudstaal ◽  
Kenan Direk ◽  
Spiros Denaxas ◽  
Rolf H. H. Groenwold ◽  
...  

Author(s):  
Bahzad Taha Chicho ◽  
◽  
Amira Bibo Sallow ◽  

Python is one of the most widely adopted programming languages, having replaced a number of those in the field. Python is popular with developers for a variety of reasons, one of which is because it has an incredibly diverse collection of libraries that users can run. The most compelling reasons for adopting Keras come from its guiding principles, particularly those related to usability. Aside from the simplicity of learning and model construction, Keras has a wide variety of production deployment options and robust support for multiple GPUs and distributed training. A strong and easy-to-use free, open-source Python library is the most important tool for developing and evaluating deep learning models. The aim of this paper is to provide the most current survey of Keras in different aspects, which is a Python-based deep learning Application Programming Interface (API) that runs on top of the machine learning framework, TensorFlow. The mentioned library is used in conjunction with TensorFlow, PyTorch, CODEEPNEATM, and Pygame to allow integration of deep learning models such as cardiovascular disease diagnostics, graph neural networks, identifying health issues, COVID-19 recognition, skin tumors, image detection, and so on, in the applied area. Furthermore, the author used Keras's details, goals, challenges, significant outcomes, and the findings obtained using this method.


2008 ◽  
Vol 1 (2) ◽  
pp. 125-133 ◽  
Author(s):  
Javed Butler ◽  
Andreas Kalogeropoulos ◽  
Vasiliki Georgiopoulou ◽  
Rhonda Belue ◽  
Nicolas Rodondi ◽  
...  

2020 ◽  
Vol 41 (Supplement_2) ◽  
Author(s):  
V Liang ◽  
H Holtstrand-Hjalm ◽  
Y Peker ◽  
E Thunstrom

Abstract Background Obstructive sleep apnoea (OSA) is highly prevalent among patients with heart failure. Accumulating research data suggest that this association is bidirectional. Less is known regarding the long-term impact of OSA and continuous positive airway pressure (CPAP) treatment on incident heart failure. Purpose We addressed the association of severe OSA with development of heart failure, and consequently addressed the impact of efficient CPAP treatment in a sleep clinic cohort. Methods The “Sleep Apnea Patients in Skaraborg (SAPIS)” project was a single center (two sites), open-label, prospective cohort study, conducted in Sweden between 2005 and 2018. All consecutive adults admitted to the Skaraborg Hospital between 2005 and 2011 were registered in a local database, and the follow-up ended in May 2018. Anthropomorphic and clinical characteristics as well as results of the diagnostic cardiorespiratory recordings were documented. Treatment of OSA was based on the clinical routines. OSA was defined as an apnoea-hypopnoea index (AHI) of at least 5 events/hr, and severe OSA consisted of patients with an AHI ≥30 events/hr. Median follow-up for the entire cohort was 8.8 years (interquartile range 7.5–10.1 years). Data regarding incident heart failure were obtained from the medical records and the Swedish Hospital Discharge Register. CPAP use (downloaded reports from the devices) of at least 4 hrs/night was defined as efficient treatment. Results Among 4239 patients with diagnostic sleep recordings, 3185 were free of a known cardiac disease at baseline. Severe OSA was observed among 953 (29.9%). Severe OSA significantly predicted incident heart failure (hazard ratio [HR] 2.42; 95% confidence interval [CI] 1.44–4.06) compared to adults with AHI <30 events/hr, adjusted for age, gender, obesity, hypertension and diabetes mellitus. The adjusted HR for severe OSA was 2.82 (95% CI 1.33–5.99) among inefficiently treated/untreated patients whereas the risk was lower but still meaningful among the individuals who were adherent to CPAP (HR 2.25; 95% CI 0.99–5.15) Conclusion Our results suggest that severe OSA is associated with increased risk for development of heart failure. More than 4 hours of CPAP use per night may be necessary for OSA patients in the primary prevention models. CHF-free survival Funding Acknowledgement Type of funding source: Public grant(s) – National budget only. Main funding source(s): ALF


Circulation ◽  
2020 ◽  
Vol 142 (Suppl_3) ◽  
Author(s):  
Chayakrit Krittanawong ◽  
Kipp W Johnson ◽  
Usman Baber ◽  
Mehmet Aydar ◽  
Zhen Wang ◽  
...  

Introduction: Heart failure (HF) is a leading cause of hospitalization, morbidity and mortality. Deep learning (DL) techniques appear to show promising results in risk stratification and prognosis in several conditions in medicine. However, few methods using DL exist to help quantitatively estimate prognosis of HF. We hypothesized that deep learning (DL) techniques could prognosis of HF using simple variables. We propose application of a custom-built deep-neural-network model to identify mortality in HF patients. Methods: Custom-built deep-neural-networks were assessed using survey data from 42,147 participants from the National Health and Nutrition Examination Survey 1999-2016 (NHANES). Variables were selected using clinical judgment and stepwise backward regressions to develop prediction models. We partitioned the data into training and testing sets and repetitive experiments. We then evaluated model performance based on discrimination and calibration including the area under the receiver-operator characteristics curve (C-statistics), balanced accuracy, probability calibration with sigmoid, and the Brier score, respectively. As sensitivity analyses, we examined results limited to cases with complete clinical information available. We validated models’ performance using Mount Sinai database. Results: Of 42,147 participants with 4,060 variables, 1,491 (3.5%) had HF and HF mortality was 51.8%. In validation cohort, of 26,333 HF patients, the mortality in HF patients was 405 (1.5%). Final model using only 20 variables (age, race, gender, BMI, smoking, alcohol consumption, HTN, COPD, SBP, DBP, HR, HDL, LDL, CRP, A1C, BUN, creatinine, hemoglobin, sodium level, on statin) was tested. A state-of-the-art deep learning models achieved high accuracy for predicting mortality in HF patients with an AUC of 0.96 (95% CI: 0.95-0.99) in the first cohort and AUC of 0.93 (95% CI: 0.91-0.96) in validation cohort. Conclusions: A deep neural network model has shown to have high predictive accuracy and discriminative and calibrative power for prediction of HF mortality. Further research can delineate the clinical implications of DL in predicting HF mortality.


Sign in / Sign up

Export Citation Format

Share Document