scholarly journals Multiscale classification of heart failure phenotypes by unsupervised clustering of unstructured electronic medical record data

2020 ◽  
Vol 10 (1) ◽  
Author(s):  
Tasha Nagamine ◽  
Brian Gillette ◽  
Alexey Pakhomov ◽  
John Kahoun ◽  
Hannah Mayer ◽  
...  

AbstractAs a leading cause of death and morbidity, heart failure (HF) is responsible for a large portion of healthcare and disability costs worldwide. Current approaches to define specific HF subpopulations may fail to account for the diversity of etiologies, comorbidities, and factors driving disease progression, and therefore have limited value for clinical decision making and development of novel therapies. Here we present a novel and data-driven approach to understand and characterize the real-world manifestation of HF by clustering disease and symptom-related clinical concepts (complaints) captured from unstructured electronic health record clinical notes. We used natural language processing to construct vectorized representations of patient complaints followed by clustering to group HF patients by similarity of complaint vectors. We then identified complaints that were significantly enriched within each cluster using statistical testing. Breaking the HF population into groups of similar patients revealed a clinically interpretable hierarchy of subgroups characterized by similar HF manifestation. Importantly, our methodology revealed well-known etiologies, risk factors, and comorbid conditions of HF (including ischemic heart disease, aortic valve disease, atrial fibrillation, congenital heart disease, various cardiomyopathies, obesity, hypertension, diabetes, and chronic kidney disease) and yielded additional insights into the details of each HF subgroup’s clinical manifestation of HF. Our approach is entirely hypothesis free and can therefore be readily applied for discovery of novel insights in alternative diseases or patient populations.

2021 ◽  
Vol 108 (Supplement_7) ◽  
Author(s):  
T Fraser ◽  
K Emslie ◽  
T Wheatley

Abstract Aim Anecdotal experience suggested patients undergoing diagnostic laparoscopy for appendicitis faced long pre-operative waiting periods. This project was undertaken to quantify the situation and determine how long patients were waiting for surgery. National guidelines recommend patients can reasonably wait 6-18 hours for surgery, as per NCEPOD classification of intervention (appendicitis considered to fall under urgent 2b category). Method A retrospective audit of 39 patients listed for diagnostic laparoscopy or appendicectomy during November 2019 was performed. Clinical notes and electronic records were reviewed to determine the timeline of clinical decision making and patient’s arrival in theatre. Operative and histopathology findings were also noted. Results The majority of patients (61%) underwent surgery within 18 hours of a documented decision to operate. Average wait was 17.5 hrs (mean). Longest wait was 41 hours (excluding isolated outlier). Documentation of decision to operate was noted to be poor (undocumented in 5 cases) and in some cases the patient was booked before a documented decision to operate. Variation between the operative and histopathological diagnosis of appendicitis was apparent. Conclusions The findings are re-assuring that once the decision was made to operate most patients had surgery within 18 hours. However, there is still room for improvement with regards to timeliness and documentation. The discrepancy between operative and histological findings highlight the challenge of diagnosing appendicitis accurately.


2021 ◽  
Author(s):  
Julius M Kernbach ◽  
Daniel Delev ◽  
Georg Neuloh ◽  
Hans Clusmann ◽  
Danilo Bzdok ◽  
...  

Background The current WHO classification integrates histological and molecular features of brain tumors. The aim of this study was to identify generalizable topological patterns with the potential to add an anatomical dimension to the classification of brain tumors. Methods We applied non-negative matrix factorization as an unsupervised pattern discovery strategy to the fine-grained topographic tumor profiles of 936 patients with primary and secondary brain tumors. From the anatomical features alone, this machine learning algorithm enabled the extraction of latent topological tumor patterns, termed meta-topologies. The optimal parts-based representation was automatically determined in 10,000 split-half iterations. We further characterized each meta-topologys unique histopathologic profile and survival probability, thus linking important biological and clinical information to the underlying anatomical patterns Results In primary brain tumors, six meta-topologies were extracted, each detailing a transpallial pattern with distinct parenchymal and ventricular compositions. We identified one infratentorial, one allopallial, three neopallial (parieto-occipital, frontal, temporal) and one unisegmental meta-topology. Each meta-topology mapped to distinct histopathologic and molecular profiles. The unisegmental meta-topology showed the strongest anatomical-clinical link demonstrating a survival advantage in histologically identical tumors. Brain metastases separated to an infra- and supratentorial meta-topology with anatomical patterns highlighting their affinity to the cortico-subcortical boundary of arterial watershed areas. Conclusions Using a novel data-driven approach, we identified generalizable topological patterns in both primary and secondary brain tumors Differences in the histopathologic profiles and prognosis of these anatomical tumor classes provide insights into the heterogeneity of tumor biology and might add to personalized clinical decision making.


Assessment ◽  
2021 ◽  
pp. 107319112199646
Author(s):  
Olivia Gratz ◽  
Duncan Vos ◽  
Megan Burke ◽  
Neelkamal Soares

To date, there is a paucity of research conducting natural language processing (NLP) on the open-ended responses of behavior rating scales. Using three NLP lexicons for sentiment analysis of the open-ended responses of the Behavior Assessment System for Children-Third Edition, the researchers discovered a moderately positive correlation between the human composite rating and the sentiment score using each of the lexicons for strengths comments and a slightly positive correlation for the concerns comments made by guardians and teachers. In addition, the researchers found that as the word count increased for open-ended responses regarding the child’s strengths, there was a greater positive sentiment rating. Conversely, as word count increased for open-ended responses regarding child concerns, the human raters scored comments more negatively. The authors offer a proof-of-concept to use NLP-based sentiment analysis of open-ended comments to complement other data for clinical decision making.


2021 ◽  
Vol 126 (3) ◽  
pp. 365-379
Author(s):  
Gianluca Pontone ◽  
Ernesto Di Cesare ◽  
Silvia Castelletti ◽  
Francesco De Cobelli ◽  
Manuel De Lazzari ◽  
...  

AbstractCardiac magnetic resonance (CMR) has emerged as new mainstream technique for the evaluation of patients with cardiac diseases, providing unique information to support clinical decision-making. This document has been developed by a joined group of experts of the Italian Society of Cardiology and Italian society of Radiology and aims to produce an updated consensus statement about the current state of technology and clinical applications of CMR. The writing committee consisted of members and experts of both societies who worked jointly to develop a more integrated approach in the field of cardiac radiology. Part 1 of the document will cover ischemic heart disease, congenital heart disease, cardio-oncology, cardiac masses and heart transplant.


Circulation ◽  
2017 ◽  
Vol 135 (suppl_1) ◽  
Author(s):  
Elizabeth J Bell ◽  
Jennifer L St. Sauver ◽  
Veronique L Roger ◽  
Nicholas B Larson ◽  
Hongfang Liu ◽  
...  

Introduction: Proton pump inhibitors (PPIs) are used by an estimated 29 million Americans. PPIs increase the levels of asymmetrical dimethylarginine, a known risk factor for cardiovascular disease (CVD). Data from a select population of patients with CVD suggest that PPI use is associated with an increased risk of stroke, heart failure, and coronary heart disease. The impact of PPI use on incident CVD is largely unknown in the general population. Hypothesis: We hypothesized that PPI users have a higher risk of incident total CVD, coronary heart disease, stroke, and heart failure compared to nonusers. To demonstrate specificity of association, we additionally hypothesized that there is not an association between use of H 2 -blockers - another commonly used class of medications with similar indications as PPIs - and CVD. Methods: We used the Rochester Epidemiology Project’s medical records-linkage system to identify all residents of Olmsted County, MN on our baseline date of January 1, 2004 (N=140217). We excluded persons who did not grant permission for their records to be used for research, were <18 years old, had a history of CVD, had missing data for any variable included in our model, or had evidence of PPI use within the previous year.We followed our final cohort (N=58175) for up to 12 years. The administrative censoring date for CVD was 1/20/2014, for coronary heart disease was 8/3/2016, for stroke was 9/9/2016, and for heart failure was 1/20/2014. Time-varying PPI ever-use was ascertained using 1) natural language processing to capture unstructured text from the electronic health record, and 2) outpatient prescriptions. An incident CVD event was defined as the first occurrence of 1) validated heart failure, 2) validated coronary heart disease, or 3) stroke, defined using diagnostic codes only. As a secondary analysis, we calculated the association between time-varying H 2 -blocker ever-use and CVD among persons not using H 2 -blockers at baseline. Results: After adjustment for age, sex, race, education, hypertension, hyperlipidemia, diabetes, and body-mass-index, PPI use was associated with an approximately 50% higher risk of CVD (hazard ratio [95% CI]: 1.51 [1.37-1.67]; 2187 CVD events), stroke (hazard ratio [95% CI]: 1.49 [1.35-1.65]; 1928 stroke events), and heart failure (hazard ratio [95% CI]: 1.56 [1.23-1.97]; 353 heart failure events) compared to nonusers. Users of PPIs had a 35% greater risk of coronary heart disease than nonusers (95% CI: 1.13-1.61; 626 coronary heart disease events). Use of H 2 -blockers was also associated with a higher risk of CVD (adjusted hazard ratio [95% CI]: 1.23 [1.08-1.41]; 2331 CVD events). Conclusions: PPI use is associated with a higher risk of CVD, coronary heart disease, stroke and heart failure. Use of a drug with no known cardiac toxicity - H 2 -blockers - was also associated with a greater risk of CVD, warranting further study.


2015 ◽  
Vol 22 (6) ◽  
pp. 1220-1230 ◽  
Author(s):  
Huan Mo ◽  
William K Thompson ◽  
Luke V Rasmussen ◽  
Jennifer A Pacheco ◽  
Guoqian Jiang ◽  
...  

Abstract Background Electronic health records (EHRs) are increasingly used for clinical and translational research through the creation of phenotype algorithms. Currently, phenotype algorithms are most commonly represented as noncomputable descriptive documents and knowledge artifacts that detail the protocols for querying diagnoses, symptoms, procedures, medications, and/or text-driven medical concepts, and are primarily meant for human comprehension. We present desiderata for developing a computable phenotype representation model (PheRM). Methods A team of clinicians and informaticians reviewed common features for multisite phenotype algorithms published in PheKB.org and existing phenotype representation platforms. We also evaluated well-known diagnostic criteria and clinical decision-making guidelines to encompass a broader category of algorithms. Results We propose 10 desired characteristics for a flexible, computable PheRM: (1) structure clinical data into queryable forms; (2) recommend use of a common data model, but also support customization for the variability and availability of EHR data among sites; (3) support both human-readable and computable representations of phenotype algorithms; (4) implement set operations and relational algebra for modeling phenotype algorithms; (5) represent phenotype criteria with structured rules; (6) support defining temporal relations between events; (7) use standardized terminologies and ontologies, and facilitate reuse of value sets; (8) define representations for text searching and natural language processing; (9) provide interfaces for external software algorithms; and (10) maintain backward compatibility. Conclusion A computable PheRM is needed for true phenotype portability and reliability across different EHR products and healthcare systems. These desiderata are a guide to inform the establishment and evolution of EHR phenotype algorithm authoring platforms and languages.


Sign in / Sign up

Export Citation Format

Share Document