Critical event prediction for proactive management in large-scale computer clusters

Author(s):  
R. K. Sahoo ◽  
A. J. Oliner ◽  
I. Rish ◽  
M. Gupta ◽  
J. E. Moreira ◽  
...  
2021 ◽  
Author(s):  
Stavros-Andreas Logothetis ◽  
Vasileios Salamalikis ◽  
Stefan Wilbert ◽  
Jan Remund ◽  
Luis Zarzalejo ◽  
...  

<p>Cloud cameras (all sky imagers/ASIs) can be used for short-term (next 20 min) forecasts of solar irradiance. For this reason, several experimental and operational solutions emerged in the last decade with different approaches in terms of instrument types and forecast algorithms. Moreover, few commercial and semi-prototype systems are already available or being investigated. So far, the uncertainty of the predictions cannot be fully compared, as previously published tests were carried out during different periods and at different locations. In this study, the results from a benchmark exercise are presented in order to qualify the current ASI-based short-term forecasting solutions and examine their accuracy. This first comparative measurement campaign carried out as part of the IEA PVPS Task 16 (https://iea-pvps.org/research-tasks/solar-resource-for-high-penetration-and-large-scale-applications/). A 3-month observation campaign (from August to December 2019) took place at Plataforma Solar de Almeria of the Spanish research center CIEMAT including five different ASI systems and a network of high-quality measurements of solar irradiance and other atmospheric parameters. Forecasted time-series of global horizontal irradiance are compared with ground-based measurements and two persistence models to identify strengths and weaknesses of each approach and define best practices of ASI-based forecasts. The statistical analysis is divided into seven cloud classes to interpret the different cloud type effect on ASIs forecast accuracy. For every cloud cluster, at least three ASIs outperform persistence models, in terms of forecast error, highlighting their performance capabilities. The feasibility of ASIs on ramp event detection is also investigated, applying different approaches of ramp event prediction. The revealed findings are promising in terms of overall performance of ASIs as well as their forecasting capabilities in ramp detection.  </p>


Author(s):  
Zhongyang Li ◽  
Xiao Ding ◽  
Ting Liu

Script event prediction requires a model to predict the subsequent event given an existing event context. Previous models based on event pairs or event chains cannot make full use of dense event connections, which may limit their capability of event prediction. To remedy this, we propose constructing an event graph to better utilize the event network information for script event prediction. In particular, we first extract narrative event chains from large quantities of news corpus, and then construct a narrative event evolutionary graph (NEEG) based on the extracted chains. NEEG can be seen as a knowledge base that describes event evolutionary principles and patterns. To solve the inference problem on NEEG, we present a scaled graph neural network (SGNN) to model event interactions and learn better event representations. Instead of computing the representations on the whole graph, SGNN processes only the concerned nodes each time, which makes our model feasible to large-scale graphs. By comparing the similarity between input context event representations and candidate event representations, we can choose the most reasonable subsequent event. Experimental results on widely used New York Times corpus demonstrate that our model significantly outperforms state-of-the-art baseline methods, by using standard multiple choice narrative cloze evaluation.


2018 ◽  
Vol 1 (1) ◽  
Author(s):  
Søren Egedorf ◽  
Hamid Reza Shaker ◽  
Rodney A. Martin ◽  
Bo Nørregaard Jørgensen

Blood ◽  
2017 ◽  
Vol 130 (3) ◽  
pp. 258-266 ◽  
Author(s):  
Robert Kridel ◽  
Laurie H. Sehn ◽  
Randy D. Gascoyne

Abstract Transformation to aggressive lymphoma is a critical event in the clinical course of follicular lymphoma (FL) patients. Yet, it is a challenge to reliably predict transformation at the time of diagnosis. Understanding the risk of transformation would be useful for guiding and monitoring patients, as well as for evaluating novel treatment strategies that could potentially prevent transformation. Herein, we review the contribution of clinical, pathological, and genetic risk factors to transformation. Patients with multiple clinical high-risk factors are at elevated risk of transformation but we are currently lacking a prognostic index that would specifically address transformation rather than disease progression or overall survival. From the biological standpoint, multiple studies have correlated individual biomarkers with transformation. However, accurate prediction of this event is currently hampered by our limited knowledge of the evolutionary pathways leading to transformation, as well as the scarcity of comprehensive, large-scale studies that assess both the genomic landscape of alterations within tumor cells and the composition of the microenvironment. Liquid biopsies hold great promise for achieving precision medicine. Indeed, mutations detected within circulating tumor DNA may be a better reflection of the inherent intratumoral heterogeneity than the biopsy of a single site. Last, we will assess whether evidence exists in the literature that transformation might be prevented altogether, based on the choice of therapy for FL.


2019 ◽  
Author(s):  
Oliver Chalkley ◽  
Oliver Purcell ◽  
Claire Grierson ◽  
Lucia Marucci

AbstractMotivationComputational biology is a rapidly developing field, and in-silico methods are being developed to aid the design of genomes to create cells with optimised phenotypes. Two barriers to progress are that in-silico methods are often only developed on a particular implementation of a specific model (e.g. COBRA metabolic models) and models with longer simulation time inhibit the large-scale in-silico experiments required to search the vast solution space of genome combinations.ResultsHere we present the genome design suite (PyGDS) which is a suite of Python tools to aid the development of in-silico genome design methods. PyGDS provides a framework with which to implement phenotype optimisation algorithms on computational models across computer clusters. The framework is abstract allowing it to be adapted to utilise different computer clusters, optimisation algorithms, or design goals. It implements an abstract multi-generation algorithm structure allowing algorithms to avoid maximum simulation times on clusters and enabling iterative learning in the algorithm. The initial case study will be genome reduction algorithms on a whole-cell model of Mycoplasma genitalium for a PBS/Torque cluster and a Slurm cluster.AvailabilityThe genome design suite is written in Python for Linux operating systems and is available from GitHub on a GPL open-source [email protected], [email protected], and [email protected].


GigaScience ◽  
2020 ◽  
Vol 9 (6) ◽  
Author(s):  
Michael Kluge ◽  
Marie-Sophie Friedl ◽  
Amrei L Menzel ◽  
Caroline C Friedel

Abstract Background Advances in high-throughput methods have brought new challenges for biological data analysis, often requiring many interdependent steps applied to a large number of samples. To address this challenge, workflow management systems, such as Watchdog, have been developed to support scientists in the (semi-)automated execution of large analysis workflows. Implementation Here, we present Watchdog 2.0, which implements new developments for module creation, reusability, and documentation and for reproducibility of analyses and workflow execution. Developments include a graphical user interface for semi-automatic module creation from software help pages, sharing repositories for modules and workflows, and a standardized module documentation format. The latter allows generation of a customized reference book of public and user-specific modules. Furthermore, extensive logging of workflow execution, module and software versions, and explicit support for package managers and container virtualization now ensures reproducibility of results. A step-by-step analysis protocol generated from the log file may, e.g., serve as a draft of a manuscript methods section. Finally, 2 new execution modes were implemented. One allows resuming workflow execution after interruption or modification without rerunning successfully executed tasks not affected by changes. The second one allows detaching and reattaching to workflow execution on a local computer while tasks continue running on computer clusters. Conclusions Watchdog 2.0 provides several new developments that we believe to be of benefit for large-scale bioinformatics analysis and that are not completely covered by other competing workflow management systems. The software itself, module and workflow repositories, and comprehensive documentation are freely available at https://www.bio.ifi.lmu.de/watchdog.


10.2196/24018 ◽  
2020 ◽  
Vol 22 (11) ◽  
pp. e24018 ◽  
Author(s):  
Akhil Vaid ◽  
Sulaiman Somani ◽  
Adam J Russak ◽  
Jessica K De Freitas ◽  
Fayzan F Chaudhry ◽  
...  

Background COVID-19 has infected millions of people worldwide and is responsible for several hundred thousand fatalities. The COVID-19 pandemic has necessitated thoughtful resource allocation and early identification of high-risk patients. However, effective methods to meet these needs are lacking. Objective The aims of this study were to analyze the electronic health records (EHRs) of patients who tested positive for COVID-19 and were admitted to hospitals in the Mount Sinai Health System in New York City; to develop machine learning models for making predictions about the hospital course of the patients over clinically meaningful time horizons based on patient characteristics at admission; and to assess the performance of these models at multiple hospitals and time points. Methods We used Extreme Gradient Boosting (XGBoost) and baseline comparator models to predict in-hospital mortality and critical events at time windows of 3, 5, 7, and 10 days from admission. Our study population included harmonized EHR data from five hospitals in New York City for 4098 COVID-19–positive patients admitted from March 15 to May 22, 2020. The models were first trained on patients from a single hospital (n=1514) before or on May 1, externally validated on patients from four other hospitals (n=2201) before or on May 1, and prospectively validated on all patients after May 1 (n=383). Finally, we established model interpretability to identify and rank variables that drive model predictions. Results Upon cross-validation, the XGBoost classifier outperformed baseline models, with an area under the receiver operating characteristic curve (AUC-ROC) for mortality of 0.89 at 3 days, 0.85 at 5 and 7 days, and 0.84 at 10 days. XGBoost also performed well for critical event prediction, with an AUC-ROC of 0.80 at 3 days, 0.79 at 5 days, 0.80 at 7 days, and 0.81 at 10 days. In external validation, XGBoost achieved an AUC-ROC of 0.88 at 3 days, 0.86 at 5 days, 0.86 at 7 days, and 0.84 at 10 days for mortality prediction. Similarly, the unimputed XGBoost model achieved an AUC-ROC of 0.78 at 3 days, 0.79 at 5 days, 0.80 at 7 days, and 0.81 at 10 days. Trends in performance on prospective validation sets were similar. At 7 days, acute kidney injury on admission, elevated LDH, tachypnea, and hyperglycemia were the strongest drivers of critical event prediction, while higher age, anion gap, and C-reactive protein were the strongest drivers of mortality prediction. Conclusions We externally and prospectively trained and validated machine learning models for mortality and critical events for patients with COVID-19 at different time horizons. These models identified at-risk patients and uncovered underlying relationships that predicted outcomes.


Sign in / Sign up

Export Citation Format

Share Document