scholarly journals Performance and Robustness of Machine Learning-based Radiomic COVID-19 Severity Prediction

Author(s):  
Stephen S.F. Yip ◽  
Zan Klanecek ◽  
Shotaro Naganawa ◽  
John Kim ◽  
Andrej Studen ◽  
...  

Objectives: This study investigated the performance and robustness of radiomics in predicting COVID-19 severity in a large public cohort. Methods: A public dataset of 1110 COVID-19 patients (1 CT/patient) was used. Using CTs and clinical data, each patient was classified into mild, moderate, and severe by two observers: (1) dataset provider and (2) a board-certified radiologist. For each CT, 107 radiomic features were extracted. The dataset was randomly divided into a training (60%) and holdout validation (40%) set. During training, features were selected and combined into a logistic regression model for predicting severe cases from mild and moderate cases. The models were trained and validated on the classifications by both observers. AUC quantified the predictive power of models. To determine model robustness, the trained models was cross-validated on the inter-observer classifications. Results: A single feature alone was sufficient to predict mild from severe COVID-19 with 〖AUC〗_valid^provider=0.85 and 〖AUC〗_valid^radiologist=0.74 (p<<0.01). The most predictive features were the distribution of small size-zones (GLSZM-SmallAreaEmphasis) for provider classification and linear dependency of neighboring voxels (GLCM-Correlation) for radiologist classification. Cross-validation showed that both 〖AUC〗_valid^ ≈0.80 (p<<0.01). In predicting moderate from severe COVID-19, first-order-Median alone had sufficient predictive power of 〖AUC〗_valid^provider=0.65 (p=0.01). For radiologist classification, the predictive power of the model increased to 〖AUC〗_valid^radiologist=0.66 (p<<0.01) as the number of features grew from 1 to 5. Cross-validation yielded 〖AUC〗_valid^radiologist=0.63 (p=0.002) and 〖AUC〗_valid^provider=0.60 (p=0.09). Conclusions: Radiomics significantly predicted different levels of COVID-19 severity. The prediction was moderately sensitive to inter-observer classifications, and thus need to be used with caution.

2019 ◽  
Vol 15 (1) ◽  
pp. 258-264 ◽  
Author(s):  
Hamid Reza Ghaieni ◽  
Saeed Tavangar ◽  
Mohammad Moein Ebrahimzadeh Qhomi

Purpose The purpose of this paper is to present simple correlation for calculating nitrated hydroxyl-terminated polybutadiene (NHTPB) enthalpy of formation. Design/methodology/approach It uses multiple linear regression methods. Findings The proposed correlation has determination coefficient 0.96. The correlation has root mean square deviation and the average absolute deviations values 53.4 and 46.1 respectively. Originality/value The predictive power of correlation is checked by cross-validation method (R2=0.96, Q L O O 2 = 0.96 ).


Author(s):  
Wenjie Liu ◽  
Shanshan Wang ◽  
Xin Chen ◽  
He Jiang

In software maintenance process, it is a fairly important activity to predict the severity of bug reports. However, manually identifying the severity of bug reports is a tedious and time-consuming task. So developing automatic judgment methods for predicting the severity of bug reports has become an urgent demand. In general, a bug report contains a lot of descriptive natural language texts, thus resulting in a high-dimensional feature set which poses serious challenges to traditionally automatic methods. Therefore, we attempt to use automatic feature selection methods to improve the performance of the severity prediction of bug reports. In this paper, we introduce a ranking-based strategy to improve existing feature selection algorithms and propose an ensemble feature selection algorithm by combining existing ones. In order to verify the performance of our method, we run experiments over the bug reports of Eclipse and Mozilla and conduct comparisons with eight commonly used feature selection methods. The experiment results show that the ranking-based strategy can effectively improve the performance of the severity prediction of bug reports by up to 54.76% on average in terms of [Formula: see text]-measure, and it also can significantly reduce the dimension of the feature set. Meanwhile, the ensemble feature selection method can get better results than a single feature selection algorithm.


2017 ◽  
Vol 65 (2) ◽  
pp. 233-245
Author(s):  
Y. Wang ◽  
M. Sun ◽  
S. Du ◽  
Z. Chen

Abstract Target manoeuvre is one of the key factors affecting guidance accuracy. To intercept highly maneuverable targets, a second-order sliding-mode guidance law, which is based on the super-twisting algorithm, is designed without depending on any information about the target motion. In the designed guidance system, the target estimator plays an essential role. Besides the existing higher-order sliding-mode observer (HOSMO), a first-order linear observer (FOLO) is also proposed to estimate the target manoeuvre, and this is the major contribution of this paper. The closed-loop guidance system can be guaranteed to be uniformly ultimately bounded (UUB) in the presence of the FOLO. The comparative simulations are carried out to investigate the overall performance resulting from these two categories of observers. The results show that the guidance law with the proposed linear observer can achieve better comprehensive criteria for the amplitude of normalised acceleration and elevator deflection requirements. The reasons for the different levels of performance of these two observer-based methods are thoroughly investigated.


1983 ◽  
Vol 20 (4) ◽  
pp. 433-438 ◽  
Author(s):  
V. Srinivasan ◽  
Arun K. Jain ◽  
Naresh K. Malhotra

The prediction of first choice preferences by the full-profile method of conjoint analysis can be improved significantly by imposing constraints on parameters based on a priori knowledge of the ordering of part worths for different levels of an attribute. Constrained estimation however, has little effect on the predictive validity of the tradeoff method because the preference judgments within rows (or columns) of tradeoff tables have largely the same role as the constraints.


2018 ◽  
Vol 33 (3) ◽  
pp. 835-855 ◽  
Author(s):  
William R. Ryerson ◽  
Joshua P. Hacker

Abstract This work develops and tests the viability of obtaining skillful short-range (&lt;20 h) visibility predictions using statistical postprocessing of a 4-km, 10-member Weather Research and Forecasting (WRF) ensemble configured to closely match the U.S. Air Force Mesoscale Ensemble Forecast System. The raw WRF predictions produce excessive forecasts of zero cloud water, which is simultaneously predicted by all ensemble members in 62% of observed fog cases, leading to zero ensemble dispersion and no skill in these cases. Adding dispersion to the clear cases by making upward adjustments to cloud water predictions from individual members not predicting fog on their own provides the best chance to increase the resolution and reliability of the ensemble. The technique leverages traits of a joint parameter space in the predictions and is generally most effective when the space is defined with a moisture parameter and a low-level stability parameter. Cross-validation shows that the method adds significant overnight skill to predictions in valley and coastal regions compared to the raw WRF forecasts, with modest skill increases after sunrise. Postprocessing does not improve the highly skillful raw WRF predictions at the mountain test sites. Since the framework addresses only systematic WRF deficiencies and identifies parameter pairs with a clear, non-site-specific physical mechanism of predictive power, it has geographical transferability with less need for recalibration or observational record compared to other statistical postprocessing approaches.


2002 ◽  
Vol 24 (2) ◽  
pp. 133-150 ◽  
Author(s):  
Susan A. Jackson ◽  
Robert C. Eklund

The Flow State Scale-2 (FSS-2) and Dispositional Flow Scale-2 (DFS-2) are presented as two self-report instruments designed to assess flow experiences in physical activity. Item modifications were made to the original versions of these scales in order to improve the measurement of some of the flow dimensions. Confirmatory factor analyses of an item identification and a cross-validation sample demonstrated a good fit of the new scales. There was support for both a 9-first-order factor model and a higher order model with a global flow factor. The item identification sample yielded mean item loadings on the first-order factor of .78 for the FSS-2 and .77 for the DFS-2. Reliability estimates ranged from .80 to .90 for the FSS-2, and .81 to .90 for the DFS-2. In the cross-validation sample, mean item loadings on the first-order factor were .80 for the FSS-2, and .73 for the DFS-2. Reliability estimates ranged between .80 to .92 for the FSS-2 and .78 to .86 for the DFS-2. The scales are presented as ways of assessing flow experienced within a particular event (FSS-2) or the frequency of flow experiences in chosen physical activity in general (DFS-2).


2020 ◽  
Vol 33 (4) ◽  
pp. 1071-1081
Author(s):  
ANTONIO VANKLANE RODRIGUES DE ALMEIDA ◽  
ALEXSANDRO OLIVEIRA DA SILVA ◽  
RAIMUNDO NONATO TÁVORA COSTA ◽  
JENYFFER DA SILVA GOMES SANTOS ◽  
GERÔNIMO FERREIRA DA SILVA

ABSTRACT In regions with limited water resources, efficient use of water has become increasingly essential for agricultural production. The objective of the present study was to evaluate the use of the carnauba palm bagana (leaf fibers) as an option of ground cover to reduce the use of water in irrigated radish. The study was conducted from July to October 2018 in two crop cycles in Pentecoste-CE, Brazil. The experiment was carried out in randomized blocks with split plots and four replicates, whose primary treatments consisted of five irrigation depths (50%; 75%; 100%; 125% and 150% of the evapotranspiration crop) and secondary treatments consisted of five different levels of ground cover using carnauba bagana (0%; 25%; 50%; 75% and 100% of 16 t ha-1), in a 5 x 5 interaction, totaling 100 experimental plots. The following variables were evaluated: fresh mass of shoots and tuber, plant height, number of leaves, tuber diameter and gas exchange. For tuber fresh mass in the first crop cycle, a first order model was obtained with the response surface, with linear increase of the factors irrigation depths (0.064 g plant-1) and ground cover (0.065 g plant-1), with the highest value (40.44 g plant-1) observed for the level of 150% ETcloc and 100% bagana. Application of 16 t ha-1 of carnauba bagana can be considered recommended, within the limits studied, for use in the radish crop.


2019 ◽  
Vol 23 (2) ◽  
pp. 130-141 ◽  
Author(s):  
Chun-Chang Lee ◽  
Chih-Min Liang ◽  
Yang-Tung Liu

This paper compares the predictive powers of hierarchical generalized linear modeling (HGLM), logistic regression, and discriminant analysis with regard to tenure choices between buying property and renting property by sampling the residents of the Greater Taipei area. The results imply that the hit rate and other indicators included in HGLM have better predictive power with regard to tenure choices than the binary logistic regression model and the discriminant analysis model. That is, using HGLM to process nested data can increase prediction accuracy regarding household tenure choices. Furthermore, cross-validation is performed to analyze hit rate stability. The hit rate sequencing from this cross-validation is found to be consistent with the HGLM results, implying that the comparison of the three models in terms of hit rate performance prediction in this study is stable and reliable.


This paper provides a result assessment of traditional JPEG picture extraction function steganalysis compared to a cross-validation picture. Four distinct algorithms are used as steganographic systems in the spatial and transform domain. They are LSB Matching, LSB Replacement, Pixel Value Differencing and F5.A 25 percentage of embedding with text embedding information is considered in this paper. The characteristics regarded for evaluation are the First Order, Second Order, Extended DCT characteristics, and Markov characteristics. Support Vector Machine is the classifier used here. In statistical recovery, six distinct kernels and four distinct sampling techniques are used for evaluation.


Molecules ◽  
2021 ◽  
Vol 26 (19) ◽  
pp. 6003
Author(s):  
Dávid Virág ◽  
Tibor Kremmer ◽  
Kende Lőrincz ◽  
Norbert Kiss ◽  
Antal Jobbágy ◽  
...  

A high-resolution HILIC-MS/MS method was developed to analyze anthranilic acid derivatives of N-glycans released from human serum alpha-1-acid glycoprotein (AGP). The method was applied to samples obtained from 18 patients suffering from high-risk malignant melanoma as well as 19 healthy individuals. It enabled the identification of 102 glycan isomers separating isomers that differ only in sialic acid linkage (α-2,3, α-2,6) or in fucose positions (core, antenna). Comparative assessment of the samples revealed that upregulation of certain fucosylated glycans and downregulation of their nonfucosylated counterparts occurred in cancer patients. An increased ratio of isomers with more α-2,6-linked sialic acids was also observed. Linear discriminant analysis (LDA) combining 10 variables with the highest discriminatory power was employed to categorize the samples based on their glycosylation pattern. The performance of the method was tested by cross-validation, resulting in an overall classification success rate of 96.7%. The approach presented here is significantly superior to serological marker S100B protein in terms of sensitivity and negative predictive power in the population studied. Therefore, it may effectively support the diagnosis of malignant melanoma as a biomarker.


Sign in / Sign up

Export Citation Format

Share Document