scholarly journals Decision Tree Approach for Soil Liquefaction Assessment

2013 ◽  
Vol 2013 ◽  
pp. 1-8 ◽  
Author(s):  
Amir H. Gandomi ◽  
Mark M. Fridline ◽  
David A. Roke

In the current study, the performances of some decision tree (DT) techniques are evaluated for postearthquake soil liquefaction assessment. A database containing 620 records of seismic parameters and soil properties is used in this study. Three decision tree techniques are used here in two different ways, considering statistical and engineering points of view, to develop decision rules. The DT results are compared to the logistic regression (LR) model. The results of this study indicate that the DTs not only successfully predict liquefaction but they can also outperform the LR model. The best DT models are interpreted and evaluated based on an engineering point of view.

2015 ◽  
Vol 54 (06) ◽  
pp. 560-567 ◽  
Author(s):  
K. Zhu ◽  
Z. Lou ◽  
J. Zhou ◽  
N. Ballester ◽  
P. Parikh ◽  
...  

SummaryIntroduction: This article is part of the Focus Theme of Methods of Information in Medicine on “Big Data and Analytics in Healthcare”.Background: Hospital readmissions raise healthcare costs and cause significant distress to providers and patients. It is, therefore, of great interest to healthcare organizations to predict what patients are at risk to be readmitted to their hospitals. However, current logistic regression based risk prediction models have limited prediction power when applied to hospital administrative data. Meanwhile, although decision trees and random forests have been applied, they tend to be too complex to understand among the hospital practitioners.Objectives: Explore the use of conditional logistic regression to increase the prediction accuracy.Methods: We analyzed an HCUP statewide in-patient discharge record dataset, which includes patient demographics, clinical and care utilization data from California. We extracted records of heart failure Medicare beneficiaries who had inpatient experience during an 11-month period. We corrected the data imbalance issue with under-sampling. In our study, we first applied standard logistic regression and decision tree to obtain influential variables and derive practically meaning decision rules. We then stratified the original data set accordingly and applied logistic regression on each data stratum. We further explored the effect of interacting variables in the logistic regression modeling. We conducted cross validation to assess the overall prediction performance of conditional logistic regression (CLR) and compared it with standard classification models.Results: The developed CLR models outperformed several standard classification models (e.g., straightforward logistic regression, stepwise logistic regression, random forest, support vector machine). For example, the best CLR model improved the classification accuracy by nearly 20% over the straightforward logistic regression model. Furthermore, the developed CLR models tend to achieve better sensitivity of more than 10% over the standard classification models, which can be translated to correct labeling of additional 400 – 500 readmissions for heart failure patients in the state of California over a year. Lastly, several key predictor identified from the HCUP data include the disposition location from discharge, the number of chronic conditions, and the number of acute procedures.Conclusions: It would be beneficial to apply simple decision rules obtained from the decision tree in an ad-hoc manner to guide the cohort stratification. It could be potentially beneficial to explore the effect of pairwise interactions between influential predictors when building the logistic regression models for different data strata. Judicious use of the ad-hoc CLR models developed offers insights into future development of prediction models for hospital readmissions, which can lead to better intuition in identifying high-risk patients and developing effective post-discharge care strategies. Lastly, this paper is expected to raise the awareness of collecting data on additional markers and developing necessary database infrastructure for larger-scale exploratory studies on readmission risk prediction.


2008 ◽  
pp. 2978-2992
Author(s):  
Jianting Zhang ◽  
Wieguo Liu ◽  
Le Gruenwald

Decision trees (DT) has been widely used for training and classification of remotely sensed image data due to its capability to generate human interpretable decision rules and its relatively fast speed in training and classification. This chapter proposes a successive decision tree (SDT) approach where the samples in the ill-classified branches of a previous resulting decision tree are used to construct a successive decision tree. The decision trees are chained together through pointers and used for classification. SDT aims at constructing more interpretable decision trees while attempting to improve classification accuracies. The proposed approach is applied to two real remotely sensed image datasets for evaluations in terms of classification accuracy and interpretability of the resulting decision rules.


Author(s):  
Jianting Zhang ◽  
Wieguo Liu ◽  
Le Gruenwald

Decision trees (DT) has been widely used for training and classification of remotely sensed image data due to its capability to generate human interpretable decision rules and its relatively fast speed in training and classification. This chapter proposes a successive decision tree (SDT) approach where the samples in the ill-classified branches of a previous resulting decision tree are used to construct a successive decision tree. The decision trees are chained together through pointers and used for classification. SDT aims at constructing more interpretable decision trees while attempting to improve classification accuracies. The proposed approach is applied to two real remotely sensed image datasets for evaluations in terms of classification accuracy and interpretability of the resulting decision rules.


2016 ◽  
Vol 20 (9) ◽  
pp. 1895-1903 ◽  
Author(s):  
Leónides Fernández ◽  
Pilar Mediano ◽  
Ricardo García ◽  
Juan M. Rodríguez ◽  
María Marín

2019 ◽  
Vol 9 (20) ◽  
pp. 4226 ◽  
Author(s):  
Mahmood Ahmad ◽  
Xiao-Wei Tang ◽  
Jiang-Nan Qiu ◽  
Feezan Ahmad

Liquefaction is considered a damaging phenomenon of earthquakes and a major cause of concern in civil engineering. Therefore, its predictory assessment is an essential task for geotechnical experts. This paper investigates the performance of Bayesian belief network (BBN) and C4.5 decision tree (DT) models to evaluate seismic soil liquefaction potential based on the updated and relatively large cone penetration test (CPT) dataset (which includes 251 case histories), comparing them to a simplified procedure and an evolutionary-based approach. The BBN model was developed using the K2 machine learning algorithm and domain knowledge (DK) with data fusion methodology, while the DT model was created using a C4.5 algorithm. This study shows that the BBN model is preferred over the others for evaluation of seismic soil liquefaction potential. Owing to its overall performance, simplicity in practice, data-driven characteristics, and ability to map interactions between variables, the use of a BBN model in assessing seismic soil liquefaction is quite promising. The results of a sensitivity analysis show that ‘equivalent clean sand penetration resistance’ is the most significant factor affecting liquefaction potential. This study also interprets the probabilistic reasoning of the robust BBN model and most probable explanation (MPE) of seismic soil liquefied sites, based on an engineering point of view.


In the article the analysis of nonsense, absurdity and paradox from the standpoint of linguistics is giv-en. Different points of view on these categories in relation to the meaning are considered. An attempt is made to reveal the commonality and specificity of nonsense, absurdity and paradox. Some researchers consider nonsense and paradox as a kind of absurdity. There is a dichotomous point of view on nonsense as one of the components of absurdity. However, there are works where these categories are differentiat-ed, for example, absurdity is understood as an ontological category, and nonsense as an epistemological category. There is a view of these categories through the allocation of "non-sense", "out-sense" and 136 "counter-sense" there is also a view that in the case of nonsense we are talking about the incompatibility of representations, and in the case of absurdity-the incompatibility of objects. If there are criteria that allow us to consider the presence of this phenomenon as natural, absurdity ceases to exist. Consequently, the view is expressed that nonsense, absurdity and paradox are different categories of thinking. Paradox is a contradiction arising from the presence of two or more common sense. The absurdity can be seen as a" counter-sense» opposing common sense and putting forward the concept of active impossibility of the latter's existence. As for nonsense, it is the meaning of metaphysical level – a meaning that goes beyond the ordinary meaning and creates new meanings. It is concluded that nonsense, absurdity and paradox are independent categories of human thinking, which is a manifestation of the cognitive function of hu-man consciousness.


Author(s):  
Dhilsath Fathima.M ◽  
S. Justin Samuel ◽  
R. Hari Haran

Aim: This proposed work is used to develop an improved and robust machine learning model for predicting Myocardial Infarction (MI) could have substantial clinical impact. Objectives: This paper explains how to build machine learning based computer-aided analysis system for an early and accurate prediction of Myocardial Infarction (MI) which utilizes framingham heart study dataset for validation and evaluation. This proposed computer-aided analysis model will support medical professionals to predict myocardial infarction proficiently. Methods: The proposed model utilize the mean imputation to remove the missing values from the data set, then applied principal component analysis to extract the optimal features from the data set to enhance the performance of the classifiers. After PCA, the reduced features are partitioned into training dataset and testing dataset where 70% of the training dataset are given as an input to the four well-liked classifiers as support vector machine, k-nearest neighbor, logistic regression and decision tree to train the classifiers and 30% of test dataset is used to evaluate an output of machine learning model using performance metrics as confusion matrix, classifier accuracy, precision, sensitivity, F1-score, AUC-ROC curve. Results: Output of the classifiers are evaluated using performance measures and we observed that logistic regression provides high accuracy than K-NN, SVM, decision tree classifiers and PCA performs sound as a good feature extraction method to enhance the performance of proposed model. From these analyses, we conclude that logistic regression having good mean accuracy level and standard deviation accuracy compared with the other three algorithms. AUC-ROC curve of the proposed classifiers is analyzed from the output figure.4, figure.5 that logistic regression exhibits good AUC-ROC score, i.e. around 70% compared to k-NN and decision tree algorithm. Conclusion: From the result analysis, we infer that this proposed machine learning model will act as an optimal decision making system to predict the acute myocardial infarction at an early stage than an existing machine learning based prediction models and it is capable to predict the presence of an acute myocardial Infarction with human using the heart disease risk factors, in order to decide when to start lifestyle modification and medical treatment to prevent the heart disease.


Sign in / Sign up

Export Citation Format

Share Document