scholarly journals Risk Prediction Models for Melanoma: A Systematic Review on the Heterogeneity in Model Development and Validation

Author(s):  
Isabelle Kaiser ◽  
Annette B. Pfahlberg ◽  
Wolfgang Uter ◽  
Markus V. Heppt ◽  
Marit B. Veierød ◽  
...  

The rising incidence of cutaneous melanoma over the past few decades has prompted substantial efforts to develop risk prediction models identifying people at high risk of developing melanoma to facilitate targeted screening programs. We review these models, regarding study characteristics, differences in risk factor selection and assessment, evaluation, and validation methods. Our systematic literature search revealed 40 studies comprising 46 different risk prediction models eligible for the review. Altogether, 35 different risk factors were part of the models with nevi being the most common one (n = 35, 78%); little consistency in other risk factors was observed. Results of an internal validation were reported for less than half of the studies (n = 18, 45%), and only 6 performed external validation. In terms of model performance, 29 studies assessed the discriminative ability of their models; other performance measures, e.g., regarding calibration or clinical usefulness, were rarely reported. Due to the substantial heterogeneity in risk factor selection and assessment as well as methodologic aspects of model development, direct comparisons between models are hardly possible. Uniform methodologic standards for the development and validation of risk prediction models for melanoma and reporting standards for the accompanying publications are necessary and need to be obligatory for that reason.

BMC Cancer ◽  
2022 ◽  
Vol 22 (1) ◽  
Author(s):  
Michele Sassano ◽  
Marco Mariani ◽  
Gianluigi Quaranta ◽  
Roberta Pastorino ◽  
Stefania Boccia

Abstract Background Risk prediction models incorporating single nucleotide polymorphisms (SNPs) could lead to individualized prevention of colorectal cancer (CRC). However, the added value of incorporating SNPs into models with only traditional risk factors is still not clear. Hence, our primary aim was to summarize literature on risk prediction models including genetic variants for CRC, while our secondary aim was to evaluate the improvement of discriminatory accuracy when adding SNPs to a prediction model with only traditional risk factors. Methods We conducted a systematic review on prediction models incorporating multiple SNPs for CRC risk prediction. We tested whether a significant trend in the increase of Area Under Curve (AUC) according to the number of SNPs could be observed, and estimated the correlation between AUC improvement and number of SNPs. We estimated pooled AUC improvement for SNP-enhanced models compared with non-SNP-enhanced models using random effects meta-analysis, and conducted meta-regression to investigate the association of specific factors with AUC improvement. Results We included 33 studies, 78.79% using genetic risk scores to combine genetic data. We found no significant trend in AUC improvement according to the number of SNPs (p for trend = 0.774), and no correlation between the number of SNPs and AUC improvement (p = 0.695). Pooled AUC improvement was 0.040 (95% CI: 0.035, 0.045), and the number of cases in the study and the AUC of the starting model were inversely associated with AUC improvement obtained when adding SNPs to a prediction model. In addition, models constructed in Asian individuals achieved better AUC improvement with the incorporation of SNPs compared with those developed among individuals of European ancestry. Conclusions Though not conclusive, our results provide insights on factors influencing discriminatory accuracy of SNP-enhanced models. Genetic variants might be useful to inform stratified CRC screening in the future, but further research is needed.


Circulation ◽  
2020 ◽  
Vol 142 (Suppl_3) ◽  
Author(s):  
Matthew W Segar ◽  
Byron Jaeger ◽  
Kershaw V Patel ◽  
Vijay Nambi ◽  
Chiadi E Ndumele ◽  
...  

Introduction: Heart failure (HF) risk and the underlying biological risk factors vary by race. Machine learning (ML) may improve race-specific HF risk prediction but this has not been fully evaluated. Methods: The study included participants from 4 cohorts (ARIC, DHS, JHS, and MESA) aged > 40 years, free of baseline HF, and with adjudicated HF event follow-up. Black adults from JHS and white adults from ARIC were used to derive race-specific ML models to predict 10-year HF risk. The ML models were externally validated in subgroups of black and white adults from ARIC (excluding JHS participants) and pooled MESA/DHS cohorts and compared to prior established HF risk scores developed in ARIC and MESA. Harrell’s C-index and Greenwood-Nam-D’Agostino chi-square were used to assess discrimination and calibration, respectively. Results: In the derivation cohorts, 288 of 4141 (7.0%) black and 391 of 8242 (4.7%) white adults developed HF over 10 years. The ML models had excellent discrimination in both black and white participants (C-indices = 0.88 and 0.89). In the external validation cohorts for black participants from ARIC (excluding JHS, N = 1072) and MESA/DHS pooled cohorts (N = 2821), 131 (12.2%) and 115 (4.1%) developed HF. The ML model had adequate calibration and demonstrated superior discrimination compared to established HF risk models (Fig A). A consistent pattern was also observed in the external validation cohorts of white participants from the MESA/DHS pooled cohorts (N=3236; 100 [3.1%] HF events) (Fig A). The most important predictors of HF in both races were NP levels. Cardiac biomarkers and glycemic parameters were most important among blacks while LV hypertrophy and prevalent CVD and traditional CV risk factors were the strongest predictors among whites (Fig B). Conclusions: Race-specific and ML-based HF risk models that integrate clinical, laboratory, and biomarker data demonstrated superior performance when compared to traditional risk prediction models.


2019 ◽  
Vol 98 (10) ◽  
pp. 1088-1095 ◽  
Author(s):  
J. Krois ◽  
C. Graetz ◽  
B. Holtfreter ◽  
P. Brinkmann ◽  
T. Kocher ◽  
...  

Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss prediction models of patients with periodontitis. Two independent cohorts (627 patients, 11,651 teeth) were followed over a mean ± SD 18.2 ± 5.6 y (Kiel cohort) and 6.6 ± 2.9 y (Greifswald cohort). Tooth loss and 10 patient- and tooth-level predictors were recorded. The impact of different model development and validation steps was evaluated: 1) model complexity (logistic regression, recursive partitioning, random forest, extreme gradient boosting), 2) sample size (full data set or 10%, 25%, or 75% of cases dropped at random), 3) prediction periods (maximum 10, 15, or 20 y or uncensored), and 4) validation schemes (internal or external by centers/time). Tooth loss was generally a rare event (880 teeth were lost). All models showed limited sensitivity but high specificity. Patients’ age and tooth loss at baseline as well as probing pocket depths showed high variable importance. More complex models (random forest, extreme gradient boosting) had no consistent advantages over simpler ones (logistic regression, recursive partitioning). Internal validation (in sample) overestimated the predictive power (area under the curve up to 0.90), while external validation (out of sample) found lower areas under the curve (range 0.62 to 0.82). Reducing the sample size decreased the predictive power, particularly for more complex models. Censoring the prediction period had only limited impact. When the model was trained in one period and tested in another, model outcomes were similar to the base case, indicating temporal validation as a valid option. No model showed higher accuracy than the no-information rate. In conclusion, none of the developed models would be useful in a clinical setting, despite high accuracy. During modeling, rigorous development and external validation should be applied and reported accordingly.


Stroke ◽  
2020 ◽  
Vol 51 (7) ◽  
pp. 2095-2102
Author(s):  
Eugene Y.H. Tang ◽  
Christopher I. Price ◽  
Louise Robinson ◽  
Catherine Exley ◽  
David W. Desmond ◽  
...  

Background and Purpose: Stroke is associated with an increased risk of dementia. To assist in the early identification of individuals at high risk of future dementia, numerous prediction models have been developed for use in the general population. However, it is not known whether such models also provide accurate predictions among stroke patients. Therefore, the aim of this study was to determine whether existing dementia risk prediction models that were developed for use in the general population can also be applied to individuals with a history of stroke to predict poststroke dementia with equivalent predictive validity. Methods: Data were harmonized from 4 stroke studies (follow-up range, ≈12–18 months poststroke) from Hong Kong, the United States, the Netherlands, and France. Regression analysis was used to test 3 risk prediction models: the Cardiovascular Risk Factors, Aging and Dementia score, the Australian National University Alzheimer Disease Risk Index, and the Brief Dementia Screening Indicator. Model performance or discrimination accuracy was assessed using the C statistic or area under the curve. Calibration was tested using the Grønnesby and Borgan and the goodness-of-fit tests. Results: The predictive accuracy of the models varied but was generally low compared with the original development cohorts, with the Australian National University Alzheimer Disease Risk Index (C-statistic, 0.66) and the Brief Dementia Screening Indicator (C-statistic, 0.61) both performing better than the Cardiovascular Risk Factors, Aging and Dementia score (area under the curve, 0.53). Conclusions: Dementia risk prediction models developed for the general population do not perform well in individuals with stroke. Their poor performance could have been due to the need for additional or different predictors related to stroke and vascular risk factors or methodological differences across studies (eg, length of follow-up, age distribution). Future work is needed to develop simple and cost-effective risk prediction models specific to poststroke dementia.


2015 ◽  
Vol 2015 ◽  
pp. 1-31 ◽  
Author(s):  
Wenda He ◽  
Arne Juette ◽  
Erika R. E. Denton ◽  
Arnau Oliver ◽  
Robert Martí ◽  
...  

Breast cancer is the most frequently diagnosed cancer in women. However, the exact cause(s) of breast cancer still remains unknown. Early detection, precise identification of women at risk, and application of appropriate disease prevention measures are by far the most effective way to tackle breast cancer. There are more than 70 common genetic susceptibility factors included in the current non-image-based risk prediction models (e.g., the Gail and the Tyrer-Cuzick models). Image-based risk factors, such as mammographic densities and parenchymal patterns, have been established as biomarkers but have not been fully incorporated in the risk prediction models used for risk stratification in screening and/or measuring responsiveness to preventive approaches. Within computer aided mammography, automatic mammographic tissue segmentation methods have been developed for estimation of breast tissue composition to facilitate mammographic risk assessment. This paper presents a comprehensive review of automatic mammographic tissue segmentation methodologies developed over the past two decades and the evidence for risk assessment/density classification using segmentation. The aim of this review is to analyse how engineering advances have progressed and the impact automatic mammographic tissue segmentation has in a clinical environment, as well as to understand the current research gaps with respect to the incorporation of image-based risk factors in non-image-based risk prediction models.


2021 ◽  
Author(s):  
Gabriella Gatt ◽  
Nikolai Attard

Abstract BackgroundDespite increasing prevalence, age specific risk predictive models for erosive tooth wear in preschool age children have not been developed. Identification of at risk groups and the timely introduction of behavioural change or treatment will stop the progression of erosive wear in the permanent dentition. The aim of this study was to identify age specific risk factors for erosive wear. Distinct risk prediction models for three year old and five year old children were developed.MethodsA prospective cohort study included clinical examinations and parent administered questionnaires for three and five-year-old children. Chi-square tests explored categorical demographic variables, Spearman Rank Order correlation tests examined changes in BEWE scores with changes in food frequencies while Wilcoxon signed rank tests evaluated the temporal effect of frequencies of consumption of dietary items. Mann-Whitney U tests compared changes in BEWE scores over time for the twenty-six bivariate categorical variables and Kruskall-Wallis tests compared changes in BEWE scores over time across the remaining 55 categorical variables representing demographic factors, oral hygiene habits and dietary habits. Change in BEWE scores for continuous variables was investigated using Spearman Rho correlation coefficient Test. Those variables showing significance with a difference in BEWE cumulative score over time were utilised to develop two risk prediction models. The models were evaluated by Receiver Operating Characteristics (ROC) analysis.ResultsRisk factors for the three-year-old cohort included the erosive wear (χ2 (1, 92) = 12.829, p < 0.001), district (χ2 (5, 92) = 17.032, p = 0.004) and family size (χ2 (1, 92) = 4.547, p = 0.033). Risk factors for the five-year-old cohort also included erosive wear (χ2 (1, 144) = 4.768, p = 0.029) gender (χ2 (1, 144) = 19.399, p <0.001), consumption of iced tea (χ2 (1, 144) = 8.872, p = 0.003) and dry mouth (χ2 (1, 144) = 9.598, p = 0.002).Conclusions: Predictive risk factors for three-year-old children are based on demographic factors and are distinct from those for the five-year-old cohort, which are based on biological and behavioural factors. The presence of erosive wear is a risk factor for further wear in both age cohorts.


Sign in / Sign up

Export Citation Format

Share Document