The Diagnostic Accuracy of Bleeding Assessment Tools for the Identification of Patients with Mild Bleeding Disorders: A Systematic Review

Blood ◽  
2015 ◽  
Vol 126 (23) ◽  
pp. 4473-4473
Author(s):  
Floor CJI Moenen ◽  
Yvonne MC Henskens ◽  
Saskia AM Schols ◽  
Patty J Nelemans ◽  
Harry C. Schouten ◽  
...  

Abstract Introduction A study of diagnostic test accuracy compares a single index test to a gold standard to determine status of disease. The observed accuracy of a test varies among patient subgroups and is sensitive to bias. To achieve reliable estimates of diagnostic accuracy, an appropriate study design in a clinically relevant population is warranted. Recently a review was published about the evolution of the bleeding assessment tool (BAT) in diagnosing patients with mild bleeding disorders (MBD) (Rydz et al. J Thromb Haemost 2012). Many validation studies have been done. However, a critical appraisal addressing the quality of these validation studies is lacking. Objective We performed a systematic review to determine the quality and applicability of studies assessing the diagnostic utility of the BAT for MBD among clinic based cohorts. Methods The literature search was conducted using the electronic database PubMed. The final search date was March 2, 2015. The search terms: 'bleeding disorder OR bleeding tendency' AND 'bleeding questionnaire' were used. All studies assessing the diagnostic accuracy of bleeding questionnaires in identifying adults (age > 18 years) with MBD were considered eligible, irrespective of study design or used reference standard. The methodological quality and applicability of each included study was assessed using a Quality Assessment of Diagnostic studies-2 (QUADAS-2) tool. This tool consists of four domains specific for patient selection, index test, reference standard and participant flow. For each domain bias was assessed using signaling questions, for the first three domains applicability was assessed. Results The search yielded 530 citations, from which 35 possible relevant full-text studies were identified. Twenty-two studies were excluded, reasons for exclusion were: letter to the editor, validation of questionnaire combined with laboratory results and primary care population. Table 1 shows the 13 included studies, the assessed BAT and the targeted bleeding condition. Risk of bias and applicability concerns are summarized in figure 1. In 77% of the studies there was a high risk of bias for patient selection and applicability concerns. Many studies used a case control design, comparing patients with a known bleeding disorder with healthy controls. This leads to spectrum bias and might generate higher estimates of sensitivity and specificity (Rutjes et al. Clin Chem 2005). In 46% there was a high risk of bias for index test due to the use of a self-administered questionnaire or because the person conducting the questionnaire was aware of the diagnosis. This leads to observer bias caused by better awareness and over-reporting of bleeding symptoms. Finally, there was high risk of bias in study flow in 38% of the studies. These studies included symptoms after diagnosis of the bleeding disorder. Since bleeding disorders are managed by interventions to prevent bleeding, underestimation of the bleeding symptoms may occur. Conclusion This review highlights the difficulties and advantages of the BAT validation studies. It provides the ability for medical practitioners to apply the BAT with full awareness of its restrictions and benefits. With the evaluation of the risks of bias in the included studies we highlighted limitations, especially in method of patient selection and use of index test, that future studies preferably should try to avoid. Disclosures No relevant conflicts of interest to declare.

BMJ Open ◽  
2019 ◽  
Vol 9 (4) ◽  
pp. e025790 ◽  
Author(s):  
Rebecca Beynon ◽  
Martha Maria Christine Elwenspoek ◽  
Athena Sheppard ◽  
John Nicholas Higgins ◽  
Angelos G Kolias ◽  
...  

ObjectiveLumbar radiculopathy (LR) often manifests as pain in the lower back radiating into one leg (sciatica). Unsuccessful back surgery is associated with significant healthcare costs and risks to patients. This review aims to examine the diagnostic accuracy of selective nerve root blocks (SNRBs) to identify patients most likely to benefit from lumbar decompression surgery.DesignSystematic review of diagnostic test accuracy studies.Eligibility criteriaPrimary research articles using a patient population with low back pain and symptoms in the leg, SNRB administered under radiological guidance as index test, and any reported reference standard for the diagnosis of LR.Information sourcesMEDLINE (Ovid), MEDLINE In-Process & Other Non-Indexed Citations, EMBASE, Science Citation Index, Biosis, LILACS, Dissertation abstracts and National Technical Information Service from inception to 2018.MethodsRisk of bias and applicability was assessed using the QUADAS-2 tool. We performed random-effects logistic regression to meta-analyse studies grouped by reference standard.Results6 studies (341 patients) were included in this review. All studies were judged at high risk of bias. There was substantial heterogeneity across studies in sensitivity (range 57%–100%) and specificity (10%–86%) estimates. Four studies were diagnostic cohort studies that used either intraoperative findings during surgery (pooled sensitivity: 93.5% [95% CI 84.0 to 97.6]; specificity: 50.0% [16.8 to 83.2]) or ‘outcome following surgery’ as the reference standard (pooled sensitivity: 90.9% [83.1 to 95.3]; specificity 22.0% [7.4 to 49.9]). Two studies had a within-patient case-control study design, but results were not pooled because different types of control injections were used.ConclusionsWe found limited evidence which was of low methodological quality indicating that the diagnostic accuracy of SNRB is uncertain and that specificity in particular may be low. SNRB is a safe test with a low risk of clinically significant complications, but it remains unclear whether the additional diagnostic information it provides justifies the cost of the test.


BMJ Open ◽  
2018 ◽  
Vol 8 (2) ◽  
pp. e018132 ◽  
Author(s):  
Carmen Phang Romero Casas ◽  
Marrissa Martyn-St James ◽  
Jean Hamilton ◽  
Daniel S Marinho ◽  
Rodolfo Castro ◽  
...  

ObjectivesTo undertake a systematic review and meta-analysis to evaluate the test performance including sensitivity and specificity of rapid immunochromatographic syphilis (ICS) point-of-care (POC) tests at antenatal clinics compared with reference standard tests (non-treponemal (TP) and TP tests) for active syphilis in pregnant women.MethodsFive electronic databases were searched (PubMed, EMBASE, CRD, Cochrane Library and LILACS) to March 2016 for diagnostic accuracy studies of ICS test and standard reference tests for syphilis in pregnant women. Methodological quality was assessed using QUADAS-2 (Quality Assessment of Diagnostic Accuracy Studies). A bivariate meta-analysis was undertaken to generate pooled estimates of diagnostic parameters. Results were presented using a coupled forest plot of sensitivity and specificity and a scatter plot.ResultsThe methodological quality of the five included studies with regards to risk of bias and applicability concern judgements was either low or unclear. One study was judged as high risk of bias for patient selection due to exclusion of pregnant women with a previous history of syphilis, and one study was judged at high risk of bias for study flow and timing as not all patients were included in the analysis. Five studies contributed to the meta-analysis, providing a pooled sensitivity and specificity for ICS of 0.85 (95% CrI: 0.73 to 0.92) and 0.98 (95% CrI: 0.95 to 0.99), respectively.ConclusionsThis review and meta-analysis observed that rapid ICS POC tests have a high sensitivity and specificity when performed in pregnant women at antenatal clinics. However, the methodological quality of the existing evidence base should be taken into consideration when interpreting these results.PROSPERO registration numberCRD42016036335.


2021 ◽  
pp. 1-8
Author(s):  
Akihiro Shiroshita ◽  
Yasuhiro Oda ◽  
Seiji Takenouchi ◽  
Noboru Hagino ◽  
Yuki Kataoka

<b><i>Background:</i></b> The sensitivity and specificity of anti-glomerular basement membrane (GBM) antibodies have not been systematically analyzed. In this systematic review, we aimed to evaluate the diagnostic accuracy of anti-GBM antibodies for anti-GBM disease. <b><i>Summary:</i></b> Potential studies were searched using MEDLINE, Embase, the Cochrane Library, and the International Clinical Trials Registry Platform based on the index test and target condition. The inclusion criteria were prospective or retrospective cohort studies or case-control studies assessing the sensitivity and specificity of anti-GBM antibodies, and the reference standard was clinical diagnosis including biopsy results. The exclusion criteria were review articles, case reports, animal studies, and in vitro studies. Quality assessment was conducted based on the Quality Assessment of Diagnostic Accuracy Studies-2. The pooled estimates of sensitivity and specificity were calculated using a bivariate random-effects model. The overall quality was evaluated using the Grades of Recommendation, Assessment, Development, and Evaluation. Six studies (1,691 patients) and 11 index tests were included in our systematic review. A high risk of bias and concerns regarding the applicability of patient selection were noted because of the case-control design in 67% of the included studies. The pooled sensitivity and specificity were 93% (95% CI: 84–97%) and 97% (95% CI: 94–99%), respectively. The certainty of evidence was low because of the high risk of bias and indirectness. <b><i>Key Messages:</i></b> Anti-GBM antibodies may exhibit high sensitivity and specificity in the diagnosis of anti-GBM disease. Further cohort studies are needed to confirm their precise diagnostic accuracy and compare diagnostic accuracies among different immunoassays.


BMJ Open ◽  
2019 ◽  
Vol 9 (8) ◽  
pp. e031558 ◽  
Author(s):  
Neda Zafari ◽  
Leonid Churilov ◽  
Richard J MacIsaac ◽  
Niloufar Torkamani ◽  
Helen Baxter ◽  
...  

IntroductionTimely detection leading to the implementation of reno-protective measures reduces the progression of diabetic kidney disease. Estimated glomerular filtration rate (eGFR) is a major surrogate of kidney function. The Chronic Kidney Disease Epidemiology Collaboration (CKD-EPI) Equation is a tool to estimate GFR. This protocol outlines a systematic-review, assessing the diagnostic accuracy of the CKD-EPI equation in adults with diabetes.Methods and analysisMEDLINE, Embase, Cochrane Central Register of Controlled Trials and grey literature will be searched for publications in English, Farsi, Dutch and Chinese from 2009 (when CKD-EPI was first introduced) to January 2019. Bridging searches will be conducted to capture literature published from January 2019 until final review publication. The inclusion criteria will be (1) study participants with diabetes; (2) age ≥18 years; (3) creatinine-based CKD-EPI eGFR as index test; (4) measured GFR using the clearance/plasma disappearance of inulin, iohexol, iothalamate, diethylenetriamine-pentaacetic acid (DTPA) or chromium labelled ethylenediaminetetraacetic acid (Cr-EDTA) as reference test; (5) report of the diagnostic accuracy of the index test. Exclusion criteria will be participants with renal transplant, chronic use of corticosteroids, chronic inflammatory diseases, pregnancy, non-diabetes related kidney disease, thalassaemia, heart failure, pregnancy and potential kidney donors as well as critically ill patients. Screening, eligibility check, risk of bias assessment and data extraction will be carried out by two independent reviewers. Any discrepancies will be discussed, and third-party opinion will be sought. The risk of bias will be assessed using the Quality Assessment of Diagnostic Accuracy Studies−2 tool. A quantitative synthesis of the aggregated-data will be used if the included studies are homogenous.Ethics and disseminationNo ethics approval is required. The outcome will be published in a peer-reviewed journal. The results will help researchers and clinicians evaluate the diagnostic accuracy of the creatinine-based CKD-EPI eGFR in adults with diabetes.PROSPERO registration numberCRD42018108776


2021 ◽  
Vol 21 (1) ◽  
Author(s):  
Chinyereugo M. Umemneku Chikere ◽  
Kevin J. Wilson ◽  
A. Joy Allen ◽  
Luke Vale

Abstract Background Staquet et al. and Brenner both developed correction methods to estimate the sensitivity and specificity of a binary-response index test when the reference standard is imperfect and its sensitivity and specificity are known. However, to our knowledge, no study has compared the statistical properties of these methods, despite their long application in diagnostic accuracy studies. Aim To compare the correction methods developed by Staquet et al. and Brenner. Methods Simulations techniques were employed to compare the methods under assumptions that the new test and the reference standard are conditionally independent or dependent given the true disease status of an individual. Three clinical datasets were analysed to understand the impact of using each method to inform clinical decision-making. Results Under the assumption of conditional independence, the Staquet et al. correction method outperforms the Brenner correction method irrespective of the prevalence of disease and whether the performance of the reference standard is better or worse than the index test. However, when the prevalence of the disease is high (> 0.9) or low (< 0.1), the Staquet et al. correction method can produce illogical results (i.e. results outside [0,1]). Under the assumption of conditional dependence; both methods failed to estimate the sensitivity and specificity of the index test especially when the covariance terms between the index test and the reference standard is not close to zero. Conclusion When the new test and the imperfect reference standard are conditionally independent, and the sensitivity and specificity of the imperfect reference standard are known, the Staquet et al. correction method outperforms the Brenner method. However, where the prevalence of the target condition is very high or low or the two tests are conditionally dependent, other statistical methods such as latent class approaches should be considered.


Haemophilia ◽  
2018 ◽  
Vol 24 (4) ◽  
pp. 525-535 ◽  
Author(s):  
F. C. J. I. Moenen ◽  
P. J. Nelemans ◽  
S. E. M. Schols ◽  
H. C. Schouten ◽  
Y. M. C. Henskens ◽  
...  

2020 ◽  
Vol 27 (8) ◽  
Author(s):  
Nishant Aggarwal ◽  
Mohil Garg ◽  
Vignesh Dwarakanathan ◽  
Nitesh Gautam ◽  
Swasthi S Kumar ◽  
...  

Abstract Infrared thermal screening, via the use of handheld non-contact infrared thermometers (NCITs) and thermal scanners, has been widely implemented all over the world. We performed a systematic review and meta-analysis to investigate its diagnostic accuracy for the detection of fever. We searched PubMed, Embase, the Cochrane Library, medRxiv, bioRxiv, ClinicalTrials.gov, COVID-19 Open Research Dataset, COVID-19 research database, Epistemonikos, EPPI-Centre, World Health Organization International Clinical Trials Registry Platform, Scopus and Web of Science databases for studies where a non-contact infrared device was used to detect fever against a reference standard of conventional thermometers. Forest plots and Hierarchical Summary Receiver Operating Characteristics curves were used to describe the pooled summary estimates of sensitivity, specificity and diagnostic odds ratio. From a total of 1063 results, 30 studies were included in the qualitative synthesis, of which 19 were included in the meta-analysis. The pooled sensitivity and specificity were 0.808 (95%CI 0.656–0.903) and 0.920 (95%CI 0.769–0.975), respectively, for the NCITs (using forehead as the site of measurement), and 0.818 (95%CI 0.758–0.866) and 0.923 (95%CI 0.823–0.969), respectively, for thermal scanners. The sensitivity of NCITs increased on use of rectal temperature as the reference. The sensitivity of thermal scanners decreased in a disease outbreak/pandemic setting. Changes approaching statistical significance were also observed on the exclusion of neonates from the analysis. Thermal screening had a low positive predictive value, especially at the initial stage of an outbreak, whereas the negative predictive value (NPV) continued to be high even at later stages. Thermal screening has reasonable diagnostic accuracy in the detection of fever, although it may vary with changes in subject characteristics, setting, index test and the reference standard used. Thermal screening has a good NPV even during a pandemic. The policymakers must take into consideration the factors surrounding the screening strategy while forming ad-hoc guidelines.


2016 ◽  
Vol 27 (02) ◽  
pp. 141-156 ◽  
Author(s):  
Andrew J. Vermiglio

Background: In the field of audiology a test protocol (index test) is used to determine the presence or absence of a target condition. The value of an index test rests in its diagnostic accuracy. Results from an index test must be verified through the use of a reference standard. The clinician and researcher should determine the quality and applicability of diagnostic accuracy studies. The Standards for Reporting of Diagnostic Accuracy (STARD) statement was published in response to the low quality of research conducted across many fields of study. It provides guidelines for the development and interpretation of diagnostic accuracy studies. Purpose: The primary purpose of this article is to assess the degree to which the main principles of diagnostic accuracy studies are used for the detection of central auditory nervous system lesions and a central auditory processing disorder (CAPD). The secondary purpose is to compare the design of these studies to select key guidelines from the STARD statement. The third purpose of this article is to present an argument against the reassignment of diagnostic accuracy values of a particular index test for one target condition (e.g., a central site of lesion) to a different target condition (e.g., a CAPD). Results: A review of diagnostic accuracy literature on the detection of a central site of lesion reveals the use of a reference standard for the independent verification of the index test results. However, diagnostic accuracy studies involving index tests for the detection of a CAPD show that independent verification of index test results is nonexistent or at best questionable. Conclusion: For a particular index test, while the diagnostic accuracy for detection of a central site of lesion may have been determined appropriately, it is inappropriate to reassign these diagnostic accuracy values to a different diagnostic target such as a CAPD.


2016 ◽  
Vol 23 (7) ◽  
pp. 629-638 ◽  
Author(s):  
Irene J Tan ◽  
Lucy P Dobson ◽  
Stephen Bartnik ◽  
Josephine Muir ◽  
Angus W Turner

Introduction Advances in imaging capabilities and the evolution of real-time teleophthalmology have the potential to provide increased coverage to areas with limited ophthalmology services. However, there is limited research assessing the diagnostic accuracy of face-to-face teleophthalmology consultation. This systematic review aims to determine if real-time teleophthalmology provides comparable accuracy to face-to-face consultation for the diagnosis of common eye health conditions. Methods A search of PubMed, Embase, Medline and Cochrane databases and manual citation review was conducted on 6 February and 7 April 2016. Included studies involved real-time telemedicine in the field of ophthalmology or optometry, and assessed diagnostic accuracy against gold-standard face-to-face consultation. The revised quality assessment of diagnostic accuracy studies (QUADAS-2) tool assessed risk of bias. Results Twelve studies were included, with participants ranging from four to 89 years old. A broad number of conditions were assessed and include corneal and retinal pathologies, strabismus, oculoplastics and post-operative review. Quality assessment identified a high or unclear risk of bias in patient selection (75%) due to an undisclosed recruitment processes. The index test showed high risk of bias in the included studies, due to the varied interpretation and conduct of real-time teleophthalmology methods. Reference standard risk was overall low (75%), as was the risk due to flow and timing (75%). Conclusion In terms of diagnostic accuracy, real-time teleophthalmology was considered superior to face-to-face consultation in one study and comparable in six studies. Store-and-forward image transmission coupled with real-time videoconferencing is a suitable alternative to overcome poor internet transmission speeds.


2018 ◽  
Vol 29 (03) ◽  
pp. 206-222 ◽  
Author(s):  
Andrew J. Vermiglio ◽  
Sigfrid D. Soli ◽  
Xiangming Fang

AbstractThe primary components of a diagnostic accuracy study are an index test, the target condition (or disorder), and a reference standard. According to the Standards for Reporting Diagnostic Accuracy statement, the reference standard should be the best method available to independently determine if the results of an index test are correct. Pure-tone thresholds have been used as the “gold standard” for the validation of some tests used in audiology. Many studies, however, have shown a lack of agreement between the audiogram and the patient’s perception of hearing ability. For example, patients with normal audiograms may report difficulty understanding speech in the presence of background noise.The primary purpose of this article is to present an argument for the use of self-report as a reference standard for diagnostic studies in the field of audiology. This will be in the form of a literature review on pure-tone threshold measures and self-report as reference standards. The secondary purpose is to determine the diagnostic accuracy of pure-tone threshold and Hearing-in-Noise Test (HINT) measures for the detection of a speech-recognition-in-noise disorder.Two groups of participants with normal pure-tone thresholds were evaluated. The King–Kopetzky syndrome (KKS) group was made up of participants with the self-report of speech-recognition-in-noise difficulties. The control group was made up of participants with no reports of speech-recognition-in-noise problems. The reference standard was self-report. Diagnostic accuracy of HINT and pure-tone threshold measures was determined by measuring group differences, sensitivity and specificity, and the area under the curve (AUC) for receiver-operating characteristic (ROC) curves.Forty-seven participants were tested. All participants were native speakers of American English. Twenty-two participants were in the control group and 25 in the KKS group. The groups were matched for age.Pure-tone threshold data were collected using the Hughson–Westlake procedure. Speech-recognition-in-noise data was collected using a software system and the standard HINT protocol. Statistical analyses were conducted using descriptive, correlational, two-sample t tests, and logistic regression.The literature review revealed that self-report has been used as a reference standard in investigations of patients with normal audiograms and the perception of difficulty understanding speech in the presence of background noise. Self-report may be a better indicator of hearing ability than pure-tone thresholds in some situations. The diagnostic accuracy investigation revealed statistically significant differences between control and KKS groups for HINT performance (p < 0.01), but not for pure-tone threshold measures. Better sensitivity was found for the HINT Composite score (88%) than pure-tone average (PTA; 28%). The specificities for the HINT Composite score and PTA were 77% and 95%, respectively. ROC curves revealed a greater AUC for the HINT Composite score (AUC = 0.87) than for PTA (AUC = 0.51).Self-report is a reasonable reference standard for studies on the diagnostic accuracy of speech-recognition-in-noise tests. For individuals with normal pure-tone thresholds, the HINT demonstrated a higher degree of diagnostic accuracy than pure-tone thresholds for the detection of speech-recognition-in-noise disorder.


Sign in / Sign up

Export Citation Format

Share Document