scholarly journals Retrospective Real-World Analysis of the Primary Safety Outcomes in Venous Thromboembolism of High-Risk Major Bleeding Cancer Patients Receiving Therapeutic Anticoagulation

Blood ◽  
2020 ◽  
Vol 136 (Supplement 1) ◽  
pp. 15-16
Author(s):  
Eric Brucks Brucks ◽  
Sumana Veeravelli ◽  
Ritika Halder ◽  
Francisco Javier Martinez ◽  
Jorge Andres Leiva ◽  
...  

Introduction The therapeutic anticoagulation of cancer-associated venous thromboembolism (VTE) is challenging because of the increased risk for thromboembolic recurrence and major bleeds (MB). Direct oral anticoagulants (DOACs) have emerged over Low Molecular Weight Heparin (LMWH) as a preferred treatment option by improving efficacy without compromising safety. Despite these advantages, major bleeding was noted to be a safety concern in recent randomized clinical trials, mainly for gastrointestinal (GI) and genitourinary (GU) high-risk cancer patients. We evaluated our cancer center's institutional experience to determine the safety signals among VTE treatments and clinical risk factors which can predict an adverse outcome to further develop our clinical practice decisions in oncology patients. Methods This is a retrospective chart review of patients receiving DOACs and LMWH with histologically proven GI or GU malignancy and symptomatic or incidental VTE treated at The University of Arizona Cancer Center from November 2013 to February 2020. Patients were excluded if DOACs were prescribed for any other reason not related to VTE, or the thrombotic event was determined to be unrelated to active malignancy. The primary safety outcome was defined as documented major bleeding by hemoglobin reduction of ≥2 g/dL, transfusion of ≥2 units of Packed Red Blood Cells, bleeding in a critical site, or bleeding contributing to death. The secondary efficacy outcome was defined as documented recurrent deep vein thrombosis (DVT), nonfatal pulmonary embolism (PE), or fatal PE. Data was collected as individually and composite outcome (MB+rVTE) before six months. Continuous non-normally distributed data are compared using the non-parametric Kruskal-Wallis Rank Sum. Categorical data are compared using χ2 or Fisher's exact test. Logistic regression was used to assess the relationship between anticoagulant therapy and the 3 outcome measurements. The relationship between therapies was adjusted for clinical risk factors using logistic regression. Statistical analyses were performed using Stata16. Results Our review included 160 patients with similar baseline characteristics who were prescribed enoxaparin (n=53), apixaban (n=55), and rivaroxaban (n=52) except for white race and active smokers [Table 1]. Primary MB outcome events at six months were 7.5%, 10.9%, and 17.3% for enoxaparin, apixaban, and rivaroxaban, with no statistical difference among therapies or cancer type [p>0.05]. Secondary rVTE outcome events at six months were 7.5%, 3.6%, and 5.7% for enoxaparin, apixaban, and rivaroxaban, respectively, with no significant difference among therapies or cancer type [p>0.05]. There were no differences in individual or composite outcomes of LMWH, apixaban, rivaroxaban, or DOAC by GI or GU cancer type. There are no statistically significant clinical associations between anticoagulation therapy by the three outcome measurements in either GI or GU cancer. Beyond six months on enoxaparin, apixaban, and rivaroxaban, there were 5.6%, 3.6%, and 5.7% added events to MB, and 3.7%, 1.8%, and 0% to rVTE, respectively. Conclusions Our retrospective review has shown similar efficacy in preventing recurrent VTE among anticoagulants and indirectly equivalent to those published by the randomized clinical trials. Notably, the rate of MB for our real-world high-risk GI/GU cancer was 2.8 times higher in patients receiving DOACs compared to those on clinical trial [MB 6% SELECT-D Young et al. 2018 and MB 3.8% Caravaggio Agnelli et al. 2020] and in agreement with a recent meta-analysis [relative risk GI 1.9 and GU 4.9, Giustozzi et al. 2020]. We recognized there are inherited selection bias in both clinical trials and retrospective studies, although baseline characteristics are similar. Future clinical trial designs should consider including real-world high-risk patients according to associated clinical risk factors to aid in the appropriate choice of therapeutic anticoagulation. Disclosures No relevant conflicts of interest to declare.

2010 ◽  
Vol 30 (1) ◽  
pp. 58-73 ◽  
Author(s):  
Paul T. Seed ◽  
Lucy C. Chappell ◽  
Michael A. Black ◽  
Katrina K. Poppe ◽  
Yuan-Chun Hwang ◽  
...  

2020 ◽  
Vol 41 (21) ◽  
pp. 1988-1999 ◽  
Author(s):  
Neal A Chatterjee ◽  
Jani T Tikkanen ◽  
Gopi K Panicker ◽  
Dhiraj Narula ◽  
Daniel C Lee ◽  
...  

Abstract Aims To determine whether the combination of standard electrocardiographic (ECG) markers reflecting domains of arrhythmic risk improves sudden and/or arrhythmic death (SAD) risk stratification in patients with coronary heart disease (CHD). Methods and results The association between ECG markers and SAD was examined in a derivation cohort (PREDETERMINE; N = 5462) with adjustment for clinical risk factors, left ventricular ejection fraction (LVEF), and competing risk. Competing outcome models assessed the differential association of ECG markers with SAD and competing mortality. The predictive value of a derived ECG score was then validated (ARTEMIS; N = 1900). In the derivation cohort, the 5-year cumulative incidence of SAD was 1.5% [95% confidence interval (CI) 1.1–1.9] and 6.2% (95% CI 4.5–8.3) in those with a low- and high-risk ECG score, respectively (P for Δ < 0.001). A high-risk ECG score was more strongly associated with SAD than non-SAD mortality (adjusted hazard ratios = 2.87 vs. 1.38 respectively; P for Δ = 0.003) and the proportion of deaths due to SAD was greater in the high vs. low risk groups (24.9% vs. 16.5%, P for Δ = 0.03). Similar findings were observed in the validation cohort. The addition of ECG markers to a clinical risk factor model inclusive of LVEF improved indices of discrimination and reclassification in both derivation and validation cohorts, including correct reclassification of 28% of patients in the validation cohort [net reclassification improvement 28 (7–49%), P = 0.009]. Conclusion For patients with CHD, an externally validated ECG score enriched for both absolute and proportional SAD risk and significantly improved risk stratification compared to standard clinical risk factors including LVEF. Clinical Trial Registration https://clinicaltrials.gov/ct2/show/NCT01114269. ClinicalTrials.gov ID NCT01114269.


Blood ◽  
2016 ◽  
Vol 128 (7) ◽  
pp. 911-922 ◽  
Author(s):  
Julie A. E. Irving ◽  
Amir Enshaei ◽  
Catriona A. Parker ◽  
Rosemary Sutton ◽  
Roland P. Kuiper ◽  
...  

Key Points Chromosomal abnormalities predict outcome after relapse in BCP-ALL, and high-risk cytogenetics takes precedence over clinical risk factors. Patients with mutations or deletions targeting TP53, NR3C1, BTG1, and NRAS were associated with clinical high risk and an inferior outcome.


2021 ◽  
Author(s):  
Jianfei Chen ◽  
Zhong-liang Liu ◽  
Song Su ◽  
Jun Fan ◽  
Shun-de Tan ◽  
...  

Abstract Background: Hepatocellular carcinoma (HCC) is one of the series malignant aggressive disease which shows elusive biological behavior and terrible prognosis. It is inadequate for the single evaluation system such as tumor-node-metastasis (TNM) staging system to predict the overall survival (OS) in HCC patients. Here, we conducted this study to identify prognosis-related genes, so as to form a reliable prognostic assessment model.Results: From three datasets in the GEO database, we identified two hub genes cytoskeleton associated protein 2 (CKAP2) and forkhead box M1 (FOXM1) among 348 differentially expressed genes (DGEs) that were differentially expressed between HCC and normal samples. The function analysis of those DEGs are enriched in cell division process (e.g., mitotic nuclear division, nuclear chromosome segregation) and metabolic process (e.g., organic acid catabolic process). Then, we established a two-gene model, that extremely distinguished the population at risk of liver cancer to high-risk and low-risk and was even viable in the TNM stage i-ii and iii-iv, vascular invasion and non-invasion subgroups (all P<0.05). Next, a nomogram was set out combined with the two hub genes and clinical risk factors, and the predictive power of the nomogram performed more outstanding than the gene expression or clinical parameters alone.Conclusions: Our two-gene-based evaluation system effectively filtered out the high-risk HCC patients, and could potentially be used for clinical decision-making and individualized management of particular HCC patients.


2020 ◽  
Vol 4 (22) ◽  
pp. 5762-5771
Author(s):  
Ho-Young Yhim ◽  
Yael Eshet ◽  
Ur Metser ◽  
Chae-Hong Lim ◽  
Katherine Lajkosz ◽  
...  

Abstract There are limited data regarding the combined value of the pretransplant Deauville score (DS) from a positron emission tomography scan and clinical risk factors in patients with relapsed/refractory aggressive non-Hodgkin lymphoma (NHL). We performed a retrospective analysis to assess the prognostic role of pretransplant DS in patients with relapsed/refractory aggressive NHL who underwent salvage chemotherapy and autologous stem cell transplantation (ASCT). We identified 174 eligible patients between January 2013 and March 2019. In multivariable analysis, pretransplant DS, B symptoms, and secondary International Prognostic Index (sIPI) were independent risk factors for event-free survival (EFS). These variables were used to derive an integrated risk score that categorized 166 patients with available information for all risk factors into 3 groups: low (n = 92; 55.4%), intermediate (n = 48; 28.9%), and high (n = 26; 15.7%). The new prognostic index showed a strong association with EFS (low-risk vs intermediate-risk hazard ratio [HR], 3.94; 95% confidence interval [CI], 2.16-7.17; P &lt; .001; low-risk vs high-risk HR, 10.83; 95% CI, 5.81-20.19; P &lt; .001) and outperformed models based on clinical risk factors or DS alone. These results were validated in 60 patients from an independent external cohort (low-risk vs intermediate-risk HR, 4.04; 95% CI, 1.51-10.82; P = .005; low-risk vs high-risk HR, 10.49; 95% CI, 4.11-26.73; P &lt; .001). We propose and validate a new prognostic index that risk-stratifies patients undergoing salvage chemotherapy followed by ASCT, thereby identifying patients at high risk for posttransplant treatment failure.


2019 ◽  
Vol 40 (Supplement_1) ◽  
Author(s):  
S E Lee ◽  
G Pontone ◽  
I Gottlieb ◽  
M Hadamitzky ◽  
J A Leipsic ◽  
...  

Abstract Background It is still debatable whether the so-called high-risk plaque (HRP) simply represents a certain phase during the natural history of coronary atherosclerotic plaques or the disease progression would differ according to the presence of HRP. Purpose We determined whether the pattern of non-obstructive lesion progression into obstructive lesions would differ according to the presence of HRP. Methods Patients with non-obstructive coronary artery disease, defined as % diameter stenosis (%DS) ≥50%, were enrolled from a prospective, multinational registry of consecutive patients who underwent serial coronary computed tomography angiography at an inter-scan interval of ≥2 years. HRP was defined as lesions with ≥2 of positive remodelling, spotty calcification, and low-attenuation plaque. The total and compositional percent atheroma volume (PAV) at baseline and annualized PAV change were compared between non-HRP and HRP lesions. Results A total of 1,115 non-obstructive lesions were identified from 327 patients (61.1±8.9 years old, 66.0% male). There were 690 non-HRP and 425 HRP lesions. HRP lesions possessed greater PAV and %DS at baseline compared to non-HRP lesions. However, the annualized total and non-calcified PAV change were greater in non-HRP lesions than in HRP lesions. On multivariate analysis, addition of baseline PAV and %DS to clinical risk factors improved the predictive power of the model (Table). When clinical risk factors, PAV, %DS, and HRP were all adjusted on Model 3, only baseline PAV and %DS independently predicted the development of obstructive lesions (hazard ratio (HR) 1.046 [95% confidence interval (CI): 1.026–1.066] and HR 1.087 [95% CI: 1.055–1.119], respectively, all p<0.001), while HRP did not (p>0.05). Comparison of C-statistics of per-lesion analysis to predict progression to obstructive lesion C-statistics (95% CI) P Model 1: Baseline PAV 0.880 (0.879–0.884) – Model 2: Model 1 + baseline %DS 0.938 (0.937–0.939) vs. Model 1: <0.001 Model 3: Model 2 + HRP 0.935 (0.934–0.937) vs. Model 2: 0.004 Adjusted for age, male sex, hypertension, diabetes mellitus, hyperlipidemia, family history of coronary artery disease, smoking, body mass index, and statin use. Conclusion The pattern of individual coronary atherosclerotic plaque progression differed according to the presence of HRP. Baseline PAV was the most important predictor for lesions developing into obstructive lesions rather than the presence of HRP features at baseline. Acknowledgement/Funding This work was supported by the National Research Foundation of Korea funded by the Ministry of Science and ICT (Grant No. 2012027176).


Sign in / Sign up

Export Citation Format

Share Document