scholarly journals Left Anterior Fascicular Block After Transcatheter Closure of Ventricular Septal Defect in Children

2021 ◽  
Vol 8 ◽  
Author(s):  
Zhijun Wu ◽  
Penghui Yang ◽  
Ping Xiang ◽  
Xiaojuan Ji ◽  
Jie Tian ◽  
...  

Background: Arrhythmia is the most common complication after transcatheter closure of a ventricular septal defect (VSD). However, the effects of postprocedural left anterior fascicular block are not clear. This study presents the clinical characteristics, prognosis, and related risk factors of left anterior fascicular block after transcatheter closure of a VSD in children.Methods: The clinical and follow-up data of the patients in the Heart Center of Children's Hospital of Chongqing Medical University from June 2009 to October 2018 were reviewed. And 30 cases were eligible out of all 1,371 cases.Results: An electrocardiogram showed a left anterior fascicular block within 3 days, and most patients gradually returned to normal within 1–2 years, showing a dynamic change. Left ventricular end-diastolic dimension Z-score ranged from −2 to 2 in all children, and no decrease of left ventricular ejection fraction was found in all children. The high ratio between VSD size and body surface area [p < 0.05, odds ratio (OR) 2.6, 95% CI: 1.136–6.113] and large diameter difference between the occluder size and VSD size (p < 0.05, OR 2.1, 95% CI: 1.036–4.609) were independent risk factors for postprocedural left anterior fascicular block.Conclusions: The incidence of postprocedural left anterior fascicular block is not that low, and the overall prognosis is quite good at the current follow-up stage. No progressive severity has been found, such as complete left bundle branch block, double (triple) bundle branch block, and atrioventricular block, to have an influence on cardiac systolic and diastolic function.

2021 ◽  
Vol 10 (11) ◽  
pp. 2284
Author(s):  
Diana Gurzău ◽  
Alexandra Dădârlat-Pop ◽  
Bogdan Caloian ◽  
Gabriel Cismaru ◽  
Horaţiu Comşa ◽  
...  

Left bundle branch block is not a benign pathology, and its presence requires the identification of a pathological substrate, such as ischemic heart disease. Left bundle branch block appears to be more commonly associated with normal coronary arteries, especially in women. The objectives of our study were to describe the particularities of left bundle branch block in women compared to men with ischemic heart disease. Result: We included seventy patients with left bundle branch block and ischemic heart disease, with a mean age of 67.01 ± 8.89 years. There were no differences in the profile of risk factors, except for smoking and uric acid. The ventricular depolarization (QRS) duration was longer in men than women (136.86 ± 8.32 vs. 132.57 ± 9.19 msec; p = 0.018) and also men were observed to have larger left ventricular diameters. Left bundle branch block duration was directly associated with ventricular diameters and indirectly associated with left ventricular ejection fraction value, especially in women (R = −0.52, p = 0.0012 vs. R = −0.50, p = 0.002). In angiography, 80% of women had normal epicardial arteries compared with 65.7% of men; all these patients presented with microvascular dysfunction. Conclusion: The differences between the sexes were not so obvious in terms of the presence of risk factors; instead, there were differences in electrocardiographic, echocardiographic, and angiographic aspects. Left bundle branch block appears to be a marker of microvascular angina and systolic dysfunction, especially in women.


EP Europace ◽  
2020 ◽  
Vol 22 (8) ◽  
pp. 1234-1239
Author(s):  
Wei Ji ◽  
Xueying Chen ◽  
Jie Shen ◽  
Diqi Zhu ◽  
Yiwei Chen ◽  
...  

Abstract Aims As a physiological pacing strategy, left bundle branch pacing (LBBP) were used to correct left bundle branch block (LBBB), however, there is no relevant report in children. We aimed to evaluate the feasibility of LBBP in children. Methods and results Left bundle branch pacing was performed in a 10-year-old girl with a second-degree atrioventricular and LBBB. Under the guide of fluoroscopy, the pacing lead was deeply screwed into the interventricular septum to gain right bundle branch block (RBBB) pattern of paced QRS. Selective LBBP was achieved with a typical RBBB pattern of paced morphology and a discrete component between stimulus and ventricular activation in intracardiac electrogram and reached the standard of the stimulus to left ventricular activation time of 56 ms. At a 3-month follow-up, the LBBP acquired the reduction of left ventricular size and enhancement of left ventricular ejection fraction. Conclusion The application of LBBP in a child was first achieved with inspiring preliminary results. The LBBP can be carried out in children by cautiousness under the premise of strict grasp of indications.


2009 ◽  
Vol 27 (15_suppl) ◽  
pp. 9569-9569
Author(s):  
A. Salar ◽  
B. Sanchez-Gonzalez ◽  
A. Alvarez-Larran ◽  
J. Comin ◽  
J. R. Gonzalez ◽  
...  

9569 Background: NT-Pro BNP provides diagnostic and prognostic information in many heart syndromes, but its role in oncologic patients is not established. Methods: We studied the association between NT-ProBNP levels and the risk of major toxicity associated with chemotherapy and death from all causes in 116 consecutive patients with lymphoma treated with curative intent. High resolution ultrasound echocardiography and serum NT-ProBNP levels were prospectively done previous the start of chemotherapy. Charlson Comorbidity Index was retrospectively calculated. A major toxicity event was defined as: when chemotherapy had to be discontinued, when chemotherapy had to be changed to a less intensive regimen, and treatment-related death. Results: High blood levels of NT-proBNP were associated with previous cardiologic history but not with left ventricular ejection fraction. With a median follow-up of 16 moths (0–49 months), 25 patients had a major toxic event with first line chemotherapy and 16 had died at last follow-up. The threshold of NT-ProBNP with better predictive accuracy for major toxicity-free and overall survival was 900 pg/mL. Patients with levels of NT-proBNP greater of 900 pg/mL had an adjusted hazard ratio (HR) for major toxicity after first-line chemotherapy for lymphoma of 6.4 (95%CI, 2.7–15.1). Two additional independent factors predicting higher major toxicity associated with chemotherapy were albumin < 3.5 g/mL (HR 3.5, P=0.008) and number of extranodal sites ≥ 2 (HR 3.1, P<0.007). For prediction of death from all causes, patients with a NT-ProBNP greater of 900 pg/mL had an HR of death of 15.3 (95%CI, 4.8–48.8; P<0.001). ECOG ≥ 2 was also significant for predicting death (HR 3.6; 95%CI, 1.1–11.6; P=0.03). The NT-Pro BNP added prognostic information beyond that provided by conventional risk factors, including IPI, left ventricular ejection fraction and Charlson comorbidity index. Conclusions: NT-ProBNP is the stronger marker for predicting major toxicity after first-line chemotherapy and death from all causes in patients with lymphoma and provides prognostic information beyond that provided by conventional lymphoma risk factors and comorbidity indexes. No significant financial relationships to disclose.


2021 ◽  
Vol 11 (4) ◽  
pp. 264-270
Author(s):  
O. V. Arsenicheva

The aim. To study the risk factors for hospital mortality in patients with acute coronary syndrome with ST-segment elevation (STEACS) complicated by cardiogenic shock (CS).Materials and methods. A total of 104 patients with STEACS complicated by CS were studied. The follow-up group (group I) included 58 (55,8%) patients who died in hospital (mean age 71,8±7,31 years), the comparison group (group II) – 46 patients, who have been treated and discharged (mean age 59,5±6,18 years). All patients underwent general clinical studies, the level of troponins, lipids, glucose, creatinine in plasma was determined, electrocardiography and echocardiography were performed. Coronary angiography and percutaneous coronary intervention (PCI) were urgently performed. The method of binary logistic regression with the determination of the odds ratio and its 95% confidence interval for each reliable variable was used to identify risk factors for hospital mortality.Results. In group I patients with CS, compared with group II, patients over the age of 70 (32 (55,2%) vs 10 (22,7%), р=0,0004), with concomitant chronic kidney disease (32 (55,2%) vs 9 (19,6%), p=0,0002), postinfarction cardiosclerosis (30 (51,7%) vs 9 (19,6%), р=0,001) and chronic heart failure of III-IV functional class (32 (55,1%) vs 11 (23,9%), p=0,001) were significantly more often observed. Baseline levels of plasma leukocytes, troponin and creatinine were significantly higher in deceased patients with CS. Left ventricular ejection fraction below 40% was observed more often in the follow-up group than in the comparison group (46 (79,3%) vs 27 (58,7%), p=0,022). In group I, compared with group II, there was a higher incidence of three-vessel coronary lesions (36 (75%) vs 12 (26,1%), p=0,0001) and chronic coronary artery occlusion unrelated to STEACS (25 (52,1%) vs 12 (26,1%), р=0,009). The same trend was observed when assessing the average number of stenoses and occlusions of the coronary arteries. PCI was performed in 43 (74,1%) of the deceased and 43 (93,5%) of the surviving STEACS patients with CS (p=0,009). The follow-up group had a higher rate of unsuccessful PCI (30,2%) vs 3 (7%), р=0,001) and performed later than 6 hours after the onset of an angina attack (28 (65,1%) vs 6 (14%), р=0,0001).Summary. Hospital mortality in patients with STEMI complicated by CS was associated with the presence left ventricular ejection fraction less than 40%, three-vessel coronary lesion and performing PCI later than 6 hours from the beginning of the pain attack.


EP Europace ◽  
2019 ◽  
Vol 21 (7) ◽  
pp. 1070-1078 ◽  
Author(s):  
Wagner L Gali ◽  
Alvaro V Sarabanda ◽  
José M Baggio ◽  
Eduardo F Silva ◽  
Gustavo G Gomes ◽  
...  

Aims Data on long-term follow-up of patients with Chagas’ heart disease (ChHD) receiving a secondary prevention implantable cardioverter-defibrillator (ICD) are limited and its benefit is controversial. The aim of this study was to evaluate the long-term outcomes of ChHD patients who received a secondary prevention ICD. Methods and results We assessed the outcomes of consecutive ChHD patients referred to our Institution from 2006 to 2014 for a secondary prevention ICD [89 patients; 58 men; mean age 56 ± 11 years; left ventricular ejection fraction (LVEF), 42 ± 12%]. The primary outcome included a composite of death from any cause or heart transplantation. After a mean follow-up of 59 ± 27 months, the primary outcome occurred in 23 patients (5.3% per year). Multivariate analysis showed that LVEF < 35% [hazard ratio (HR) 4.64; P < 0.01] and age ≥ 65 years (HR 3.19; P < 0.01) were independent predictors of the primary outcome. Using these two risk factors, a risk score was developed, and lower- (no risk factors), intermediate- (one risk factor), and higher-risk (two risk factors) groups were recognized with an annual rate of primary outcome of 1.4%, 7.4%, and 20.4%, respectively. A high burden of appropriate ICD therapies (16% per year) and electrical storms were documented, however, ICD interventions did not impact on the primary outcome. Conclusion Among ChHD patients receiving a secondary prevention ICD, older age (≥65 years) and left ventricular dysfunction (LVEF < 35%) portend a poor outcome and were associated with increased risk of death or heart transplantation. Most patients received appropriate ICD therapies, however, ICD interventions did not impact on the primary outcome.


2019 ◽  
Vol 40 (Supplement_1) ◽  
Author(s):  
H Mizutani ◽  
T Kurita ◽  
S Kasuya ◽  
T Mori ◽  
H Ito ◽  
...  

Abstract Background Aortic valve stenosis (AS) is associated with the presence and severity of coronary artery disease independently of clinical risk factors, which leads to increased cardiovascular mortality. However, the prevalence of AS and its prognostic value among patients with acute myocardial infarction (AMI) remain unknown. Purpose The purpose of this study was to investigate the prevalence and prognostic impact of AS in AMI patients. Methods We studied 2,803 AMI patients using data from Mie ACS registry, a prospective and multicenter registry. Patients were divided into subgroups according to the presence and severity of AS based on maximal aortic flow rate by Doppler echocardiography before hospital discharge: non-AS <2.0 m/s, 2.0 m/s≤mild AS <3.0 m/s, 3.0 m/s≤moderate AS <4.0m/s and severe AS≥4.0 m/s. The primary outcome was defined as 2-year all-cause mortality. Results AS was detected in 79 patients (2.8%) including 49 mild AS, 23 moderate AS and 6 severe AS. AS patients were significantly older (79.9±9.8 versus 68.3±12.6 years), and higher killip classification than non-AS patients (P<0.01, respectively). However, left ventricular ejection fraction, and prevalence of primary PCI was similar between the 2 groups. During the follow-up periods (median 725 days), 333 (11.9%) patients experienced all-cause death. AS patients demonstrated the higher all-cause mortality rate compared to that of non-AS patients during follow up (47.3% versus 11.3%, P<0.0001, chi square). Kaplan-Meier curves showed that the probability of all-cause mortality was significantly higher among AS patients than non-AS patients, and was highest among moderate and severe AS (See figure A and B). Cox regression analyses for all-cause mortality demonstrated that the severity of AS was the strongest and independent poor prognostic factor (HR 1.71, 95% CI 1.30–2.24, P<0.001, See table). Cox hazard regression analysis Hazard ratio 95% Confidential interval P-value Severity of aortic valve stenosis 1.71 1.30–2.24 <0.001 Killip classification 1.63 1.46–1.82 <0.001 Age 1.07 1.06–1.09 <0.001 Serum creatinine level 1.05 1.03–1.08 <0.001 Max CPK level 1.00 1.00–1.01 <0.001 Left ventricular ejection fraction 0.96 0.95–0.97 <0.001 Primary percutaneous coronary intervention 0.67 0.47–0.96 0.03 CPK suggests creatinine phosphokinase. All cause mortality Conclusions The presence of AS of any severity contributes to worsening of patients' prognosis following AMI independently of other known risk factors. Acknowledgement/Funding None


Author(s):  
К.А. Ерусланова ◽  
А.В. Лузина ◽  
Ю.С. Онучина ◽  
В.С. Остапенко ◽  
Н.В. Шарашкина ◽  
...  

В последние годы появляется все больше работ, посвященных снижению воздействия классических факторов риска, негативно сказывающихся на выживаемости с возрастом. Целью исследования была оценка влияния сердечно-сосудистых заболеваний, их факторов риска и структурно-функциональных характеристик сердца на трехлетнюю выживаемость лиц 95 лет и старше. В исследовании участвовали 69 пациентов 95 лет и старше (98±1,9 года), из них 61 (88,4 %) женщина и 8 (11,6 %) мужчин. Через 3 года были получены данные о статусе жизни участников: 25 (36,2 %) были живы и 44 (63,8 %) умерли. По результатам проведенного однофакторного регрессионного анализа было определено, что факторы риска и анамнез сердечно-сосудистых заболеваний не ассоциированы с трехлетней выживаемостью. Однако в трехлетнем периоде риск смерти увеличивался в 3 раза при снижении ДАД <75 мм рт. ст., в 7,8 раза - при снижении ФВ ЛЖ <62 % и в 4,9 раза - при увеличении конечного диастолического размера правого желудочка >2,9 см. In recent years, more and more works have appeared that with age, classic risk factors that negatively affect the prognosis (cardiovascular diseases) lose their influence on life expectancy. The study aimed to assess the influence of cardiovascular diseases and their risk factors and structural and functional characteristics of the heart on three-year survival in people 95 years and older. The study involved 69 patients 95 years and older (98±1,9 years), 61 (88,4 %) were women. After 36 months, data were obtained on the participants’ status of life: 25 (36,2 %) were alive, and 44 (63,8 %) died. Based on the regression analysis results, it was determined that risk factors and history of cardiovascular diseases were not associated with 3-year survival. With a 3-year follow-up, the risk of death increases three times with a decrease in diastolic blood pressure less than 75 mm/Hg, 7,8 times with a decrease in left ventricular ejection fraction below 62 %, and 4,9 times with an increase in the end-diastolic size of the right ventricle by more than 2,9 cm.


2021 ◽  
Vol 39 (15_suppl) ◽  
pp. e24060-e24060
Author(s):  
Alma Farooque ◽  
Cibele Carroll ◽  
Amye Juliet Tevaarwerk ◽  
Priyanka Avinash Pophali

e24060 Background: Anthracyclines are known to cause long-term cardiotoxicity. There are no specific guidelines for CV screening and follow-up of AYA patients treated with anthracyclines. Pediatric guidelines focus on long-term imaging surveillance, while for adults, LVEF assessment prior to anthracyclines is recommended. Multiple studies have demonstrated LVEF assessment rarely impacts treatment decisions, especially in the absence of CV symptoms/risk factors, adds to unnecessary costs and delays treatment initiation. Our study aimed to determine the pre-treatment LVEF assessment practices in AYA lymphoma patients treated with anthracyclines and its association with long-term cardiotoxicity. Methods: AYA survivors diagnosed with lymphoma > 5 years ago and treated with anthracyclines at age 15-39 years were identified in a retrospective single institution registry. To ensure adequate follow-up, at least 2 follow-up visits during 2015-19 were required. Data abstracted on eligible subjects included documentation of pre-treatment LVEF evaluation, clinical rationale and treatment regimen. CV risk factors and events were collected pre-treatment and during follow-up. Descriptive statistics were used to summarize data. Results: 64/115 (56%) of AYA lymphoma patients underwent pre-treatment LVEF assessment. Rationale for/against LVEF assessment was rarely documented: low CV risk was recorded as rationale for no LVEF assessment in 2 subjects. Among AYAs who underwent pre-treatment LVEF assessment, no significant abnormalities were detected and no changes in subsequent treatment plans were found. During median follow-up of 6.7 (inter-quartile range 5.4-9.5) years, 6/115 (5%) experienced CV events. Only 2 (1.7%) survivors experienced potential anthracycline-related CV events: 1 moderate cardiomyopathy at 9 years, 1 peri-partum cardiomyopathy and atrial fibrillation due to post-radiation SVC occlusion at 15 years post-treatment. Both these AYAs (aged 38 and 31 years at time of CV events) also had other CV risk factors- family history, smoking, obesity, and hyperlipidemia. Four (3.5%) survivors’ experienced CV events (1 sinus tachyarrhythmia, 1 junctional rhythm, 2 acute/asymptomatic drop in LVEF) unrelated to anthracyclines with clear alternative etiology e.g. sepsis/symptom burden. There was no correlation between having pre-treatment LVEF assessment and occurrence of CV events. 13/115 (11.3%) developed new CV risk factors: 4 hypertension, 6 hyperlipidemia, 3 diabetes. Conclusions: Pre-treatment LVEF assessment is done inconsistently in AYA lymphoma patients but does not impact initial treatment or predict late cardiotoxicity. CV events in long-term AYA lymphoma survivors are rare but evaluation of CV risk factors, early detection and management may be more important than focusing on LVEF assessment.


2020 ◽  
Vol 58 (5) ◽  
pp. 923-931
Author(s):  
Federico Sertic ◽  
Dieynaba Diagne ◽  
Lexy Chavez ◽  
Thomas Richards ◽  
Ashley Berg ◽  
...  

Abstract OBJECTIVES There has been increasing interest in using extracorporeal membrane oxygenation (ECMO) to rescue patients with pulmonary embolism (PE) in the advanced stages of respiratory or haemodynamic decompensation. We examined mid-term outcomes and risk factors for in-hospital mortality. METHODS We conducted a retrospective study of 36 patients who required ECMO placement (32 veno-arterial ECMO, 4 veno-venous) following acute PE. Survival curves were estimated using the Kaplan–Meier method. Risk factors for in-hospital mortality were assessed by logistic regression analysis. Functional status and quality of life were assessed by phone questionnaire. RESULTS Overall survival to hospital discharge was 44.4% (16/36). Two-year survival conditional to discharge was 94% (15/16). Two-year survival after veno-arterial ECMO was 39% (13/32). In patients supported with veno-venous ECMO, survival to discharge was 50%, and both patients were alive at follow-up. In univariable analysis, a history of recent surgery (P = 0.064), low left ventricular ejection fraction (P = 0.029), right ventricular dysfunction ≥ moderate at weaning (P = 0.083), on-going cardiopulmonary resuscitation at ECMO placement (P = 0.053) and elevated lactate at weaning (P = 0.002) were risk factors for in-hospital mortality. In multivariable analysis, recent surgery (P = 0.018) and low left ventricular ejection fraction at weaning (P = 0.013) were independent factors associated with in-hospital mortality. At a median follow-up of 23 months, 10 patients responded to our phone survey; all had acceptable functional status and quality of life. CONCLUSIONS Massive acute PE requiring ECMO support is associated with high early mortality, but patients surviving to hospital discharge have excellent mid-term outcomes with acceptable functional status and quality of life. ECMO can provide a stable platform to administer other intervention with the potential to improve outcomes. Risk factors for in-hospital mortality after PE and veno-arterial ECMO support were identified.


EP Europace ◽  
2020 ◽  
Vol 22 (Supplement_2) ◽  
pp. ii54-ii60
Author(s):  
Yuqiu Li ◽  
Lirong Yan ◽  
Yan Dai ◽  
Yu’an Zhou ◽  
Qi Sun ◽  
...  

Abstract Aims The present study was to evaluate the feasibility and clinical outcomes of left bundle branch area pacing (LBBAP) in cardiac resynchronization therapy (CRT)-indicated patients. Methods and results LBBAP was performed via transventricular septal approach in 25 patients as a rescue strategy in 5 patients with failed left ventricular (LV) lead placement and as a primary strategy in the remaining 20 patients. Pacing parameters, procedural characteristics, electrocardiographic, and echocardiographic data were assessed at implantation and follow-up. Of 25 enrolled CRT-indicated patients, 14 had left bundle branch block (LBBB, 56.0%), 3 right bundle branch block (RBBB, 12.0%), 4 intraventricular conduction delay (IVCD, 16.0%), and 4 ventricular pacing dependence (16.0%). The QRS duration (QRSd) was significantly shortened by LBBAP (intrinsic 163.6 ± 29.4 ms vs. LBBAP 123.0 ± 10.8 ms, P &lt; 0.001). During the mean follow-up of 9.1 months, New York Heart Association functional class was improved to 1.4 ± 0.6 from baseline 2.6 ± 0.6 (P &lt; 0.001), left ventricular ejection fraction (LVEF) increased to 46.9 ± 10.2% from baseline 35.2 ± 7.0% (P &lt; 0.001), and LV end-diastolic dimensions (LVEDD) decreased to 56.8 ± 9.7 mm from baseline 64.1 ± 9.9 mm (P &lt; 0.001). There was a significant improvement (34.1 ± 7.4% vs. 50.0 ± 12.2%, P &lt; 0.001) in LVEF in patients with LBBB. Conclusion The present study demonstrates the clinical feasibility of LBBAP in CRT-indicated patients. Left bundle branch area pacing generated narrow QRSd and led to reversal remodelling of LV with improvement in cardiac function. LBBAP may be an alternative to CRT in patients with failure of LV lead placement and a first-line option in selected patients such as those with LBBB and heart failure.


Sign in / Sign up

Export Citation Format

Share Document