ABSTRACT
We evaluated the liver transplantation (LT) criteria in acute-on-chronic liver failure (ACLF), incorporating an urgent living-donor LT (LDLT) program. Critically ill patients with a Chronic Liver Failure Consortium (CLIF-C) ACLF score (CLIF-C_ACLF_score) ≥65, previously considered unsuitable for LT, were included to explore the excess mortality threshold of the CLIF-C_ACLF_score (CLIF-C_ACLF_score_threshold). We followed 854 consecutive patients with ACLF (276 ACLF grade 2 and 215 ACLF grade 3) over 10 years among 4432 LT recipients between 2008 and 2019. For advanced ACLF patients without immediate deceased-donor (DD) allocation, an urgent LDLT program was expedited. The CLIF-C_ACLF_score_threshold was determined by the metrics of transplant survival benefit: >60% 1-year and >50% 5-year survival rate. In predicting post-LT mortality, the CLIF-C_ACLF_score outperformed the (model for end-stage liver disease-sodium) MELD-Na and (model for end-stage liver disease) MELD-3.0 scores but was comparable to the Sundaram ACLF-LT-mortality score. A CLIF-C_ACLF_score ≥65 (n = 54) demonstrated posttransplant survival benefits, with 1-year and 5-year survival rates of 66.7% and 50.4% (P < .001), respectively. Novel CLIF-C_ACLF_score_threshold for 1-year and 5-year mortalities was 70 and 69, respectively. A CLIF-C_ACLF_score-based nomogram for predicting survival probabilities, integrating cardiovascular disease, diabetes, and donor type (LDLT vs DDLT), was generated. This study suggests reconsidering the criteria for unsuitable LT with a CLIF-C_ACLF_score ≥65. Implementing a timely salvage LT strategy, and incorporating urgent LDLT, can enhance survival rates.
ABSTRACT
Low skeletal muscle mass may develop in children with end-stage liver disease, affecting postoperative outcomes. We retrospectively investigated whether preoperative low muscle mass was associated with early postoperative outcomes in pediatric patients undergoing living donor liver transplantation (LDLT). Electronic medical records of children (age below 12 y) who underwent LDLT between February 1, 2007, and January 31, 2018, were reviewed. The cross-sectional areas of psoas, quadratus lumborum, and erector spinae muscles at the level of fourth-fifth lumbar intervertebral disks were measured using abdominal CT images, divided by the square of the height and were added to obtain the total skeletal muscle index (TSMI). The patients were divided into two groups according to the median TSMI in the second quintile (1859.1 mm 2 /m 2 ). Complications in the early postoperative period (within 30 d after surgery) classified as Clavien-Dindo grade 3 or higher were considered major complications. Logistic regression analyses were performed to determine the association between preoperative low muscle mass and early postoperative outcomes. In the study population of 123 patients (median age, 14 mo; range, 8-38 mo) who underwent LDLT, 29% and 71% were classified in the low (mean TSMI, 1642.5 ± 187.0 mm 2 /m 2 ) and high (mean TSMI 2188.1 ± 273.5 mm 2 /m 2 ) muscle mass groups, respectively. The rates of major complications, mechanical ventilation >96 hours, intensive care unit stay >14 days, hospital stay >30 days, and in-hospital mortality were not significantly different between the 2 groups. Additionally, adverse outcomes according to pediatric end-stage liver disease scores and sex were not significantly different between the 2 groups. In conclusion, preoperative low muscle mass defined by TSMI was not associated with early postoperative outcomes in pediatric patients undergoing LDLT.
Subject(s)
End Stage Liver Disease , Liver Transplantation , Humans , Child , Adolescent , Liver Transplantation/methods , Retrospective Studies , End Stage Liver Disease/complications , Living Donors , Psoas Muscles/diagnostic imaging , Severity of Illness Index , Postoperative Period , Postoperative Complications/epidemiology , Postoperative Complications/etiologyABSTRACT
BACKGROUND: Acute kidney injury (AKI) is one of the most common complications after liver transplantation (LT) and can significantly impact outcomes. The presence of hepatitis C virus (HCV) infection increases the risk of AKI development. However, the impact of HCV on AKI after LT has not been evaluated. The aim of this study was to assess the effect of HCV on AKI development in patients who underwent LT. METHODS: Between January 2008 and April 2023, 2183 patients who underwent living donor LT (LDLT) were included. Patients were divided into 2 groups based on the presence of chronic HCV infection. We compared LT recipients using the propensity score matching (PSM) method. Factors associated with AKI development were evaluated using multiple logistic regression analysis. In addition, 1-year mortality and graft failure were assessed using a Cox proportional regression model. RESULTS: Among 2183 patients, the incidence of AKI was 59.2%. After PSM, the patients with HCV showed a more frequent development of AKI (71.9% vs 63.9%, P = .026). In multivariate analysis after PSM, HCV was associated with AKI development (odds ratio [OR], 1.53; 95% confidence interval [CI], 1.06-2.20, P = .022), 1-year mortality (Hazard ratio [HR], 1.98; 95% CI, 1.12-3.52, P = .019), and graft failure (HR, 2.12; 95% CI, 1.22-3.69, P = .008). CONCLUSIONS: The presence of HCV was associated with increased risk for the development of AKI, 1-year mortality, and graft failure after LT.
ABSTRACT
INTRODUCTION AND OBJECTIVES: Acute kidney injury (AKI) is prevalent and has deleterious effects on postoperative outcomes following liver transplantation (LT). The impact of nonselective beta-blockers (NSBBs) in patients with liver cirrhosis remains controversial. This study investigated the association between preoperative NSBB use and AKI after living donor LT (LDLT). PATIENTS AND METHODS: We evaluated 2,972 adult LDLT recipients between January 2012 and July 2022. The patients were divided into two groups based on the preoperative NSBB use. Propensity score matched (PSM) and inverse probability of treatment weighting (IPTW) analyses were performed to evaluate the association between preoperative NSBB use and postoperative AKI. Multiple logistic regression analyses were also used to identify the risk factors for AKI. RESULTS: The overall incidence of AKI was 1,721 (57.9%) cases. The NSBB group showed a higher incidence of AKI than the non-NSBB group (62.4% vs. 56.7%; P = 0.011). After PSM and IPTW analyses, no significant difference in the incidence of AKI was found between the two groups (Odds ratio, OR 1.13, 95% confidence interval, CI 0.93-1.37, P = 0.230, PSM analysis; OR 1.20, 95% CI 0.99-1.44, P = 0.059, IPTW analysis). In addition, preoperative NSBB use was not associated with AKI after multivariate logistic regression analysis (OR 1.16, 95% CI 0.96-1.40, P = 0.118). CONCLUSIONS: Preoperative NSBB use was not associated with AKI after LDLT. Further studies are needed to validate our results.
Subject(s)
Acute Kidney Injury , Adrenergic beta-Antagonists , Liver Transplantation , Living Donors , Propensity Score , Humans , Acute Kidney Injury/epidemiology , Acute Kidney Injury/chemically induced , Acute Kidney Injury/diagnosis , Acute Kidney Injury/etiology , Liver Transplantation/adverse effects , Female , Male , Middle Aged , Incidence , Risk Factors , Adrenergic beta-Antagonists/therapeutic use , Adrenergic beta-Antagonists/adverse effects , Retrospective Studies , Adult , Postoperative Complications/epidemiology , Postoperative Complications/etiology , Preoperative Care/methods , Liver Cirrhosis/surgery , Liver Cirrhosis/complications , Risk AssessmentABSTRACT
BACKGROUND: Diastolic dysfunction is regarded as an important predictor of outcome after liver transplantation (LT). We investigated the influence of liver disease severity on left ventricular diastolic properties using end-diastolic pressure-volume relationship (EDPVR) analysis in patients with end-stage liver disease (ESLD). Association between alterations of the EDPVR and mortality after LT was evaluated. METHODS: In this observational retrospective cohort study, 3,211 patients who underwent LT for ESLD were included in analysis. Variables derived from single-beat EDPVR (diastolic stiffness-coefficient [ß] and end-diastolic volume at an end-diastolic pressure of 20 mmHg [EDVI20] indicating ventricular capacitance) were estimated using preoperative echocardiographic data. Alterations in EDPVR with increased stiffness (ß > 6.16) were categorized into 3 groups; leftward-shifted (EDVI20 <51 mL/m2), rightward-shifted (EDVI20 > 69.7 mL/m2), and intermediate (EDVI20 51-69.7 mL/m2). RESULTS: As the model for ESLD score increases, both EDVI20 and ß gradually increased, which indicated ventricular remodeling with larger capacitance and higher diastolic stiffness. Among patients with increased stiffness (ß > 6.16, n = 1,090), survival rates after LT were lower in leftward-shifted EDPVR than in rightward-shifted EDPVR (73.7% vs 82.9%; log-rank P = 0.002). In the adjusted Cox proportional hazard model, risk of cumulative all-cause mortality at 11 years was the highest in leftward-shifted EDPVR (hazard ratio [HR]: 1.93; 95% confidence interval [CI]: 1.27-2.92), followed by intermediate EDPVR (HR: 1.55; 95% CI: 1.12-2.26), compared with rightward-shifted EDPVR. The SHapley Additive exPlanation model revealed that the variables associated with leftward-shifted EDPVR were diabetes, female sex, old age, and hypertension. CONCLUSIONS: As ESLD advances, diastolic ventricular properties are characterized by increased EDVI20 and ß on rightward-shifted EDPVR, indicating larger capacitance and higher stiffness. However, leftward-shifted EDPVR with left ventricle remodeling failure is associated with poor post-LT survival.
Subject(s)
End Stage Liver Disease , Ventricular Remodeling , Humans , Female , Retrospective Studies , Blood Pressure , End Stage Liver Disease/surgery , Diastole , Stroke Volume , Ventricular Function, LeftABSTRACT
BACKGROUND: Excessive visceral obesity in recipients of living donor liver transplantation (LDLT) is associated with mortality, and a recent study reported the correlation between visceral adiposity of male LDLT recipients and hepatocellular carcinoma (HCC) recurrence. However, there is no study on the relationship between the donor's visceral adiposity and surgical outcomes in LDLT recipients. We investigated the association of the visceral-to-subcutaneous fat area ratio (VSR) in donors and recipients with HCC recurrence and mortality in LDLT. METHODS: We analyzed 1386 sets of donors and recipients who underwent LDLT between January 2008 and January 2018. The maximal chi-square method was used to determine the optimal cutoff values for VSR for predicting overall HCC recurrence and mortality. Cox regression analyses were performed to evaluate the association of donor VSR and recipient VSR with overall HCC recurrence and mortality in recipients. RESULTS: The cutoff values of VSR was determined as 0.73 in males and 0.31 in females. High donor VSR was significantly associated with overall HCC recurrence (adjusted hazard ratio [HR]: 1.43, 95% confidence interval [CI]: 1.06-1.93, p = 0.019) and mortality (HR: 1.35, 95% CI: 1.03-1.76, p = 0.030). High recipient VSR was significantly associated with overall HCC recurrence (HR: 1.40, 95% CI: 1.04-1.88, p = 0.027) and mortality (HR: 1.50, 95% CI: 1.14-1.96, p = 0.003). CONCLUSIONS: Both recipient VSR and donor VSR were significant risk factors for HCC recurrence and mortality in LDLT recipients. Preoperative donor VSR and recipient VSR may be strong predictors of the surgical outcomes of LDLT recipients with HCC.
Subject(s)
Carcinoma, Hepatocellular , Liver Neoplasms , Liver Transplantation , Female , Male , Humans , Carcinoma, Hepatocellular/surgery , Carcinoma, Hepatocellular/pathology , Liver Transplantation/adverse effects , Liver Transplantation/methods , Liver Neoplasms/surgery , Liver Neoplasms/pathology , Living Donors , Neoplasm Recurrence, Local/epidemiology , Neoplasm Recurrence, Local/etiology , Obesity, Abdominal/etiology , Retrospective Studies , Treatment OutcomeABSTRACT
BACKGROUND: A recent study reported a correlation between the muscle mass of male donors and graft failure in living donor liver transplantation (LDLT) recipients. We investigated the association of sex-specific donor skeletal muscle index (SMI) with mortality and graft failure in LDLT recipients. METHODS: We retrospectively analysed 2750 sets of donors and recipients between January 2008 and January 2018. The recipient outcomes were analysed by dividing the data according to donor sex. Cox regression analyses were performed to evaluate the association between donor SMI by sex and 1-year mortality and graft failure in recipients. RESULTS: In the male donor group, robust donor (increased SMI) was significantly associated with higher risks for mortality (hazard ratio [HR]: 1.03, 95% confidence interval [CI]: 1.00-1.06, p = .023) and graft failure (HR: 1.04, 95% CI: 1.01-1.06, p = .007) at 1 year. In the female donor group, the robust donor was significantly associated with lower risks for mortality (HR: 0.92, 95% CI: 0.87-0.97, p = .003) and graft failure (HR: 0.95, 95% CI: 0.90-1.00, p = .032) at 1 year. CONCLUSIONS: Donor SMI was associated with surgical outcomes in recipients. Robust male and female donors were a significant negative and protective factor for grafts respectively.
Subject(s)
Liver Transplantation , Humans , Male , Female , Living Donors , Retrospective Studies , Treatment Outcome , Muscle, Skeletal , Graft SurvivalABSTRACT
Recent studies have reported that sarcopenia influences morbidity and mortality in surgical patients. However, few studies have investigated the associations of sarcopenia with short-term and long-term graft failure in recipients after living donor liver transplantation (LDLT). In this study, we investigated the associations between sarcopenia and graft failure/mortality in patients undergoing LDLT. We retrospectively examined 2816 recipients who underwent LDLT between January 2008 and January 2018. Cox regression analysis was performed to evaluate the associations between sarcopenia and graft failure/mortality in recipients at 60 days, 180 days, and 1 year and overall. Sarcopenia in the recipient was significantly associated with 60-day graft failure (adjusted hazard ratio [HR], 1.98; 95% confidence interval [CI], 1.09-3.61; p = 0.03), 180-day graft failure (HR, 1.85; 95% CI, 1.19-2.88; p = 0.01), 1-year graft failure (HR, 1.45; 95% CI, 1.01-2.17; p = 0.05), and overall graft failure (HR, 1.42; 95% CI, 1.08-1.87; p = 0.01). In addition, recipient sarcopenia was associated with 180-day mortality (HR, 1.88; 95% CI, 1.17-3.01; p = 0.01), 1-year mortality (HR, 1.53; 95% CI, 1.01-2.29; p = 0.04), and overall mortality (HR, 1.43; 95% CI, 1.08-1.90; p = 0.01). Preoperative sarcopenia was associated with high rates of graft failure and mortality in LDLT recipients. Therefore, preoperative sarcopenia may be a strong predictor of the surgical prognosis in LDLT recipients.
Subject(s)
Liver Transplantation , Sarcopenia , Graft Survival , Humans , Liver Transplantation/adverse effects , Living Donors , Retrospective Studies , Sarcopenia/complications , Sarcopenia/epidemiology , Treatment OutcomeABSTRACT
BACKGROUND AND AIMS: Despite frequent cirrhotic cardiomyopathy or subclinical heart failure (HF), the prognostic value of peri-liver transplant (LT) B-type natriuretic peptide (BNP) has been poorly studied in advanced liver disease. We examined the association between BNP and mortality in a large cohort of LT patients and identified risk factors for peri-LT BNP increase. APPROACH AND RESULTS: Using prospectively collected data from the Asan LT Registry between 2008 and 2019, 3,811 patients who measured serial pretransplant BNP (preBNP) and peak BNP levels within the first 3 posttransplant days (postBNPPOD3 ) were analyzed. Thirty-day all-cause mortality predicted by adding preBNP and/or postBNPPOD3 to the traditional Revised Cardiac Risk Index (RCRI) was evaluated. PreBNP > 400 pg/mL (known cutoff of acute HF) was found in 298 (7.8%); however, postBNPPOD3 > 400 pg/mL was identified in 961 (25.2%) patients, specifically in 40.4% (531/1,315) of those with a Model for End-Liver Disease score (MELDs) > 20. Strong predictors of postBNPPOD3 > 400 pg/mL were preBNP, hyponatremia, and MELDs, whereas those of preBNP > 400 pg/mL were MELDs, kidney failure, and respiratory failure. Among 100 (2.6%) post-LT patients who died within 30 days, patients with postBNPPOD3 ≤ 150 pg/mL (43.1%, reference group), 150-400 pg/mL (31.7%), 400-1,000 pg/mL (18.5%), 1,000-2,000 pg/mL (4.7%), and >2,000 pg/mL (2.0%) had 30-day mortalities of 0.9%, 2.2%, 4.0%, 7.7%, and 22.4%, respectively. Adding preBNP, postBNPPOD3 , and both BNP to RCRI improved net reclassification index to 22.5%, 29.5%, and 33.1% of 30-day mortality, respectively. CONCLUSIONS: PostBNPPOD3 > 400 pg/mL after LT was markedly prevalent in advanced liver disease and mainly linked to elevated preBNP. Routine monitoring of peri-LT BNP provides incremental prognostic information; therefore, it could help risk stratification for mortality as a practical and useful biomarker in LT.
Subject(s)
End Stage Liver Disease/surgery , Liver Transplantation/adverse effects , Natriuretic Peptide, Brain/blood , Biomarkers/blood , End Stage Liver Disease/blood , End Stage Liver Disease/diagnosis , End Stage Liver Disease/mortality , Female , Hospital Mortality , Humans , Liver Transplantation/statistics & numerical data , Male , Middle Aged , Perioperative Period , Prognosis , Prospective Studies , Republic of Korea/epidemiology , Risk Assessment/methods , Risk Assessment/statistics & numerical data , Risk Factors , Severity of Illness IndexABSTRACT
BACKGROUND: Although living donor liver transplantation (LDLT) is the standard treatment option for patients with end-stage liver disease, it always entails ethical concerns about the risk of living donors. Recent studies have reported a correlation between sarcopenia and surgical prognosis in recipients. However, there are few studies of donor sarcopenia and the surgical prognosis of donors. This study investigated the association between sarcopenia and postoperative acute kidney injury in liver donors. METHODS: This retrospective study analysed 2892 donors who underwent donor hepatectomy for LDLT between January 2008 and January 2018. Sarcopenia was classified into pre-sarcopenia and severe sarcopenia, which were determined to be -1 standard deviation (SD), and -2 SD from the mean baseline of the skeletal muscle index, respectively. Multivariate regression analysis was performed to evaluate the association between donor sarcopenia and postoperative AKI. Additionally, we assessed the association between donor sarcopenia and delayed recovery of liver function (DRHF). RESULTS: In the multivariate analysis, donor sarcopenia was significantly associated a higher incidence of postoperative AKI (adjusted odds ratio [OR]: 2.65, 95% confidence interval [CI]: 1.15-6.11, P = .022 in pre-sarcopenia, OR: 5.59, 95% CI: 1.11-28.15, P = .037 in severe sarcopenia, respectively). Additionally, hypertension and synthetic colloid use were significantly associated with postoperative AKI. In the multivariate analysis, risk factors of DRHF were male gender, indocyanine green retention rate at 15 minutes, and graft type, however, donor sarcopenia was not a risk factor. CONCLUSIONS: Donor sarcopenia is associated with postoperative AKI following donor hepatectomy.
Subject(s)
Acute Kidney Injury , Liver Transplantation , Acute Kidney Injury/epidemiology , Acute Kidney Injury/etiology , Cohort Studies , Hepatectomy/adverse effects , Humans , Liver/surgery , Liver Transplantation/adverse effects , Living Donors , Male , Muscle, Skeletal , Retrospective StudiesABSTRACT
BACKGROUND AND AIMS: Enhanced sympathetic nervous activation and peripheral vasodilation in end-stage liver disease (ESLD) may limit the importance of left ventricular ejection fraction (LVEF) as an influential prognosticator. We sought to understand the LVEF and cardiac dimensions in ESLD patients in order to define the LVEF threshold to predict all-cause mortality after liver transplantation (LT). APPROACH AND RESULTS: Data were collected prospectively from the Asan LT Registry between 2008 and 2016, and outcomes were retrospectively reviewed. LVEF, end-diastolic volume index (EDVI), and end-diastolic elastance (Eed) were measured by preoperative echocardiography. Of 2,799 patients, 452 (16.2%) had LVEF ≤ 60%, with 29 (1.0%) having LVEF < 55% and 269 (9.6%) had LVEF ≥ 70%. Over a median of 5.4-year follow-up, 329 (11.8%) patients died: 104 (3.7%) died within 90 days. LVEF (range, 30%-81%) was directly proportionate to Model for End-stage Liver Disease (MELD) scores, an index of liver disease severity, in survivors but showed a fixed flat-line pattern in nonsurvivors (interaction P = 0.004 between groups), with lower EDVI (P = 0.013) and higher Eed (P = 0.001) in the MELD ≥ 20 group. Patients with LVEF ≤ 60% had higher 90-day (13% vs. 7.4%; log rank, P = 0.03) and median 5.4-year (26.7% vs. 16.2%; log rank, P = 0.003) mortality rates in the MELD ≥ 20 group, respectively, compared to those with LVEF > 60%. Specifically, in the MELD > 35 group, median 5.4-year mortality rate was 53.3% in patients with LVEF ≤ 60% versus 24% in those with LVEF > 60% (log rank P < 0.001). By contrast, mortality rates of LVEF ≤ 60% and > 60% were similar in the MELD < 20 group (log rank P = 0.817). CONCLUSIONS: LVEF ≤ 60% is strongly associated with higher post-LT mortality rates in the MELD ≥ 20 group, indicating the need to appraise both LVEF and liver disease severity simultaneously. Enhanced diastolic elastance with low EDVI provides insights into pathogenesis of low LVEF in nonsurvivors with MELD ≥ 20.
Subject(s)
End Stage Liver Disease/mortality , End Stage Liver Disease/surgery , Liver Transplantation , Stroke Volume , Female , Humans , Male , Middle Aged , Registries , Severity of Illness Index , Ventricular Function, LeftABSTRACT
BACKGROUND AND AIM: The proportional increase of corrected QT interval (QTc) along end-stage liver disease (ESLD) severity may lead to inconsistent outcome reporting if based on conventional threshold of prolonged QTc. We investigated the comprehensive QTc distribution among ESLD patients and assessed the association between QTc > 500 ms, a criterion for diagnosing severe long-QT syndrome, and the 30-day major adverse cardiovascular event (MACE) after liver transplantation (LT) and identified the risk factors for developing QTc > 500 ms. METHODS: Data were collected prospectively from the Asan LT Registry between 2011 and 2018, and outcomes were retrospectively reviewed. Multivariable analysis and propensity score-weighted adjusted odds ratios (ORs) were calculated. Thirty-day MACEs were defined as the composite of cardiovascular mortality, arrhythmias, myocardial infarction, pulmonary thromboembolism, and/or stroke. RESULTS: Of 2579 patients, 194 (7.5%) had QTc > 500 ms (QTc500_Group), and 1105 (42.8%) had prolonged QTc (QTcP_Group), defined as QTc > 470 ms for women and >450 ms for men. The 30-day MACE occurred in 336 (13%) patients. QTc500_Group showed higher 30-day MACE than did those without (20.1% vs 12.5%, P = 0.003), with corresponding adjusted OR of 1.24 (95% CI: 1.06-1.46, P = 0.007). However, QTcP_Group showed comparable 30-day MACE (13.3% vs 12.8% without prolonged QTc, P = 0.764). Significant risk factors for QTc > 500 ms development were advanced liver disease, female sex, hypokalemia, hypocalcemia, high left ventricular end-diastolic volume, and tachycardia. CONCLUSION: Our results revealed that, among ESLD patients, a novel threshold of QTc > 500 ms was associated with post-LT 30-day MACE but not with conventional threshold, indicating that a longer QTc threshold should be considered for this unique patient population.
Subject(s)
Cardiovascular Diseases/etiology , End Stage Liver Disease/complications , End Stage Liver Disease/surgery , Liver Transplantation/adverse effects , Long QT Syndrome/etiology , Adult , Aged , Cardiac Volume , Diastole , Female , Humans , Hypocalcemia , Hypokalemia , Male , Middle Aged , Risk Factors , Severity of Illness Index , Sex Factors , Tachycardia , Time FactorsABSTRACT
BACKGROUND: We aimed to create a novel model using a deep learning method to estimate stroke volume variation (SVV), a widely used predictor of fluid responsiveness, from arterial blood pressure waveform (ABPW). METHODS: In total, 557 patients and 8,512,564 SVV datasets were collected and were divided into three groups: training, validation, and test. Data was composed of 10 s of ABPW and corresponding SVV data recorded every 2 s. We built a convolutional neural network (CNN) model to estimate SVV from the ABPW with pre-existing commercialized model (EV1000) as a reference. We applied pre-processing, multichannel, and dimension reduction to improve the CNN model with diversified inputs. RESULTS: Our CNN model showed an acceptable performance with sample data (r = 0.91, MSE = 6.92). Diversification of inputs, such as normalization, frequency, and slope of ABPW significantly improved the model correlation (r = 0.95), lowered mean squared error (MSE = 2.13), and resulted in a high concordance rate (96.26%) with the SVV from the commercialized model. CONCLUSIONS: We developed a new CNN deep-learning model to estimate SVV. Our CNN model seems to be a viable alternative when the necessary medical device is not available, thereby allowing a wider range of application and resulting in optimal patient management.
Subject(s)
Arterial Pressure , Neural Networks, Computer , Blood Pressure , Humans , Stroke VolumeABSTRACT
In patients at high risk of respiratory complications, pulse oximetry may not adequately detect hypoventilation events. Previous studies have proposed using thermography, which relies on infrared imaging, to measure respiratory rate (RR). These systems lack support from real-world feasibility testing for widespread acceptance. This study enrolled 101 spontaneously ventilating patients in a post-anesthesia recovery unit. Patients were placed in a 45° reclined position while undergoing pulse oximetry and bioimpedance-based RR monitoring. A thermography camera was placed approximately 1 m from the patient and pointed at the patient's face, recording continuously at 30 frames per second for 2 min. Simultaneously, RR was manually recorded. Offline imaging analysis identified the nares as a region of interest and then quantified nasal temperature changes frame by frame to estimate RR. The manually calculated RR was compared with both bioimpedance and thermographic estimates. The Pearson correlation coefficient between direct measurement and bioimpedance was 0.69 (R2 = 0.48), and that between direct measurement and thermography was 0.95 (R2 = 0.90). Limits of agreement analysis revealed a bias of 1.3 and limits of agreement of 10.8 (95% confidence interval 9.07 to 12.5) and - 8.13 (- 6.41 to - 9.84) between direct measurements and bioimpedance, and a bias of -0.139 and limits of agreement of 2.65 (2.14 to 3.15) and - 2.92 (- 2.41 to 3.42) between direct measurements and thermography. Thermography allowed tracking of the manually measured RR in the post-anesthesia recovery unit without requiring patient contact. Additional work is required for image acquisition automation and nostril identification.
Subject(s)
Anesthetics , Respiratory Rate , Humans , Monitoring, Physiologic , Oximetry , ThermographyABSTRACT
OBJECTIVE: This study aimed to assess the effects of remote ischemic preconditioning (RIPC) on liver function in donors and recipients after living donor liver transplantation (LDLT). BACKGROUND: Ischemia reperfusion injury (IRI) is known to be associated with graft dysfunction after liver transplantation. RIPC is used to lessen the harmful effects of IRI. METHODS: A total of 148 donors were randomly assigned to RIPC (n = 75) and control (n = 73) groups. RIPC involves 3 cycles of 5-minute inflation of a blood pressure cuff to 200 mm Hg to the upper arm, followed by 5-minute reperfusion with cuff deflation. The primary aim was to assess postoperative liver function in donors and recipients and the incidence of early allograft dysfunction and graft failure in recipients. RESULTS: RIPC was not associated with any differences in postoperative aspartate aminotransferase (AST) and alanine aminotransferase levels after living donor hepatectomy, and it did not decrease the incidence of delayed graft hepatic function (6.7% vs 0.0%, P = 0.074) in donors. AST level on postoperative day 1 [217.0 (158.0, 288.0) vs 259.5 (182.0, 340.0), P = 0.033] and maximal AST level within 7 postoperative days [244.0 (167.0, 334.0) vs 296.0 (206.0, 395.5), P = 0.029) were significantly lower in recipients who received a preconditioned graft. No differences were found in the incidence of early allograft dysfunction (4.1% vs 5.6%, P = 0.955) or graft failure (1.4% vs 5.6%, P = 0.346) among recipients. CONCLUSIONS: RIPC did not improve liver function in living donor hepatectomy. However, RIPC performed in liver donors may be beneficial for postoperative liver function in recipients after living donor liver transplantation.
Subject(s)
Ischemic Preconditioning , Liver Transplantation , Living Donors , Reperfusion Injury/prevention & control , Adult , Double-Blind Method , Female , Graft Rejection , Hepatectomy , Humans , Liver Function Tests , Male , Middle Aged , Postoperative Period , Retrospective Studies , Transplantation, HomologousABSTRACT
We aimed to determine if the severity of computed tomographic coronary angiography (CTCA)-diagnosed coronary artery disease (CAD) is associated with postliver transplantation (LT) myocardial infarction (MI) within 30 days and early mortality. We retrospectively evaluated 2118 consecutive patients who underwent CAD screening using CTCA. Post-LT type-2 MI, elicited by oxygen supply-and-demand mismatch within a month after LT, was assessed according to the severity of CTCA-diagnosed CAD. Obstructive CAD (>50% narrowing, 9.2% prevalence) was identified in 21.7% of patients with 3 or more known CAD risk factors of the American Heart Association. Post-LT MI occurred in 60 (2.8%) of total patients in whom 90-day mortality rate was 16.7%. Rates of post-LT MI were 2.1%, 3.1%, 3.4%, 4.3%, and 21.4% for normal, nonobstructive CAD, and 1-, 2-, and 3-vessel obstructive CAD, respectively. Two-vessel or 3-vessel obstructive CAD showed a 4.9-fold higher post-LT MI risk compared to normal coronary vessels. The sensitivity and negative predictive value of obstructive CAD in detecting post-LT MI were, respectively, 20% and 97.5%. In conclusion, negative CTCA finding in suspected patients can successfully exclude post-LT MI, whereas proceeding with invasive angiography is needed to further risk-stratify in patients with significant CTCA-diagnosed CAD. Prognostic role of CTCA in predicting post-LT MI needs further research.
Subject(s)
Coronary Angiography/methods , Coronary Artery Disease/surgery , Liver Transplantation/adverse effects , Myocardial Infarction/etiology , Postoperative Complications/etiology , Risk Assessment/methods , Tomography, X-Ray Computed/methods , Coronary Artery Disease/diagnosis , Coronary Artery Disease/diagnostic imaging , Female , Follow-Up Studies , Humans , Male , Middle Aged , Myocardial Infarction/pathology , Postoperative Complications/pathology , Predictive Value of Tests , Prospective Studies , Retrospective Studies , Risk FactorsABSTRACT
BACKGROUND & AIMS: Early allograft dysfunction (EAD) is predictive of poor graft and patient survival following living donor liver transplantation (LDLT). Considering the impact of the inflammatory response on graft injury extent following LDLT, we investigated the association between neutrophil-to-lymphocyte ratio (NLR) and EAD, 1-year graft failure, and mortality following LDLT, and compared it to C-reactive protein (CRP), procalcitonin, platelet-to-lymphocyte ratio and the Glasgow prognostic score. METHODS: A total of 1960 consecutive adult LDLT recipients (1531/429 as development/validation cohort) were retrospectively evaluated. Cut-offs were derived using the area under the receiver operating characteristic curve (AUROC), and multivariable regression and Cox proportional hazard analyses were performed. RESULTS: The risk of EAD increased proportionally with increasing NLR, and the NLR AUROC was 0.73, similar to CRP and procalcitonin and higher than the rest. NLR ≥ 2.85 (best cut-off) showed a significantly higher EAD occurrence (20.5% vs 5.8%, P < 0.001), higher 1-year graft failure (8.2% vs 4.9%, log-rank P = 0.009) and higher 1-year mortality (7% vs 4.5%, log-rank P = 0.039). NLR ≥ 2.85 was an independent predictor of EAD (odds ratio, 1.89 [1.26-2.84], P = 0.002) after multivariable adjustment, whereas CRP and procalcitonin were not. Increasing NLR was independently associated with higher 1-year graft failure and mortality (both P < 0.001). Consistent results in the validation cohort strengthened the prognostic value of NLR. CONCLUSIONS: Preoperative NLR ≥ 2.85 predicted higher risk of EAD, 1-year graft failure and 1-year mortality following LDLT, and NLR was superior to other parameters, suggesting that preoperative NLR may be a practical index for predicting graft function following LDLT.
Subject(s)
Liver Transplantation/mortality , Primary Graft Dysfunction/immunology , Female , Humans , Living Donors , Lymphocyte Count , Male , Middle Aged , Preoperative Period , Republic of Korea/epidemiology , Retrospective StudiesABSTRACT
AIMS: Avoiding propofol in patients with Brugada syndrome has been suggested because of the theoretical risk of provoking ventricular arrhythmias, although propofol may be selected for conscious sedation during electrophysiological procedures in catheterization laboratories. This study aimed to document periprocedural electrocardiographic changes and adverse events in patients with Brugada syndrome undergoing implantable cardioverter defibrillator (ICD) implantation using propofol sedation. METHODS: We reviewed the clinical data of 53 consecutive patients who underwent ICD implantation during 1998-2011. Sedation was achieved by combining propofol with either midazolam or fentanyl, and a bolus propofol dose (0.5-1 mg/kg) was administered to induce deep sedation. Periprocedural events, including arrhythmias, defibrillations, and hyperthermia episodes, were evaluated, and electrocardiogram (ECG) variables were measured. The need for emergency anesthetic support/intubation and incidence of perioperative complications or mortality were analyzed. RESULTS: Procedure time and cumulative propofol dose for each patient was 125.2 (42.8) min and 204.6 (212.7) mg, respectively. During deep sedation, blood pressure, heart rate, and oxygen saturation were significantly decreased (P < 0.001) such that eight (15.1%) patients required manual ventilation and one (1.9%) needed atropine injection. No significant ECG changes were observed. Only two (3.7%) patients showed newly developed ST elevation in the anterior precordial lead, whereas three (5.6%) had isolated premature ventricular contractions. CONCLUSION: ICD implantation without significant ECG changes or adverse outcomes is feasible under propofol sedation in patients with Brugada syndrome. However, because of significant hemodynamic changes and respiratory compromise, close monitoring and meticulous propofol dose titration is warranted.
Subject(s)
Brugada Syndrome/therapy , Conscious Sedation/methods , Defibrillators, Implantable , Hypnotics and Sedatives/administration & dosage , Propofol/administration & dosage , Adjuvants, Anesthesia/administration & dosage , Adult , Female , Fentanyl/administration & dosage , Humans , Male , Midazolam/administration & dosage , Monitoring, Physiologic , Treatment OutcomeABSTRACT
BACKGROUND: Although desflurane and sevoflurane, the most commonly used inhalational anesthetics, have been linked to postoperative liver injury, their impact on liver regeneration remains unclear. We compared the influence of these anesthetics on the postoperative liver regeneration index (LRI) after living donor hepatectomy (LDH). METHODS: We conducted a retrospective chart review of 1629 living donors who underwent right hepatectomy for LDH between January 2008 and August 2016. The patients were divided into sevoflurane (n = 1206) and desflurane (n = 423) groups. Factors associated with LRI were investigated using multivariable logistic regression analysis. Propensity score matching analysis compared early (1 postoperative week) and late (within 1-2 months) LRIs and delayed recovery of hepatic function between the 2 groups. RESULTS: The mean early and late LRIs in the 1629 patients were 63.3% ± 41.5% and 93.7% ± 48.1%, respectively. After propensity score matching (n = 403 pairs), there were no significant differences in early and late LRIs between the sevoflurane and desflurane groups (early LRI: 61.2% ± 41.5% vs 58.9% ± 42.4%, P = .438; late LRI: 88.3% ± 44.3% vs 94.6% ± 52.4%, P = .168). Male sex (regression coefficient [ß], 4.6; confidence interval, 1.6-7.6; P = .003) and remnant liver volume (ß, -4.92; confidence interval, -5.2 to -4.7; P < .001) were associated with LRI. The incidence of delayed recovery of hepatic function was 3.6% (n = 29) with no significant difference between the 2 groups (3.0% vs 4.2%, P = .375) after LDH. CONCLUSIONS: Both sevoflurane and desflurane can be safely used without affecting liver regeneration and delaying liver function recovery after LDH.
Subject(s)
Anesthetics, Inhalation/administration & dosage , Hepatectomy/trends , Liver Regeneration/drug effects , Liver Regeneration/physiology , Living Donors , Propensity Score , Adult , Desflurane/administration & dosage , Female , Humans , Male , Retrospective Studies , Sevoflurane/administration & dosageABSTRACT
BACKGROUND: Postreperfusion syndrome (PRS) has been shown to be related to postoperative morbidity and graft failure in orthotopic liver transplantation. To date, little is known about the impact of PRS on the prevalence of postoperative acute kidney injury (AKI) and the postoperative outcomes after living donor liver transplantation (LDLT). The purpose of our study was to determine the impact of PRS on AKI and postoperative outcomes after LDLT surgery. METHODS: Between January 2008 and October 2015, we retrospectively collected and evaluated the records of 1865 patients who underwent LDLT surgery. We divided the patients into 2 groups according to the development of PRS: PRS group (n = 715) versus no PRS group (n = 1150). Risk factors for AKI and mortality were investigated by multivariable logistic and Cox proportional hazards regression model analysis. Propensity score (PS) analysis (PS matching and inverse probability of treatment weighting analysis) was designed to compare the outcomes between the 2 groups. RESULTS: The prevalence of PRS and the mortality rate were 38% and 7%, respectively. In unadjusted analyses, the PRS group showed more frequent development of AKI (P < .001), longer hospital stay (P = .010), and higher incidence of intensive care unit stay over 7 days (P < .001) than the no PRS group. After PS matching and inverse probability of treatment weighting analysis, the PRS group showed a higher prevalence of postoperative AKI (P = .023 and P = .017, respectively) and renal dysfunction 3 months after LDLT (P = .036 and P = .006, respectively), and a higher incidence of intensive care unit stay over 7 days (P = .014 and P = .032, respectively). CONCLUSIONS: We demonstrated that the magnitude and duration of hypotension caused by PRS is a factor contributing to the development of AKI and residual renal dysfunction 3 months after LDLT.