RESUMEN
BACKGROUND: Surgical video recording provides the opportunity to acquire intraoperative data that can subsequently be used for a variety of quality improvement, research, and educational applications. Various recording devices are available for standard operating room camera systems. Some allow for collateral data acquisition including activities of the OR staff, kinematic measurements (motion of surgical instruments), and recording of the endoscopic video streams. Additional analysis through computer vision (CV), which allows software to understand and perform predictive tasks on images, can allow for automatic phase segmentation, instrument tracking, and derivative performance-geared metrics. With this survey, we summarize available surgical video acquisition technologies and associated performance analysis platforms. METHODS: In an effort promoted by the SAGES Artificial Intelligence Task Force, we surveyed the available video recording technology companies. Of thirteen companies approached, nine were interviewed, each over an hour-long video conference. A standard set of 17 questions was administered. Questions spanned from data acquisition capacity, quality, and synchronization of video with other data, availability of analytic tools, privacy, and access. RESULTS: Most platforms (89%) store video in full-HD (1080p) resolution at a frame rate of 30 fps. Most (67%) of available platforms store data in a Cloud-based databank as opposed to institutional hard drives. CV powered analysis is featured in some platforms: phase segmentation in 44% platforms, out of body blurring or tool tracking in 33%, and suture time in 11%. Kinematic data are provided by 22% and perfusion imaging in one device. CONCLUSION: Video acquisition platforms on the market allow for in depth performance analysis through manual and automated review. Most of these devices will be integrated in upcoming robotic surgical platforms. Platform analytic supplementation, including CV, may allow for more refined performance analysis to surgeons and trainees. Most current AI features are related to phase segmentation, instrument tracking, and video blurring.
Asunto(s)
Inteligencia Artificial , Procedimientos Quirúrgicos Robotizados , Humanos , Endoscopía , Programas Informáticos , Privacidad , Grabación en VideoRESUMEN
BACKGROUND: Many surgical adverse events, such as bile duct injuries during laparoscopic cholecystectomy (LC), occur due to errors in visual perception and judgment. Artificial intelligence (AI) can potentially improve the quality and safety of surgery, such as through real-time intraoperative decision support. GoNoGoNet is a novel AI model capable of identifying safe ("Go") and dangerous ("No-Go") zones of dissection on surgical videos of LC. Yet, it is unknown how GoNoGoNet performs in comparison to expert surgeons. This study aims to evaluate the GoNoGoNet's ability to identify Go and No-Go zones compared to an external panel of expert surgeons. METHODS: A panel of high-volume surgeons from the SAGES Safe Cholecystectomy Task Force was recruited to draw free-hand annotations on frames of prospectively collected videos of LC to identify the Go and No-Go zones. Expert consensus on the location of Go and No-Go zones was established using Visual Concordance Test pixel agreement. Identification of Go and No-Go zones by GoNoGoNet was compared to expert-derived consensus using mean F1 Dice Score, and pixel accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV). RESULTS: A total of 47 frames from 25 LC videos, procured from 3 countries and 9 surgeons, were annotated simultaneously by an expert panel of 6 surgeons and GoNoGoNet. Mean (± standard deviation) F1 Dice score were 0.58 (0.22) and 0.80 (0.12) for Go and No-Go zones, respectively. Mean (± standard deviation) accuracy, sensitivity, specificity, PPV and NPV for the Go zones were 0.92 (0.05), 0.52 (0.24), 0.97 (0.03), 0.70 (0.21), and 0.94 (0.04) respectively. For No-Go zones, these metrics were 0.92 (0.05), 0.80 (0.17), 0.95 (0.04), 0.84 (0.13) and 0.95 (0.05), respectively. CONCLUSIONS: AI can be used to identify safe and dangerous zones of dissection within the surgical field, with high specificity/PPV for Go zones and high sensitivity/NPV for No-Go zones. Overall, model prediction was better for No-Go zones compared to Go zones. This technology may eventually be used to provide real-time guidance and minimize the risk of adverse events.
Asunto(s)
Colecistectomía Laparoscópica , Cirujanos , Humanos , Colecistectomía Laparoscópica/efectos adversos , Inteligencia Artificial , Recolección de Datos , ColecistectomíaRESUMEN
BACKGROUND: Although beta-adrenergic receptor blockade may improve outcomes after traumatic brain injury (TBI), its early use is not routine. We hypothesize that judicious early low-dose propranolol after TBI (EPAT) will improve outcomes without altering bradycardia or hypotensive events. METHODS: We conducted a prospective, observational study on all patients who presented with moderate-to-severe TBI from March 2010-August 2013. Ten initial patients did not receive propranolol (control). Subsequent patients received propranolol at 1-mg intravenous every 6 h starting within 12 h of intensive care unit (ICU) admission (EPAT) for a minimum of 48 h. Heart rate and blood pressure were recorded hourly for the first 72 h. Bradycardia and hypotensive events, mortality, and length of stay (LOS) were compared between cohorts to determine significant differences. RESULTS: Thirty-eight patients were enrolled; 10 control and 28 EPAT. The two cohorts were similar when compared by gender, emergency department (ED) systolic blood pressure, ED heart rate, and mortality. ED Glasgow coma scale was lower (4.2 versus 10.7, P < 0.01) and injury severity score higher in control. EPAT patients received a mean of 10 ± 14 doses of propranolol. Hypotensive events were similar between cohorts, whereas bradycardia events were higher in control (5.8 versus 1.6, P = 0.05). ICU LOS (15.4 versus 30.4 d, P = 0.02) and hospital LOS (10 versus 19.1 d, P = 0.05) were lower in EPAT. Mortality rates were similar between groups (10% versus 10.7%, P = 0.9). The administration of propranolol led to no recorded complications. CONCLUSIONS: Although bradycardia and hypotensive events occur early after TBI, low-dose intravenous propranolol does not increase their number or severity. Early use of propranolol after TBI appears to be safe and may be associated with decreased ICU and hospital LOS.
Asunto(s)
Antagonistas Adrenérgicos beta/uso terapéutico , Lesiones Encefálicas/tratamiento farmacológico , Propranolol/uso terapéutico , Adulto , Anciano , Bradicardia/inducido químicamente , Lesiones Encefálicas/complicaciones , Esquema de Medicación , Femenino , Humanos , Hipotensión/inducido químicamente , Inyecciones Intravenosas , Puntaje de Gravedad del Traumatismo , Tiempo de Internación/estadística & datos numéricos , Masculino , Persona de Mediana Edad , Estudios Prospectivos , Resultado del TratamientoRESUMEN
BACKGROUND: Although it is assumed that obese patients are naturally protected against anterior abdominal stab wounds, the relationship has never been formally studied. We sought to examine the impact of body mass index (BMI) on severity of sustained injury, need for operation, and patient outcomes. MATERIALS AND METHODS: We conducted a review of all patients presenting with abdominal stab wounds at an urban level I trauma center from January 2000-December 2012. Patients were divided into groups based on their BMI (<18.5, 18.5-29.9, 30-35, and >35). Data abstracted included baseline demographics, physiologic data, and characterization of whether the stab wound had violated the peritoneum, caused intra-abdominal injury, or required an operation that was therapeutic. The one-sided Cochran-Armitage trend test was used for significance testing of the protective effect. RESULTS: Of 281 patients with abdominal stab wounds, 249 had complete data for evaluation. Chest and abdomen abbreviated injury scale trends decreased with increasing BMI, as did overall injury severity score, the percent of patients severely injured (injury severity score ≥ 25), and length of intensive care unit stay. Rates of peritoneal violation (100%, 84%, 77%, and 74%; P = 0.077), visceral injury (83%, 56%, 50%, and 30%; P = 0.022), and injury requiring a therapeutic operation (67%, 45%, 40%, and 20%; P = 0.034) all decreased with increasing BMI. Patients in the thinnest group required an operation three times more often than those in the most obese. CONCLUSIONS: Increased BMI protects patients with abdominal stab wounds and is associated with lower incidence of severe injury and need for operation. Heavier patients may be more suitable to observation and serial examinations, whereas very thin patients are more likely to require an operation and be critically injured.
Asunto(s)
Traumatismos Abdominales/cirugía , Índice de Masa Corporal , Puntaje de Gravedad del Traumatismo , Obesidad , Heridas Punzantes/cirugía , Adulto , Anciano , Anciano de 80 o más Años , Femenino , Humanos , Masculino , Persona de Mediana Edad , Resultado del TratamientoRESUMEN
INTRODUCTION: Hemorrhage is assessed, at least in part, via hematocrit testing. To differentiate unexpected drops in hematocrit because of ongoing hemorrhage versus expected drops as a result of known hemorrhage and intravenous fluid administration, we model expected post-operative hematocrit values accounting for fluid balance and intraoperative estimated blood loss (EBL) among patients without substantial post-operative bleeding. MATERIALS AND METHODS: We reviewed patient-level data from the electronic health record of an academic medical center for all non-pregnant adults admitted for elective knee or hip arthroplasty from November 2013 to September 2022 who did not require blood products. We used linear regression to evaluate the association between post-operative hematocrit and predictor variables including pre-operative hematocrit, intraoperative net fluid intake, blood volume, time from surgery to lab testing, EBL, patient height, and patient weight. RESULTS: We included 6,648 cases. Mean (SD) estimated blood volume was 4,804 mL (1023), mean net fluid intake was 1,121 mL (792), and mean EBL was 144 mL (194). Each 100 mL of EBL and 1,000 mL net positive fluid intake was associated with a decrease of 0.52 units (95% CI, 0.51-0.53) and 2.4 units (2.2-2.7) in post-operative hematocrit. Pre-operative hematocrit was the strongest predictor of post-operative hematocrit. Each 1-unit increase in pre-operative hematocrit was associated with a 0.70-unit increase (95% CI, 0.67-0.73) in post-operative hematocrit. Our estimates were robust to sensitivity analyses, and all variables included in the model were statistically significant with P <.005. CONCLUSION: Patient-specific data, including fluid received since the time of initial hemorrhage, can aid in estimating expected post-hemorrhage hematocrit values, and thus in assessing for the ongoing hemorrhage.
RESUMEN
INTRODUCTION: The aim was to leverage electronic health record (EHR) smartphrases to improve compliance with the modified Brain Injury Guidelines (mBIG). METHODS: Smartphrases were developed for the trauma team and radiology and implemented December 2022. Traumatic brain injury (TBI) patients meeting mBIG inclusion from 03/2021- 07/2023 were reviewed. Smartphrase usage and clinical compliance with mBIG (measured by percent reduction of repeat head imaging, ICU admissions, and neurosurgery consults) were compared pre- and post-intervention. RESULTS: 268 cases were examined. Post-intervention, mBIG1 patients had significantly fewer neurosurgery consults (82.4 â% vs. 50.0 â%, OR â= â0.21, p â= â0.03), while mBIG2 patients had significantly fewer repeat head CTs (91.0 â% vs. 66.7 â%, OR â= â0.2, p â= â0.01), ICU admissions (66.7 â% vs. 38.5 â%, OR â= â0.31, p â= â0.02) and neurosurgery consults (93.9 â% vs. 56.4 â%, OR â= â0.08, p â< â0.01). CONCLUSION: Standardized smartphrases can streamline workflow and significantly improve trauma team compliance with best practice guidelines for TBI and reduce unnecessary imaging, consults, and costly ICU admissions.
RESUMEN
BACKGROUND: In adult trauma, mortality varies with race and insurance status. In the elderly, insurance type has little impact on mortality after trauma and the influence of race is reduced. How race and insurance affect pediatric trauma requires further attention. We hypothesized that mortality after pediatric trauma is influenced by insurance type and not race. METHODS: We reviewed all cases of blunt trauma in children ≤13 y requiring admission, using the National Trauma Data Bank Research Data Sets for 2007 and 2008. Exclusions included an Abbreviated Injury Score of 6 for any body region, dead on arrival, and missing data. Our primary outcome measure was in-hospital mortality. RESULTS: We identified 831 Asian (1.2%), 10,592 black (15.5%), 45,173 white (66.2%), and 8498 Hispanic (12.5%) children, and 3161 children (4.6%) classified as other race. Mean age was 7.4 ± 4.5 y, 11.9% were uninsured, and overall in-hospital mortality was 1.4%. Multivariable modeling indicated that race was not associated with increased mortality (Asian versus white, adjusted odds ratio [AOR] 1.05, P = 0.88; black versus white, AOR 0.92, P = 0.42; Hispanic versus white, AOR 0.87, P = 0.26; and other race versus white, AOR 1.01, P = 0.96). In contrast, insurance status (any insurance versus no insurance, AOR 0.6, P < 0.01) and insurance type (private insurance versus no insurance, AOR 0.47, P < 0.01; Medicaid versus no insurance, 0.67, P < 0.01) predicted reduced mortality. CONCLUSIONS: Insurance status and insurance type are important predictors of mortality after pediatric trauma while, in contrast, race is not.
Asunto(s)
Seguro de Salud/estadística & datos numéricos , Medicaid/estadística & datos numéricos , Grupos Raciales/estadística & datos numéricos , Heridas y Lesiones/etnología , Heridas y Lesiones/mortalidad , Asiático/estadística & datos numéricos , Población Negra/estadística & datos numéricos , Niño , Preescolar , Bases de Datos Factuales/estadística & datos numéricos , Femenino , Hispánicos o Latinos/estadística & datos numéricos , Mortalidad Hospitalaria , Humanos , Masculino , Análisis Multivariante , Valor Predictivo de las Pruebas , Estudios Retrospectivos , Estados Unidos/epidemiología , Población Blanca/estadística & datos numéricosRESUMEN
The rodent septohippocampal system contains "theta cells," which burst rhythmically at 4-12 Hz, but the functional significance of this rhythm remains poorly understood (Buzsáki, 2006). Theta rhythm commonly modulates the spike trains of spatially tuned neurons such as place (O'Keefe and Dostrovsky, 1971), head direction (Tsanov et al., 2011a), grid (Hafting et al., 2005), and border cells (Savelli et al., 2008; Solstad et al., 2008). An "oscillatory interference" theory has hypothesized that some of these spatially tuned neurons may derive their positional firing from phase interference among theta oscillations with frequencies that are modulated by the speed and direction of translational movements (Burgess et al., 2005, 2007). This theory is supported by studies reporting modulation of theta frequency by movement speed (Rivas et al., 1996; Geisler et al., 2007; Jeewajee et al., 2008a), but modulation of theta frequency by movement direction has never been observed. Here we recorded theta cells from hippocampus, medial septum, and anterior thalamus of freely behaving rats. Theta cell burst frequencies varied as the cosine of the rat's movement direction, and this directional tuning was influenced by landmark cues, in agreement with predictions of the oscillatory interference theory. Computer simulations and mathematical analysis demonstrated how a postsynaptic neuron can detect location-dependent synchrony among inputs from such theta cells, and thereby mimic the spatial tuning properties of place, grid, or border cells. These results suggest that theta cells may serve a high-level computational function by encoding a basis set of oscillatory signals that interfere with one another to synthesize spatial memory representations.
Asunto(s)
Potenciales de Acción/fisiología , Encéfalo/citología , Neuronas/fisiología , Orientación , Percepción Espacial/fisiología , Ritmo Teta/fisiología , Animales , Biofisica , Encéfalo/fisiología , Simulación por Computador , Señales (Psicología) , Conducta Exploratoria , Masculino , Modelos Neurológicos , Movimiento/fisiología , Ratas , Ratas Long-Evans , Grabación en Video/métodos , VigiliaRESUMEN
Gastric pneumatosis, the presence of air within the stomach wall, is a very rare occurrence with poor outcomes. One of the most common mechanisms for gastric pneumatosis is gastric ischaemia, also a rare entity. Although patients with gastric ischaemia may require surgical intervention, they can often be treated with conservative measures such as a proton pump inhibitor, broad-spectrum antibiotics, nasogastric tube decompression, fluid resuscitation and total parenteral nutrition. We report a rare case of gastric ischaemia and pneumatosis following therapeutic left gastric artery argon plasma coagulation that was treated with conservative measures.
Asunto(s)
Tratamiento Conservador , Artería Gástrica , Humanos , Intubación Gastrointestinal/efectos adversos , Isquemia , Estómago/irrigación sanguíneaRESUMEN
Sodium glucose cotransporter-2 (SGLT2) inhibitors are glucose-lowering drugs with proven efficacy in treating type 2 diabetes mellitus, and more recently, have been shown to improve heart failure outcomes in patients without diabetes. A rare complication of SGLT2 inhibitor use is the development of euglycaemic diabetic ketoacidosis (EDKA), characterised by euglycaemia (blood glucose level <250 mg/dL), metabolic acidosis (arterial pH <7.3 and serum bicarbonate <18 mEq/L), and ketonaemia. Given patients with EDKA do not present with the typical manifestations of diabetic ketoacidosis, including marked hyperglycaemia and dehydration, the diagnosis of EDKA may be missed and initiation of treatment delayed. We present the case of a man with recent SGLT2 inhibitor use and multiple other risk factors who developed EDKA.
Asunto(s)
Diabetes Mellitus Tipo 2 , Cetoacidosis Diabética , Pancreatitis , Inhibidores del Cotransportador de Sodio-Glucosa 2 , Compuestos de Bencidrilo , Diabetes Mellitus Tipo 2/complicaciones , Diabetes Mellitus Tipo 2/tratamiento farmacológico , Cetoacidosis Diabética/inducido químicamente , Cetoacidosis Diabética/diagnóstico , Cetoacidosis Diabética/tratamiento farmacológico , Glucósidos , Humanos , Masculino , Pancreatitis/complicaciones , Inhibidores del Cotransportador de Sodio-Glucosa 2/efectos adversosRESUMEN
BACKGROUND: Donor liver biopsy (DLBx) in liver transplantation provides information on allograft quality; however, predicting outcomes from these allografts remains difficult. METHODS: Between 2006 and 2015, 16 691 transplants with DLBx were identified from the Standard Transplant Analysis and Research database. Cox proportional hazard regression analyses identified donor and recipient characteristics associated with 30-d, 90-d, 1-y, and 3-y graft survival. A composite model, the Liver Transplant After Biopsy (LTAB) score, was created. The Mini-LTAB was then derived consisting of only donor age, macrosteatosis on DLBx, recipient model for end-stage liver disease score, and cold ischemic time. Risk groups were identified for each score and graft survival was evaluated. P values <0.05 were considered significant. RESULTS: The LTAB model used 14 variables and 5 risk groups and identified low-, mild-, moderate-, high-, and severe-risk groups. Compared with moderate-risk recipients, severe-risk recipients had increased risk of graft loss at 30 d (hazard ratio, 3.270; 95% confidence interval, 2.568-4.120) and at 1 y (2.258; 1.928-2.544). The Mini-LTAB model identified low-, moderate-, and high-risk groups. Graft survival in Mini-LTAB high-risk transplants was significantly lower than moderate- or low-risk transplants at all time points. CONCLUSIONS: The LTAB and Mini-LTAB scores represent guiding principles and provide clinically useful tools for the successful selection and utilization of marginal allografts in liver transplantation.
RESUMEN
Background: Cleft lip/palate (CLP) is a congenital orofacial anomaly appearing in approximately one in 700 births worldwide. While in high-income countries CLP is normally addressed surgically during infancy, in developing countries CLP is often left unoperated, potentially impacting multiple dimensions of life quality. Previous research has frequently compared CLP outcomes to those of the general population. But because local environmental and genetic factors contribute to the risk of CLP and also may influence life outcomes, such studies may downwardly bias estimates of both CLP status and correction. Objectives: This research represents the first study to use causal econometric methods to estimate the effects of both CLP status and CLP correction on the physical, social, and mental well-being of children. Methods: Data were collected first-hand from 1,118 Indian children, where we obtained first-hand data on height, weight, grip strength, cognitive ability, reading, and math ability. A professional speech therapist reviewed digital recordings of speech taken at the interview to obtain four measures of speech quality. Using this data, the household fixed-effects model we employ jointly estimates effects of CLP status and CLP surgical intervention. Findings: Our results indicate that adolescents with median-level CLP severity show statistically significant losses in indices of speech quality (-1.59σ), physical well-being (0.32σ), academic and cognitive ability (-0.37σ), and social integration (-0.32σ). We find strong evidence that CLP surgery significantly restores speech if performed before five years of age. The first surgeries performed on less-severe CLP cases significantly restore social integration, psychological well-being, academic/cognitive ability, and a general index of human flourishing. Conclusions: Children born with CLP in India face statistically significant losses in speech, physical health, mental health, and social inclusion. CLP surgical intervention significantly restores speech quality if carried out at an early age. Surgeries with the most significant impact on life outcomes are the first surgeries performed on less-severe CLP cases.
Asunto(s)
Labio Leporino , Fisura del Paladar , Enfermedades Musculoesqueléticas , Adolescente , Niño , Labio Leporino/cirugía , Fisura del Paladar/cirugía , Cara , Humanos , Calidad de VidaRESUMEN
PURPOSE: Recent literature suggests an increasing use of systemic treatment in patients with advanced cancer near the end of life (EOL), partially driven by the increasing adoption of immune checkpoint inhibitors (ICIs). While studies have identified this trend, additional variables associated with ICI use at EOL are limited. Our aim was to characterize a population of patients who received a dose of ICI in the last 30 days of life. METHODS: We performed a manual retrospective chart review of patients ≥ 18 years who died within 30 days of receiving a dose of ICI. Metrics such as Eastern Cooperative Oncology Group performance status (ECOG PS), number of ICI doses, need for hospitalization, and numerous other variables were evaluated. RESULTS: Over a 4-year time period, 97 patients received an ICI at EOL. For 40% of patients, the ICI given in the 30 days before death was their only dose. Over 50% of patients had an ECOG PS of ≥ 2, including 17% of patients with an ECOG PS of 3. Over 60% were hospitalized, 65% visited the emergency department, 20% required intensive care unit admission, and 25% died in the hospital. CONCLUSION: Our study contributes to the ongoing literature regarding the risks and benefits of ICI use in patients with advanced cancer near the EOL. While accurate predictions regarding the EOL are challenging, oncologists may routinely use clinical factors such as ECOG PS along with patient preferences to guide recommendations and shared decision making. Ultimately, further follow-up studies to better characterize and prognosticate this population of patients are needed.
Asunto(s)
Inhibidores de Puntos de Control Inmunológico , Neoplasias , Muerte , Humanos , Inhibidores de Puntos de Control Inmunológico/farmacología , Inhibidores de Puntos de Control Inmunológico/uso terapéutico , Neoplasias/terapia , Prioridad del Paciente , Estudios RetrospectivosRESUMEN
BACKGROUND: This study investigates the impact of standing electric scooter-related injuries within an entire integrated hospital system. METHODS: We performed a retrospective review of patients involved in standing electric scooter incidents presenting throughout an urban hospital network over a 10 month period. Rates of Google searches of scooter-related terms performed locally were used as a surrogate for ride frequency. Injury, mechanism, and cost data were analyzed. RESULTS: Data on 248 patients were reviewed. Twenty-three (9%) were under 18 years old. Loss of balance was the most common cause of injury accounting for nearly half, while tripping over a scooter 14 (6%) affected the elderly disproportionately. Eight (3%) riders wore helmets. All TBI and closed head injuries occurred in unhelmeted patients. Most incidents occurred in the street, only one in a bicycle lane. Facilities costs were greater for patients under the influence of alcohol and marijuana. CONCLUSION: Policies related to the use of mandated safety equipment, dedicated bicycle lanes, and the proper storage of empty vehicles should be further investigated.
Asunto(s)
Lesiones Accidentales/epidemiología , Vehículos a Motor Todoterreno , Adolescente , Adulto , Anciano , Anciano de 80 o más Años , Niño , Preescolar , Femenino , Humanos , Masculino , Persona de Mediana Edad , Estudios Retrospectivos , Estados Unidos , Adulto JovenRESUMEN
BACKGROUND: The Warkentin 4-T scoring system for determining the pretest probability of heparin-induced thrombocytopenia (HIT) has been shown to be inaccurate in the ICU and does not take into account body mass index (BMI). METHODS: Prospectively collected data on patients in the surgical and cardiac ICU between January 2007 and February 2016 who were presumed to have HIT by clinical suspicion were reviewed. Patients were categorized into 3 BMI groups and assigned scores: Normal weight, overweight, and obese. Multivariate analyses were used to identify independent predictors of HIT. RESULTS: A total of 523 patients met inclusion criteria. Multivariate analysis showed that only BMI, Timing, and oTher variables were independently associated with HIT. This new 3-T model was better than a five-component model consisting of the entire 4-T scoring system plus BMI (AUCâ¯=â¯0.791). CONCLUSIONS: Incorporating patient 'T'hickness into a pretest probability model along with platelet 'T'iming and the exclusion of o'T'her causes of thrombocytopenia yields a simplified "3-T" scoring system that has increased predictive accuracy in the ICU.
Asunto(s)
Índice de Masa Corporal , Fibrinolíticos/efectos adversos , Heparina/efectos adversos , Modelos Teóricos , Trombocitopenia/inducido químicamente , Adulto , Anciano , Femenino , Predicción , Humanos , Unidades de Cuidados Intensivos , Masculino , Persona de Mediana Edad , Estudios RetrospectivosRESUMEN
Background and study aims Endoscopic ultrasound (EUS) has been used for portal vein sampling in patients with pancreaticobiliary cancers for enumerating circulating tumor cells but is not yet a standard procedure. Further evaluation is needed to refine the methodology. Therefore, we evaluated the feasibility and safety of 19-gauge (19G) versus a 22-gauge (22âG) EUS fine-needle aspiration needles for portal vein sampling in a swine model. Methods Celiotomy was performed on two farm pigs. Portal vein sampling occurred transhepatically. We compared 19âG and 22âG needles coated interiorly with saline, heparin or ethylenediaminetetraacetic acid (EDTA). Small- (10âmL) and large- (25âmL) volume blood collections were evaluated. Two different collection methods were tested: direct-to-vial and suction syringe. A bleeding risk trial for saline-coated 19âG and 22âG needles was performed by puncturing the portal vein 20 times. Persistent bleeding after 3 minutes was considered significant. Results All small-volume collection trials were successful except for 22âG saline-coated needles with direct-to-vial method. All large-volume collection trials were successful when using suction syringe; direct-to-vial method for both 19âG and 22âG needles were unsuccessful. Collection times were shorter for 19âG vs. 22âG needles for both small and large-volume collections ( P â<â0.05). Collection times for saline-coated 22âG needles were longer compared to heparin/EDTA-coated ( P â<â0.05). Bleeding occurred in 10â% punctures with 19âG needles compared to 0â% with 22âG needles. Conclusion The results of this animal study demonstrate the feasibility and the safety of using 22âG needles for portal vein sampling and can form the basis for a pilot study in patients.
RESUMEN
BACKGROUND: Historically, liver allografts with >30% macrosteatosis (MaS) on donor biopsy have been associated with early allograft dysfunction and worse graft survival; however, successful outcomes have been reported in small cohorts. This study proposes an elevated MaS threshold for organ utilization without detriment to graft survival. METHODS: The UNOS Standard Transplant Analysis and Research database was evaluated for transplants between 2006-2015. Graft survival up to 1-year was evaluated by Kaplan-Meier (KM) survival analyses, and by univariate and multivariable logistic regression analyses, including donor and recipient characteristics. Odds ratios (OR) with 95% confidence intervals (CI) for risk of graft loss are reported. RESULTS: Thirty-day risk of graft loss was increased with MaS as low as 10-19% (OR [95% CI] 1.301 [1.055-1.605], p<0.0001) and peaked with MaS 50-59% (2.921 [1.672-5.103]). At 1-year, risk of graft loss remained elevated with MaS 40-49% (1.465 [1.002-2.142]) and MaS 50-59% (1.978 [1.281-3.056], p = 0.0224). Multivariable models were created for Lower and Higher MELD recipients and MaS cutoffs were established. In Lower MELD recipients, organs with ≥50% MaS had increased risk of graft loss at 30 days (2.451 [1.541-3.897], p = 0.0008) and 1-year post-transplant (1.720 [1.224-2.418], p = 0.0125). Higher MELD recipients had increased risk of graft loss at 30 days with allografts showing MaS ≥40% (4.204 [1.440-5.076], p = 0.0016). At 1-year the risk remained increased, but MaS was not significant predictor of graft loss.048 [1.131-3.710], p = 0.0616). In both MELD cohorts, organs with MaS levels below threshold had similar survival to those transplanted without a donor biopsy. CONCLUSIONS: In conjunction with recipient selection, organs with MaS up to 50% may be safely used without detriment to outcomes.
Asunto(s)
Aloinjertos/cirugía , Supervivencia de Injerto/fisiología , Trasplante de Hígado/mortalidad , Adulto , Bases de Datos Factuales , Selección de Donante/métodos , Femenino , Humanos , Estimación de Kaplan-Meier , Masculino , Persona de Mediana Edad , Oportunidad Relativa , Estudios Retrospectivos , Factores de Riesgo , Donantes de Tejidos , Trasplante Homólogo/mortalidad , Estados Unidos , Adulto JovenRESUMEN
BACKGROUND: Genomic classifiers (GC) have been shown to improve risk stratification post prostatectomy. However, their clinical benefit has not been prospectively demonstrated. We sought to determine the impact of GC testing on postoperative management in men with prostate cancer post prostatectomy. METHODS: Two prospective registries of prostate cancer patients treated between 2014 and 2019 were included. All men underwent Decipher tumor testing for adverse features post prostatectomy (Decipher Biosciences, San Diego, CA). The clinical utility cohort, which measured the change in treatment decision-making, captured pre- and postgenomic treatment recommendations from urologists across diverse practice settings (n = 3455). The clinical benefit cohort, which examined the difference in outcome, was from a single academic institution whose tumor board predefined "best practices" based on GC results (n = 135). RESULTS: In the clinical utility cohort, providers' recommendations pregenomic testing were primarily observation (69%). GC testing changed recommendations for 39% of patients, translating to a number needed to test of 3 to change one treatment decision. In the clinical benefit cohort, 61% of patients had genomic high-risk tumors; those who received the recommended adjuvant radiation therapy (ART) had 2-year PSA recurrence of 3 vs. 25% for those who did not (HR 0.1 [95% CI 0.0-0.6], p = 0.013). For the genomic low/intermediate-risk patients, 93% followed recommendations for observation, with similar 2-year PSA recurrence rates compared with those who received ART (p = 0.93). CONCLUSIONS: The use of GC substantially altered treatment decision-making, with a number needed to test of only 3. Implementing best practices to routinely recommend ART for genomic-high patients led to larger than expected improvements in early biochemical endpoints, without jeopardizing outcomes for genomic-low/intermediate-risk patients.
Asunto(s)
Biomarcadores de Tumor/genética , Toma de Decisiones , Selección de Paciente , Prostatectomía/métodos , Neoplasias de la Próstata/genética , Neoplasias de la Próstata/terapia , Medición de Riesgo/métodos , Adulto , Anciano , Anciano de 80 o más Años , Algoritmos , Estudios de Seguimiento , Perfilación de la Expresión Génica , Genómica , Humanos , Masculino , Persona de Mediana Edad , Pronóstico , Neoplasias de la Próstata/clasificación , Neoplasias de la Próstata/patología , Tasa de SupervivenciaRESUMEN
BACKGROUND: Obesity, defined as a high body mass index (hBMI) of 30 kg/m or greater, is a growing epidemic worldwide and is associated with multiple comorbidities. High BMI individuals account for an increasing portion of potential liver donors. Here we evaluate trends in the utilization and outcomes of hBMI donors on a national and regional level and the potential role of liver biopsy in donor evaluation. METHODS: United Network for Organ Sharing Standard Transplant Analysis and Research database was evaluated for deceased donor liver transplants between 2006 and 2016 across 11 Organ Procurement and Transplantation Network regions. High BMI donors were compared with lower BMI counterparts and evaluated for biopsy rates, utilization rates and allograft outcomes. Univariate and multivariable analyses were performed. RESULTS: Seventy-seven thousand fifty potential donors were identified and 60 200 transplants were evaluated. Utilization rates for hBMI donors were 66.1% versus 78.1% for lower BMI donors (P < 0.001). Pretransplant biopsy was performed more frequently in hBMI donors (52.1% vs 33.1%, P < 0.001) and macrosteatosis of 30% or greater was identified more often (21.1% vs 12.2%, P < 0.001). Biopsy performance increased utilization rate of hBMI donors in 7 of 11 Organ Procurement and Transplantation Network regions. region 6 showed the highest rate of biopsy performance, high rate of hBMI donor utilization, and highest 5-year estimated graft survival rates of all regions. CONCLUSIONS: High BMI donors have not previously been associated with worse graft survival in multivariable analyses; however, they are used much less frequently. Liver biopsy may increase the utilization rate of hBMI donors and improve donor selection. Further evaluation of regions with high rates of utilization and good outcomes is warranted.