Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 693
Filter
1.
BMC Med Ethics ; 25(1): 107, 2024 Oct 07.
Article in English | MEDLINE | ID: mdl-39375660

ABSTRACT

BACKGROUND: Artificial intelligence-driven Clinical Decision Support Systems (AI-CDSS) are being increasingly introduced into various domains of health care for diagnostic, prognostic, therapeutic and other purposes. A significant part of the discourse on ethically appropriate conditions relate to the levels of understanding and explicability needed for ensuring responsible clinical decision-making when using AI-CDSS. Empirical evidence on stakeholders' viewpoints on these issues is scarce so far. The present study complements the empirical-ethical body of research by, on the one hand, investigating the requirements for understanding and explicability in depth with regard to the rationale behind them. On the other hand, it surveys medical students at the end of their studies as stakeholders, of whom little data is available so far, but for whom AI-CDSS will be an important part of their medical practice. METHODS: Fifteen semi-structured qualitative interviews (each lasting an average of 56 min) were conducted with German medical students to investigate their perspectives and attitudes on the use of AI-CDSS. The problem-centred interviews draw on two hypothetical case vignettes of AI-CDSS employed in nephrology and surgery. Interviewees' perceptions and convictions of their own clinical role and responsibilities in dealing with AI-CDSS were elicited as well as viewpoints on explicability as well as the necessary level of understanding and competencies needed on the clinicians' side. The qualitative data were analysed according to key principles of qualitative content analysis (Kuckartz). RESULTS: In response to the central question about the necessary understanding of AI-CDSS tools and the emergence of their outputs as well as the reasons for the requirements placed on them, two types of argumentation could be differentiated inductively from the interviewees' statements: the first type, the clinician as a systemic trustee (or "the one relying"), highlights that there needs to be empirical evidence and adequate approval processes that guarantee minimised harm and a clinical benefit from the employment of an AI-CDSS. Based on proof of these requirements, the use of an AI-CDSS would be appropriate, as according to "the one relying", clinicians should choose those measures that statistically cause the least harm. The second type, the clinician as an individual expert (or "the one controlling"), sets higher prerequisites that go beyond ensuring empirical evidence and adequate approval processes. These higher prerequisites relate to the clinician's necessary level of competence and understanding of how a specific AI-CDSS works and how to use it properly in order to evaluate its outputs and to mitigate potential risks for the individual patient. Both types are unified in their high esteem of evidence-based clinical practice and the need to communicate with the patient on the use of medical AI. However, the interviewees' different conceptions of the clinician's role and responsibilities cause them to have different requirements regarding the clinician's understanding and explicability of an AI-CDSS beyond the proof of benefit. CONCLUSIONS: The study results highlight two different types among (future) clinicians regarding their view of the necessary levels of understanding and competence. These findings should inform the debate on appropriate training programmes and professional standards (e.g. clinical practice guidelines) that enable the safe and effective clinical employment of AI-CDSS in various clinical fields. While current approaches search for appropriate minimum requirements of the necessary understanding and competence, the differences between (future) clinicians in terms of their information and understanding needs described here can lead to more differentiated approaches to solutions.


Subject(s)
Artificial Intelligence , Decision Support Systems, Clinical , Qualitative Research , Students, Medical , Humans , Artificial Intelligence/ethics , Students, Medical/psychology , Germany , Female , Male , Attitude of Health Personnel , Clinical Decision-Making/ethics , Physician's Role , Adult , Interviews as Topic
2.
JAMIA Open ; 7(4): ooae102, 2024 Dec.
Article in English | MEDLINE | ID: mdl-39386064

ABSTRACT

Objective: This study investigates the concordance of patient information collected using a medical history app compared to in-person interviews. Materials and Methods: In this cross-sectional study we used an app to collect medical data from patients in family practice in Germany. Collected information included age, height, weight, perceived severity of complaints, and 38 current complaints. Subsequently, in-person interviews based on the query structure of the app were conducted with patients directly after the patient finished filling out the app. Concordance was assessed as exact matches between the data collected app-based and in-person interviews, with the in-person interview as a reference. Regression analysis examined which patient characteristics were associated with mismatching and underreporting of complaints. Results: Three hundred ninety-nine patients were included in the study. Concordance of reported age, weight, and height, as well as perceived severity of complaints ranged from 76.2% to 96.7%. Across all 38 complaints, 64.4% of participants showed completely identical complaint selection in app-based and in-person interviews; 18.5% of all participants overreported; and 17.0% underreported at least 1 complaint when using the app. Male sex, higher age, and higher number of stated complaints were associated with higher odds of underreporting at least one complaint in the app. Discussion: App-collected data regarding age, weight, height, and perceived severity of complaints showed high concordance. The discordance shown concerning various complaints should be examined regarding their potential for medical errors. Conclusion: The introduction of apps for gathering information on complaints can improve the efficiency and quality of care but must first be improved. Trial registration: The study was registered at the German Clinical Trials Register No. DRKS00026659 registered November 3, 2021. World Health Organization Trial Registration Data Set, https://trialsearch.who.int/Trial2.aspx?TrialID=DRKS00026659.

3.
Int J Nurs Stud ; 161: 104918, 2024 Sep 27.
Article in English | MEDLINE | ID: mdl-39388847

ABSTRACT

BACKGROUND: Assessment of signs and symptoms in hospitalized children presents unique challenges due to the children's age-related differences, such as vital signs and the broad range of medical conditions that affects children. Early detection of clinical changes in children is crucial to prevent deterioration, and while standardized tools exist, there is a growing recognition of the need to consider subjective factors based on experienced nurses' knowledge and intuition. OBJECTIVE: To explore which signs and symptoms, apart from vital signs, that trigger nurses' concern regarding deterioration of hospitalized children and adolescents. DESIGN: This study used a descriptive qualitative design. SETTINGS: The study was conducted at three pediatric departments in Denmark and a nursing department of a university in Norway, offering post graduate education programs for health care professions working with children and adolescents throughout Norway. PARTICIPANTS: A total sample of 29 registered nurses with varying levels of experience participated. METHOD: Four focus group interviews were used to collect data and analyzed with inductive content analysis approach. RESULTS: Nurses' knowledge about children's clinical conditions is influenced by the nurses experience, their use of senses like touching the child with their hands, and the use of various approaches. Information from parents about the child's normal behavior are considered valuable. These sources of information, often difficult to verbalize, might be referred to as intuition or "gut feeling" and often guides the nurses' actions when vital signs appear normal, and nurses rely on their senses to assess the child's condition. Specific indicators triggering concern include changes in respiration, circulation, level of consciousness, and facial expressions. Challenges arise from nighttime assessments, interactions with parents, the presence of electronic devices, and children's ability to compensate. Clinical experience is a significant factor in nurses' ability to recognize changes in in the child's condition. CONCLUSION: This study highlights the multifaceted nature of nurses' assessments of clinical conditions in hospitalized children. Nurses draw on their experiences, intuition, and interactions with parents to complement vital signs-based assessments. Their intuition, or "gut feeling" serves as a valuable tool when vital signs do not fully capture the child's clinical status. Specific signs and symptoms that trigger nurses' concern, along with the challenges they face, contribute to a comprehensive understanding of the complexity of assessing children's clinical conditions. These findings, emphasize the role of nurses in early recognition of clinical deterioration in hospitalized children and the need for assessments that go beyond vital signs. TWEETABLE ABSTRACT: Both objective assessments and intuitive clinical judgment play an important role in identifying potential deterioration in pediatric patients.

4.
Article in English | MEDLINE | ID: mdl-39401245

ABSTRACT

OBJECTIVES: Successful implementation of machine learning-augmented clinical decision support systems (ML-CDSS) in perioperative care requires the prioritization of patient-centric approaches to ensure alignment with societal expectations. We assessed general public and surgical patient attitudes and perspectives on ML-CDSS use in perioperative care. MATERIALS AND METHODS: A sequential explanatory study was conducted. Stage 1 collected public opinions through a survey. Stage 2 ascertained surgical patients' experiences and attitudes via focus groups and interviews. RESULTS: For Stage 1, a total of 281 respondents' (140 males [49.8%]) data were considered. Among participants without ML awareness, males were almost three times more likely than females to report more acceptance (OR = 2.97; 95% CI, 1.36-6.49) and embrace (OR = 2.74; 95% CI, 1.23-6.09) of ML-CDSS use by perioperative teams. Males were almost twice as likely as females to report more acceptance across all perioperative phases with ORs ranging from 1.71 to 2.07. In Stage 2, insights from 10 surgical patients revealed unanimous agreement that ML-CDSS should primarily serve a supportive function. The pre- and post-operative phases were identified explicitly as forums where ML-CDSS can enhance care delivery. Patients requested for education on ML-CDSS's role in their care to be disseminated by surgeons across multiple platforms. DISCUSSION AND CONCLUSION: The general public and surgical patients are receptive to ML-CDSS use throughout their perioperative care provided its role is auxiliary to perioperative teams. However, the integration of ML-CDSS into perioperative workflows presents unique challenges for healthcare settings. Insights from this study can inform strategies to support large-scale implementation and adoption of ML-CDSS by patients in all perioperative phases. Key strategies to promote the feasibility and acceptability of ML-CDSS include clinician-led discussions about ML-CDSS's role in perioperative care, established metrics to evaluate the clinical utility of ML-CDSS, and patient education.

5.
JMIR Hum Factors ; 11: e56949, 2024 Oct 15.
Article in English | MEDLINE | ID: mdl-39405513

ABSTRACT

BACKGROUND: Sepsis is a common cause of serious illness and death. Sepsis management remains challenging and suboptimal. To support rapid sepsis diagnosis and treatment, screening tools have been embedded into hospital digital systems to appear as digital alerts. The implementation of digital alerts to improve the management of sepsis and deterioration is a complex intervention that has to fit with team workflow and the views and practices of hospital staff. Despite the importance of human decision-making and behavior in optimal implementation, there are limited qualitative studies that explore the views and experiences of health care professionals regarding digital alerts as sepsis or deterioration computerized clinician decision support systems (CCDSSs). OBJECTIVE: This study aims to explore the views and experiences of health care professionals on the use of sepsis or deterioration CCDSSs and to identify barriers and facilitators to their implementation and use in National Health Service (NHS) hospitals. METHODS: We conducted a qualitative, multisite study with unstructured observations and semistructured interviews with health care professionals from emergency departments, outreach teams, and intensive or acute units in 3 NHS hospital trusts in England. Data from both interviews and observations were analyzed together inductively using thematic analysis. RESULTS: A total of 22 health care professionals were interviewed, and 12 observation sessions were undertaken. A total of four themes regarding digital alerts were identified: (1) support decision-making as nested in electronic health records, but never substitute professionals' knowledge and experience; (2) remind to take action according to the context, such as the hospital unit and the job role; (3) improve the alerts and their introduction, by making them more accessible, easy to use, not intrusive, more accurate, as well as integrated across the whole health care system; and (4) contextual factors affecting views and use of alerts in the NHS trusts. Digital alerts are more optimally used in general hospital units with a lower senior decision maker:patient ratio and by health care professionals with experience of a similar technology. Better use of the alerts was associated with quality improvement initiatives and continuous sepsis training. The trusts' features, such as the presence of a 24/7 emergency outreach team, good technological resources, and staffing and teamwork, favored a more optimal use. CONCLUSIONS: Trust implementation of sepsis or deterioration CCDSSs requires support on multiple levels and at all phases of the intervention, starting from a prego-live analysis addressing organizational needs and readiness. Advancements toward minimally disruptive and smart digital alerts as sepsis or deterioration CCDSSs, which are more accurate and specific but at the same time scalable and accessible, require policy changes and investments in multidisciplinary research.


Subject(s)
Decision Support Systems, Clinical , Health Personnel , Qualitative Research , Sepsis , State Medicine , Humans , Sepsis/therapy , Sepsis/diagnosis , England , Attitude of Health Personnel
6.
JAMIA Open ; 7(4): ooae092, 2024 Dec.
Article in English | MEDLINE | ID: mdl-39415945

ABSTRACT

Objectives: Conducting simulation testing with end-users is essential for facilitating successful implementation of new health information technologies. This study designed a standardized simulation testing process with a system prototype prior to implementation to help study teams identify the system's interpretability and feasibility from the end-user perspective and to effectively integrate new innovations into real-world clinical settings and workflows. Materials and Methods: A clinical simulation model was developed to test a new Clinical Decision Support (CDS) system outside of the clinical environment while maintaining high fidelity. A web-based CDS prototype, the "CONCERN Smart Application," which leverages clinical data to measure and express a patient's risk of deterioration on a 3-level scale ("low," "moderate," or "high"), and audiovisual-integrated materials, were used to lead simulation sessions. Results: A total of 6 simulation sessions with 17 nurses were held to investigate how nurses interact with the CONCERN Smart application and how it influences their critical thinking, and clinical responses. Four themes were extracted from the simulation debriefing sessions and used to inform implementation strategies. The strategies include how the CDS should be improved for practical real-world use. Discussion and Conclusions: Standardized simulation testing procedures identified and informed the necessary CDS improvements, the enhancements needed for real-world use, and the training requirements to effectively prepare end-users for system go-live.

7.
Ther Adv Drug Saf ; 15: 20420986241272846, 2024.
Article in English | MEDLINE | ID: mdl-39421007

ABSTRACT

Background: Polypharmacy and potentially inappropriate medications are significant challenges in older adults' medication management. The Consolidated Framework for Implementation Research (CFIR) is a comprehensive approach used to explore barriers and enablers to the healthcare system in guiding the effective implementation of evidence-based practices. Objectives: This study examines the barriers and enablers to promote safe medication management among older adults in Qatar from healthcare professionals' perspectives. This includes identifying critical factors within the healthcare system influencing medication management and suggesting practical solutions to improve it. Design: The study employs a qualitative design. Focus Groups (FGs) were conducted with healthcare professionals from the geriatric, mental health and medicine departments of Hamad Medical Corporation (HMC), the leading governmental sector in Qatar serving the older adult population. Methods: Utilising the CFIR, this study analysed feedback from healthcare professionals through FGs at HMC. A combined inductive and deductive thematic analysis was applied to transcripts from five FGs, focusing on identifying barriers and enablers to safe medication management among older adults. Two researchers transcribed the audio-recorded FG discussions verbatim, and two researchers analysed the data using a mixed inductive and deductive thematic analysis approach utilising CFIR constructs. Results: We engaged 53 healthcare professionals (31 physicians, 10 nurses and 12 clinical pharmacists) in FGs. The analysis identified current barriers and enabler themes under different CFIR constructs, including inner settings, outer settings, individual characteristics and intervention characteristics. We identified 44 themes, with 25 classifieds as barriers and 19 as enablers. The findings revealed that barriers and enablers within the inner settings were primarily related to structural characteristics, resources, policies, communication and culture. On the other hand, barriers and enablers from the outer settings included patients and caregivers, care coordination, policies and laws, and resources. Conclusion: This study identified several barriers and enablers to promote medication management for older adults using the CFIR constructs from the perspective of healthcare professionals. The multifaceted findings emphasise involving stakeholders like clinical leaders, policymakers and decision-makers to address medication safety factors. A robust action plan, continuously monitored under Qatar's national strategy, is vital. Further research is needed to implement recommended interventions.


Medication management challenges and solutions for older adults in Qatar: insights from healthcare professionals As people age, they often need multiple medications to manage their health conditions. However, taking medications that are not needed can cause harm. To improve medication management in this vulnerable population, it is essential to understand the barriers and enablers that healthcare professionals (HCPs) face. Our study used focus groups to explore these factors from the perspectives of healthcare providers in Qatar's Hamad Medical Corporation (HMC). We used the Consolidated Framework for Implementation Research (CFIR) to collect and analyse the data. Healthcare Professionals emphasised that the significant barriers to safe medication management in older adults include: • The missing medication history in electronic health records in many cases. • There is a lack of clinical decision support systems that guide and save prescribers time. • There is limited access to services such as medication therapy management and telemedicine. These services could facilitate managing complex or urgent cases. • Sometimes, communication between healthcare providers, patients, and caregivers is inadequate. It could be due to limited clinic time, HCPs' experience, or patients' health literacy. • There are unclear guidelines and policies regarding prescribing, dispensing, and stopping medications for older adults. • There is insufficient education for sub-specialists, junior HCPs, patients, and caregivers about the challenges of managing older adults' medications. • Limited patient engagement in their medication management plans could be due to low health literacy, social support, or physical or cognitive disabilities. • In addition to overcoming the previous challenges, HCPs suggested implementing a national strategy to utilise, guide, and monitor all the efforts. In conclusion, through our study, HCPs highlight the need for tailored national interventions to optimise safe medication management in older adults. The findings can inform the need for developing long-term and comprehensive strategies to help healthcare systems manage older adults' medications, leading to better health outcomes for this vulnerable population.

8.
Integr Pharm Res Pract ; 13: 139-153, 2024.
Article in English | MEDLINE | ID: mdl-39220215

ABSTRACT

The field of healthcare is experiencing a significant transformation driven by technological advancements, scientific breakthroughs, and a focus on personalized patient care. At the forefront of this evolution is artificial intelligence-driven pharmacy practice (IDPP), which integrates data science and technology to enhance pharmacists' capabilities. This prospective article introduces the concept of "pharmacointelligence", a paradigm shift that synergizes artificial intelligence (AI), data integration, clinical decision support systems (CDSS), and pharmacy informatics to optimize medication-related processes. Through a comprehensive literature review and analysis, this research highlights the potential of pharmacointelligence to revolutionize pharmacy practice by addressing the complexity of pharmaceutical data, changing healthcare demands, and technological advancements. This article identifies the critical need for integrating these technologies to enhance medication management, improve patient outcomes, and streamline pharmacy operations. It also underscores the importance of regulatory and ethical considerations in implementing pharmacointelligence, ensuring patient privacy, data security, and equitable healthcare delivery.

9.
Comput Struct Biotechnol J ; 24: 533-541, 2024 Dec.
Article in English | MEDLINE | ID: mdl-39220685

ABSTRACT

Objectives: Urinary tract infections (UTIs) are common infections within the Emergency Department (ED), causing increased laboratory workloads and unnecessary antibiotics prescriptions. The aim of this study was to improve UTI diagnostics in clinical practice by application of machine learning (ML) models for real-time UTI prediction. Methods: In a retrospective study, patient information and outcomes from Emergency Department patients, with positive and negative culture results, were used to design models - 'Random Forest' and 'Neural Network' - for the prediction of UTIs. The performance of these predictive models was validated in a cross-sectional study. In a quasi-experimental study, the impact of UTI risk assessment was investigated by evaluating changes in the behaviour of clinicians, measuring changes in antibiotic prescriptions and urine culture requests. Results: First, we trained and tested two different predictive models with 8692 cases. Second, we investigated the performance of the predictive models in clinical practice with 962 cases (Area under the curve was between 0.81 to 0.88). The best performance was the combination of both models. Finally, the assessment of the risk for UTIs was implemented into clinical practice and allowed for the reduction of unnecessary urine cultures and antibiotic prescriptions for patients with a low risk of UTI, as well as targeted diagnostics and treatment for patients with a high risk of UTI. Conclusion: The combination of modern urinalysis diagnostic technologies with digital health solutions can help to further improve UTI diagnostics with positive impact on laboratory workloads and antimicrobial stewardship.

10.
Br J Clin Pharmacol ; 2024 Sep 10.
Article in English | MEDLINE | ID: mdl-39256034

ABSTRACT

AIMS: Computerized decision support systems (CDSSs) aim to prevent adverse drug events. However, these systems generate an overload of alerts that are not always clinically relevant. Anticoagulants are frequently involved in these alerts. The aim of this study was to investigate the efficiency of CDSS alerts on anticoagulants in Dutch hospital pharmacies. METHODS: A multicentre, single-day, cross-sectional study was conducted using a flashmob design in Dutch hospital pharmacies, which have CDSSs that operate on both a national medication surveillance database and on self-developed clinical rules. Hospital pharmacists and pharmacy technicians collected data on the number and type of alerts and time needed for assessing these alerts. The primary outcome was the CDSS efficiency on anticoagulants, defined as the percentage of alerts on anticoagulants that led to an intervention. Secondary outcomes where among other CDSSs efficiency related to any medications and the time expenditure. Descriptive data-analysis was used. RESULTS: Of the 69 hospital pharmacies invited, 42 (61%) participated. The efficiency of CDSS alerts on anticoagulants was 4.0% (interquartile range [IQR] 14.0%) for the national medication surveillance database alerts and 14.3% (IQR 40.0%) for alerts from clinical rules. For any medication, the efficiency was lower: 1.8% (IQR 7.5%) and 13.4% (IQR 21.5%) respectively. The median time for assessing the relevance of all alerts was 2 (IQR 1:21) h/day for pharmacists and 6 (IQR 5:01) h/day for pharmacy technicians. CONCLUSION: CDSS efficiency is generally low, both for anticoagulants and any medication, while the time investment is high. Optimization of CDSSs is needed.

11.
Acute Crit Care ; 39(3): 400-407, 2024 Aug.
Article in English | MEDLINE | ID: mdl-39266275

ABSTRACT

BACKGROUND: Diagnosing pediatric septic shock is difficult due to the complex and often impractical traditional criteria, such as systemic inflammatory response syndrome (SIRS), which result in delays and higher risks. This study aims to develop a deep learning-based model using SIRS data for early diagnosis in pediatric septic shock cases. METHODS: The study analyzed data from pediatric patients (<18 years old) admitted to a tertiary hospital from January 2010 to July 2023. Vital signs, lab tests, and clinical information were collected. Septic shock cases were identified using SIRS criteria and inotrope use. A deep learning model was trained and evaluated using the area under the receiver operating characteristics curve (AUROC) and area under the precision-recall curve (AUPRC). Variable contributions were analyzed using the Shapley additive explanation value. RESULTS: The analysis, involving 9,616,115 measurements, identified 34,696 septic shock cases (0.4%). Oxygen supply was crucial for 41.5% of the control group and 20.8% of the septic shock group. The final model showed strong performance, with an AUROC of 0.927 and AUPRC of 0.879. Key influencers were age, oxygen supply, sex, and partial pressure of carbon dioxide, while body temperature had minimal impact on estimation. CONCLUSIONS: The proposed deep learning model simplifies early septic shock diagnosis in pediatric patients, reducing the diagnostic workload. Its high accuracy allows timely treatment, but external validation through prospective studies is needed.

12.
medRxiv ; 2024 Sep 01.
Article in English | MEDLINE | ID: mdl-39252910

ABSTRACT

Background: Guidelines recommend pharmacological venous thromboembolism (VTE) prophylaxis only for high-risk patients, but the probability of VTE considered "high-risk" is not specified. Our objective was to define an appropriate probability threshold (or range) for VTE risk stratification and corresponding prophylaxis in medical inpatients. Methods: Patients were adults admitted to any of 10 Cleveland Clinic Health System hospitals between December 2020 and August 2021 (N = 41,036). Hospital medicine physicians and internal medicine residents from included hospitals were surveyed between June and November 2023 (N = 214). We compared five approaches to determining a threshold: decision analysis, maximizing the sensitivity and specificity of a logistic regression model, deriving a probability from a point-based model, surveying physicians' understanding of VTE risk, and deriving a probability from physician behavior. For each approach, we determined the probability threshold above which a patient would be considered high-risk for VTE. We applied each threshold to the Cleveland Clinic VTE risk assessment model (CCM) and calculated the percentage of the 41,036 patients in our cohort who would be considered eligible for prophylaxis due to their high-risk status. We compared these hypothetical prophylaxis rates with physicians' observed prophylaxis rates. Results: The different approaches yielded thresholds ranging from 0.3% to 5.4%, corresponding inversely with hypothetical prophylaxis rates of 0.2% to 75%. Multiple thresholds clustered between 0.52% to 0.55%, suggesting an average hypothetical prophylaxis rate of approximately 30%, whereas physicians' observed prophylaxis rates ranged from 48% to 76%. Conclusions: Multiple approaches to determining a probability threshold for VTE prophylaxis converged to suggest an optimal threshold of approximately 0.5%. Other approaches yielded extreme thresholds that are unrealistic for clinical practice. Physicians prescribed prophylaxis much more frequently than the suggested rate of 30%, indicating opportunity to reduce unnecessary prophylaxis. To aid in these efforts, guidelines should explicitly quantify high-risk.

13.
Healthcare (Basel) ; 12(17)2024 Aug 26.
Article in English | MEDLINE | ID: mdl-39273719

ABSTRACT

BACKGROUND: COVID-19 has had a substantial influence on healthcare systems, requiring early prognosis for innovative therapies and optimal results, especially in individuals with comorbidities. AI systems have been used by healthcare practitioners for investigating, anticipating, and predicting diseases, through means including medication development, clinical trial analysis, and pandemic forecasting. This study proposes the use of AI to predict disease severity in terms of hospital mortality among COVID-19 patients. METHODS: A cross-sectional study was conducted at King Abdulaziz University, Saudi Arabia. Data were cleaned by encoding categorical variables and replacing missing quantitative values with their mean. The outcome variable, hospital mortality, was labeled as death = 0 or survival = 1, with all baseline investigations, clinical symptoms, and laboratory findings used as predictors. Decision trees, SVM, and random forest algorithms were employed. The training process included splitting the data set into training and testing sets, performing 5-fold cross-validation to tune hyperparameters, and evaluating performance on the test set using accuracy. RESULTS: The study assessed the predictive accuracy of outcomes and mortality for COVID-19 patients based on factors such as CRP, LDH, Ferritin, ALP, Bilirubin, D-Dimers, and hospital stay (p-value ≤ 0.05). The analysis revealed that hospital stay, D-Dimers, ALP, Bilirubin, LDH, CRP, and Ferritin significantly influenced hospital mortality (p ≤ 0.0001). The results demonstrated high predictive accuracy, with decision trees achieving 76%, random forest 80%, and support vector machines (SVMs) 82%. CONCLUSIONS: Artificial intelligence is a tool crucial for identifying early coronavirus infections and monitoring patient conditions. It improves treatment consistency and decision-making via the development of algorithms.

14.
BMC Med Inform Decis Mak ; 24(1): 241, 2024 Sep 02.
Article in English | MEDLINE | ID: mdl-39223512

ABSTRACT

BACKGROUND: Successful deployment of clinical prediction models for clinical deterioration relates not only to predictive performance but to integration into the decision making process. Models may demonstrate good discrimination and calibration, but fail to match the needs of practising acute care clinicians who receive, interpret, and act upon model outputs or alerts. We sought to understand how prediction models for clinical deterioration, also known as early warning scores (EWS), influence the decision-making of clinicians who regularly use them and elicit their perspectives on model design to guide future deterioration model development and implementation. METHODS: Nurses and doctors who regularly receive or respond to EWS alerts in two digital metropolitan hospitals were interviewed for up to one hour between February 2022 and March 2023 using semi-structured formats. We grouped interview data into sub-themes and then into general themes using reflexive thematic analysis. Themes were then mapped to a model of clinical decision making using deductive framework mapping to develop a set of practical recommendations for future deterioration model development and deployment. RESULTS: Fifteen nurses (n = 8) and doctors (n = 7) were interviewed for a mean duration of 42 min. Participants emphasised the importance of using predictive tools for supporting rather than supplanting critical thinking, avoiding over-protocolising care, incorporating important contextual information and focusing on how clinicians generate, test, and select diagnostic hypotheses when managing deteriorating patients. These themes were incorporated into a conceptual model which informed recommendations that clinical deterioration prediction models demonstrate transparency and interactivity, generate outputs tailored to the tasks and responsibilities of end-users, avoid priming clinicians with potential diagnoses before patients were physically assessed, and support the process of deciding upon subsequent management. CONCLUSIONS: Prediction models for deteriorating inpatients may be more impactful if they are designed in accordance with the decision-making processes of acute care clinicians. Models should produce actionable outputs that assist with, rather than supplant, critical thinking.


Subject(s)
Clinical Decision-Making , Clinical Deterioration , Early Warning Score , Humans , Critical Care/standards , Attitude of Health Personnel , Female , Male , Adult , Physicians
15.
Curr Oncol ; 31(9): 4984-5007, 2024 Aug 27.
Article in English | MEDLINE | ID: mdl-39329997

ABSTRACT

The integration of multidisciplinary tumor boards (MTBs) is fundamental in delivering state-of-the-art cancer treatment, facilitating collaborative diagnosis and management by a diverse team of specialists. Despite the clear benefits in personalized patient care and improved outcomes, the increasing burden on MTBs due to rising cancer incidence and financial constraints necessitates innovative solutions. The advent of artificial intelligence (AI) in the medical field offers a promising avenue to support clinical decision-making. This review explores the perspectives of clinicians dedicated to the care of cancer patients-surgeons, medical oncologists, and radiation oncologists-on the application of AI within MTBs. Additionally, it examines the role of AI across various clinical specialties involved in cancer diagnosis and treatment. By analyzing both the potential and the challenges, this study underscores how AI can enhance multidisciplinary discussions and optimize treatment plans. The findings highlight the transformative role that AI may play in refining oncology care and sustaining the efficacy of MTBs amidst growing clinical demands.


Subject(s)
Artificial Intelligence , Oncologists , Radiation Oncologists , Humans , Neoplasms/therapy , Surgeons , Medical Oncology/methods , Radiation Oncology/methods
16.
Sci Rep ; 14(1): 21820, 2024 09 18.
Article in English | MEDLINE | ID: mdl-39294200

ABSTRACT

Feature Selection (FS) is essential in the Internet of Things (IoT)-based Clinical Decision Support Systems (CDSS) to improve the accuracy and efficiency of the system. With the increasing number of sensors and devices used in healthcare, the volume of data generated is vast and complex. Relevant FS from this data is crucial in reducing computational overhead, improving the system's interpretability, and enhancing the Decision-Making System (DMS) quality. FS also aids in addressing the problems of data redundancy and noise, which can negatively impact the system's performance. FS is critical to developing practical and dependable CDSS in IoT-based healthcare sectors. This research proposes a two-phase FS model. Phase-I employs an ensemble of five Filter Methods (FM), followed by a Pearson Correlation Method (PCM). Phase-II uses the Binary Optimized Genetic Grey Wolf Optimization Algorithm (BOGGWOA) as a Wrapper Method (WM). This recommended model integrates the most valuable features of each filter. Then, it uses the Pearson Correlation Coefficient (PCC) to get rid of features that aren't needed, a Support Vector Machine (SVM) to guess how accurate their classification will be, and BOGGWOA as the Wrapper Method (WM) to pick the most essential features with the best CA.


Subject(s)
Algorithms , Decision Support Systems, Clinical , Internet of Things , Humans , Support Vector Machine , Delivery of Health Care
17.
J Clin Transl Sci ; 8(1): e132, 2024.
Article in English | MEDLINE | ID: mdl-39345695

ABSTRACT

Background: Central venous lines (CVLs) are frequently utilized in critically ill patients and confer a risk of central line-associated bloodstream infections (CLABSIs). CLABSIs are associated with increased mortality, extended hospitalization, and increased costs. Unnecessary CVL utilization contributes to CLABSIs. This initiative sought to implement a clinical decision support system (CDSS) within an electronic health record (EHR) to quantify the prevalence of potentially unnecessary CVLs and improve their timely removal in six adult intensive care units (ICUs). Methods: Intervention components included: (1) evaluating existing CDSS' effectiveness, (2) clinician education, (3) developing/implementing an EHR-based CDSS to identify potentially unnecessary CVLs, (4) audit/feedback, and (5) reviewing EHR/institutional data to compare rates of removal of potentially unnecessary CVLs, device utilization, and CLABSIs pre- and postimplementation. Data was evaluated with statistical process control charts, chi-square analyses, and incidence rate ratios. Results: Preimplementation, 25.2% of CVLs were potentially removable, and the mean weekly proportion of these CVLs that were removed within 24 hours was 20.0%. Postimplementation, a greater proportion of potentially unnecessary CVLs were removed (29%, p < 0.0001), CVL utilization decreased, and days between CLABSIs increased. The intervention was most effective in ICUs staffed by pulmonary/critical care physicians, who received monthly audit/feedback, where timely CVL removal increased from a mean of 18.0% to 30.5% (p < 0.0001) and days between CLABSIs increased from 17.3 to 25.7. Conclusions: A significant proportion of active CVLs were potentially unnecessary. CDSS implementation, in conjunction with audit and feedback, correlated with a sustained increase in timely CVL removal and an increase in days between CLABSIs.

18.
J Med Internet Res ; 26: e54737, 2024 Sep 16.
Article in English | MEDLINE | ID: mdl-39283665

ABSTRACT

BACKGROUND: Despite the emerging application of clinical decision support systems (CDSS) in pregnancy care and the proliferation of artificial intelligence (AI) over the last decade, it remains understudied regarding the role of AI in CDSS specialized for pregnancy care. OBJECTIVE: To identify and synthesize AI-augmented CDSS in pregnancy care, CDSS functionality, AI methodologies, and clinical implementation, we reported a systematic review based on empirical studies that examined AI-augmented CDSS in pregnancy care. METHODS: We retrieved studies that examined AI-augmented CDSS in pregnancy care using database queries involved with titles, abstracts, keywords, and MeSH (Medical Subject Headings) terms. Bibliographic records from their inception to 2022 were retrieved from PubMed/MEDLINE (n=206), Embase (n=101), and ACM Digital Library (n=377), followed by eligibility screening and literature review. The eligibility criteria include empirical studies that (1) developed or tested AI methods, (2) developed or tested CDSS or CDSS components, and (3) focused on pregnancy care. Data of studies used for review and appraisal include title, abstract, keywords, MeSH terms, full text, and supplements. Publications with ancillary information or overlapping outcomes were synthesized as one single study. Reviewers independently reviewed and assessed the quality of selected studies. RESULTS: We identified 30 distinct studies of 684 studies from their inception to 2022. Topics of clinical applications covered AI-augmented CDSS from prenatal, early pregnancy, obstetric care, and postpartum care. Topics of CDSS functions include diagnostic support, clinical prediction, therapeutics recommendation, and knowledge base. CONCLUSIONS: Our review acknowledged recent advances in CDSS studies including early diagnosis of prenatal abnormalities, cost-effective surveillance, prenatal ultrasound support, and ontology development. To recommend future directions, we also noted key gaps from existing studies, including (1) decision support in current childbirth deliveries without using observational data from consequential fetal or maternal outcomes in future pregnancies; (2) scarcity of studies in identifying several high-profile biases from CDSS, including social determinants of health highlighted by the American College of Obstetricians and Gynecologists; and (3) chasm between internally validated CDSS models, external validity, and clinical implementation.


Subject(s)
Artificial Intelligence , Decision Support Systems, Clinical , Humans , Pregnancy , Female , Prenatal Care/methods
19.
J Med Syst ; 48(1): 93, 2024 Sep 30.
Article in English | MEDLINE | ID: mdl-39347841

ABSTRACT

Fixed and broad screening intervals for drug-drug interaction (DDI) alerts lead to false positive alerts, thereby contributing to alert fatigue among healthcare professionals. Hence, we aimed to investigate the impact of customized screening intervals on the daily incidence of DDI alerts. An interrupted time series analysis was performed at the University Hospitals Leuven to evaluate the impact of a pragmatic intervention on the daily incidence of DDI alerts per 100 prescriptions. The study period encompassed 100 randomly selected days between April 2021 and December 2022. Preceding the intervention, a fixed and broad screening interval of 7 days before and after prescribing an interacting drug was applied. The intervention involved implementing customized screening intervals for a subset of highly prevalent or clinically relevant DDIs into the hospital information system. Additionally, the sensitivity of the tailored approach was evaluated. During the study period, a mean of 5731 (± 2909) new prescriptions per day was generated. The daily incidence of DDI alerts significantly decreased from 9.8% (95% confidence interval (CI) 8.4;11.1) before the intervention, to 6.3% (95% CI 5.4;7.2) afterwards, p < 0.0001. This corresponded to avoiding 201 (0.035*5731) false positive DDI alerts per day. Sensitivity was not compromised by our intervention. Defining and implementing customized screening intervals was feasible and effective in reducing the DDI alert burden without compromising sensitivity.


Subject(s)
Drug Interactions , Interrupted Time Series Analysis , Medical Order Entry Systems , Humans , Medication Errors/prevention & control , Alert Fatigue, Health Personnel/prevention & control , Hospital Information Systems , Time Factors , Belgium
20.
Health Serv Res Manag Epidemiol ; 11: 23333928241275292, 2024.
Article in English | MEDLINE | ID: mdl-39211386

ABSTRACT

Objective: Diabetes mellitus is an important chronic disease that is prevalent around the world. Different countries and diverse cultures use varying approaches to dealing with this chronic condition. Also, with the advancement of computation and automated decision-making, many tools and technologies are now available to patients suffering from this disease. In this work, the investigators attempt to analyze approaches taken towards managing this illness in India and the United States. Methods: In this work, the investigators have used available literature and data to compare the use of artificial intelligence in diabetes management. Findings: The article provides key insights to comparison of diabetes management in terms of the nature of the healthcare system, availability, electronic health records, cultural factors, data privacy, affordability, and other important variables. Interestingly, variables such as quality of electronic health records, and cultural factors are key impediments in implementing an efficiency-driven management system for dealing with this chronic disease. Conclusion: The article adds to the body of knowledge associated with the management of this disease, establishing a critical need for using artificial intelligence in diabetes care management.

SELECTION OF CITATIONS
SEARCH DETAIL