Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 104
Filter
1.
Sci Rep ; 14(1): 10594, 2024 05 08.
Article in English | MEDLINE | ID: mdl-38719953

ABSTRACT

Colorectal liver metastases (CRLM) are the predominant factor limiting survival in patients with colorectal cancer and liver resection with complete tumor removal is the best treatment option for these patients. This study examines the predictive ability of three-dimensional lung volumetry (3DLV) based on preoperative computerized tomography (CT), to predict postoperative pulmonary complications in patients undergoing major liver resection for CRLM. Patients undergoing major curative liver resection for CRLM between 2010 and 2021 with a preoperative CT scan of the thorax within 6 weeks of surgery, were included. Total lung volume (TLV) was calculated using volumetry software 3D-Slicer version 4.11.20210226 including Chest Imaging Platform extension ( http://www.slicer.org ). The area under the curve (AUC) of a receiver-operating characteristic analysis was used to define a cut-off value of TLV, for predicting the occurrence of postoperative respiratory complications. Differences between patients with TLV below and above the cut-off were examined with Chi-square or Fisher's exact test and Mann-Whitney U tests and logistic regression was used to determine independent risk factors for the development of respiratory complications. A total of 123 patients were included, of which 35 (29%) developed respiratory complications. A predictive ability of TLV regarding respiratory complications was shown (AUC 0.62, p = 0.036) and a cut-off value of 4500 cm3 was defined. Patients with TLV < 4500 cm3 were shown to suffer from significantly higher rates of respiratory complications (44% vs. 21%, p = 0.007) compared to the rest. Logistic regression analysis identified TLV < 4500 cm3 as an independent predictor for the occurrence of respiratory complications (odds ratio 3.777, 95% confidence intervals 1.488-9.588, p = 0.005). Preoperative 3DLV is a viable technique for prediction of postoperative pulmonary complications in patients undergoing major liver resection for CRLM. More studies in larger cohorts are necessary to further evaluate this technique.


Subject(s)
Colorectal Neoplasms , Hepatectomy , Liver Neoplasms , Postoperative Complications , Tomography, X-Ray Computed , Humans , Female , Male , Colorectal Neoplasms/pathology , Colorectal Neoplasms/surgery , Middle Aged , Liver Neoplasms/surgery , Liver Neoplasms/secondary , Aged , Hepatectomy/adverse effects , Hepatectomy/methods , Postoperative Complications/etiology , Lung/pathology , Lung/diagnostic imaging , Lung/surgery , Retrospective Studies , Imaging, Three-Dimensional , Lung Volume Measurements , Risk Factors , Preoperative Period
2.
J Med Internet Res ; 26: e54948, 2024 May 01.
Article in English | MEDLINE | ID: mdl-38691404

ABSTRACT

This study demonstrates that GPT-4V outperforms GPT-4 across radiology subspecialties in analyzing 207 cases with 1312 images from the Radiological Society of North America Case Collection.


Subject(s)
Radiology , Radiology/methods , Radiology/statistics & numerical data , Humans , Image Processing, Computer-Assisted/methods
3.
Article in English | MEDLINE | ID: mdl-38627537

ABSTRACT

Liver cancer has high incidence and mortality globally. Artificial intelligence (AI) has advanced rapidly, influencing cancer care. AI systems are already approved for clinical use in some tumour types (for example, colorectal cancer screening). Crucially, research demonstrates that AI can analyse histopathology, radiology and natural language in liver cancer, and can replace manual tasks and access hidden information in routinely available clinical data. However, for liver cancer, few of these applications have translated into large-scale clinical trials or clinically approved products. Here, we advocate for the incorporation of AI in all stages of liver cancer management. We present a taxonomy of AI approaches in liver cancer, highlighting areas with academic and commercial potential, and outline a policy for AI-based liver cancer management, including interdisciplinary training of researchers, clinicians and patients. The potential of AI in liver cancer is immense, but effort is required to ensure that AI can fulfil expectations.

4.
Eur Radiol ; 2024 Apr 16.
Article in English | MEDLINE | ID: mdl-38627289

ABSTRACT

OBJECTIVES: Large language models (LLMs) have shown potential in radiology, but their ability to aid radiologists in interpreting imaging studies remains unexplored. We investigated the effects of a state-of-the-art LLM (GPT-4) on the radiologists' diagnostic workflow. MATERIALS AND METHODS: In this retrospective study, six radiologists of different experience levels read 40 selected radiographic [n = 10], CT [n = 10], MRI [n = 10], and angiographic [n = 10] studies unassisted (session one) and assisted by GPT-4 (session two). Each imaging study was presented with demographic data, the chief complaint, and associated symptoms, and diagnoses were registered using an online survey tool. The impact of Artificial Intelligence (AI) on diagnostic accuracy, confidence, user experience, input prompts, and generated responses was assessed. False information was registered. Linear mixed-effect models were used to quantify the factors (fixed: experience, modality, AI assistance; random: radiologist) influencing diagnostic accuracy and confidence. RESULTS: When assessing if the correct diagnosis was among the top-3 differential diagnoses, diagnostic accuracy improved slightly from 181/240 (75.4%, unassisted) to 188/240 (78.3%, AI-assisted). Similar improvements were found when only the top differential diagnosis was considered. AI assistance was used in 77.5% of the readings. Three hundred nine prompts were generated, primarily involving differential diagnoses (59.1%) and imaging features of specific conditions (27.5%). Diagnostic confidence was significantly higher when readings were AI-assisted (p > 0.001). Twenty-three responses (7.4%) were classified as hallucinations, while two (0.6%) were misinterpretations. CONCLUSION: Integrating GPT-4 in the diagnostic process improved diagnostic accuracy slightly and diagnostic confidence significantly. Potentially harmful hallucinations and misinterpretations call for caution and highlight the need for further safeguarding measures. CLINICAL RELEVANCE STATEMENT: Using GPT-4 as a virtual assistant when reading images made six radiologists of different experience levels feel more confident and provide more accurate diagnoses; yet, GPT-4 gave factually incorrect and potentially harmful information in 7.4% of its responses.

5.
Commun Med (Lond) ; 4(1): 71, 2024 Apr 11.
Article in English | MEDLINE | ID: mdl-38605106

ABSTRACT

BACKGROUND: The field of Artificial Intelligence (AI) holds transformative potential in medicine. However, the lack of universal reporting guidelines poses challenges in ensuring the validity and reproducibility of published research studies in this field. METHODS: Based on a systematic review of academic publications and reporting standards demanded by both international consortia and regulatory stakeholders as well as leading journals in the fields of medicine and medical informatics, 26 reporting guidelines published between 2009 and 2023 were included in this analysis. Guidelines were stratified by breadth (general or specific to medical fields), underlying consensus quality, and target research phase (preclinical, translational, clinical) and subsequently analyzed regarding the overlap and variations in guideline items. RESULTS: AI reporting guidelines for medical research vary with respect to the quality of the underlying consensus process, breadth, and target research phase. Some guideline items such as reporting of study design and model performance recur across guidelines, whereas other items are specific to particular fields and research stages. CONCLUSIONS: Our analysis highlights the importance of reporting guidelines in clinical AI research and underscores the need for common standards that address the identified variations and gaps in current guidelines. Overall, this comprehensive overview could help researchers and public stakeholders reinforce quality standards for increased reliability, reproducibility, clinical validity, and public trust in AI research in healthcare. This could facilitate the safe, effective, and ethical translation of AI methods into clinical applications that will ultimately improve patient outcomes.


Artificial Intelligence (AI) refers to computer systems that can perform tasks that normally require human intelligence, like recognizing patterns or making decisions. AI has the potential to transform healthcare, but research on AI in medicine needs clear rules so caregivers and patients can trust it. This study reviews and compares 26 existing guidelines for reporting on AI in medicine. The key differences between these guidelines are their target areas (medicine in general or specific medical fields), the ways they were created, and the research stages they address. While some key items like describing the AI model recurred across guidelines, others were specific to the research area. The analysis shows gaps and variations in current guidelines. Overall, transparent reporting is important, so AI research is reliable, reproducible, trustworthy, and safe for patients. This systematic review of guidelines aims to increase the transparency of AI research, supporting an ethical and safe progression of AI from research into clinical practice.

6.
Comput Biol Med ; 175: 108410, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38678938

ABSTRACT

Latent diffusion models (LDMs) have emerged as a state-of-the-art image generation method, outperforming previous Generative Adversarial Networks (GANs) in terms of training stability and image quality. In computational pathology, generative models are valuable for data sharing and data augmentation. However, the impact of LDM-generated images on histopathology tasks compared to traditional GANs has not been systematically studied. We trained three LDMs and a styleGAN2 model on histology tiles from nine colorectal cancer (CRC) tissue classes. The LDMs include 1) a fine-tuned version of stable diffusion v1.4, 2) a Kullback-Leibler (KL)-autoencoder (KLF8-DM), and 3) a vector quantized (VQ)-autoencoder deploying LDM (VQF8-DM). We assessed image quality through expert ratings, dimensional reduction methods, distribution similarity measures, and their impact on training a multiclass tissue classifier. Additionally, we investigated image memorization in the KLF8-DM and styleGAN2 models. All models provided a high image quality, with the KLF8-DM achieving the best Frechet Inception Distance (FID) and expert rating scores for complex tissue classes. For simpler classes, the VQF8-DM and styleGAN2 models performed better. Image memorization was negligible for both styleGAN2 and KLF8-DM models. Classifiers trained on a mix of KLF8-DM generated and real images achieved a 4% improvement in overall classification accuracy, highlighting the usefulness of these images for dataset augmentation. Our systematic study of generative methods showed that KLF8-DM produces the highest quality images with negligible image memorization. The higher classifier performance in the generatively augmented dataset suggests that this augmentation technique can be employed to enhance histopathology classifiers for various tasks.


Subject(s)
Colorectal Neoplasms , Humans , Colorectal Neoplasms/pathology , Colorectal Neoplasms/diagnostic imaging , Image Interpretation, Computer-Assisted/methods , Image Processing, Computer-Assisted/methods , Algorithms
7.
Eur Radiol Exp ; 8(1): 53, 2024 May 01.
Article in English | MEDLINE | ID: mdl-38689178

ABSTRACT

BACKGROUND: To compare denoising diffusion probabilistic models (DDPM) and generative adversarial networks (GAN) for recovering contrast-enhanced breast magnetic resonance imaging (MRI) subtraction images from virtual low-dose subtraction images. METHODS: Retrospective, ethically approved study. DDPM- and GAN-reconstructed single-slice subtraction images of 50 breasts with enhancing lesions were compared to original ones at three dose levels (25%, 10%, 5%) using quantitative measures and radiologic evaluations. Two radiologists stated their preference based on the reconstruction quality and scored the lesion conspicuity as compared to the original, blinded to the model. Fifty lesion-free maximum intensity projections were evaluated for the presence of false-positives. Results were compared between models and dose levels, using generalized linear mixed models. RESULTS: At 5% dose, both radiologists preferred the GAN-generated images, whereas at 25% dose, both radiologists preferred the DDPM-generated images. Median lesion conspicuity scores did not differ between GAN and DDPM at 25% dose (5 versus 5, p = 1.000) and 10% dose (4 versus 4, p = 1.000). At 5% dose, both readers assigned higher conspicuity to the GAN than to the DDPM (3 versus 2, p = 0.007). In the lesion-free examinations, DDPM and GAN showed no differences in the false-positive rate at 5% (15% versus 22%), 10% (10% versus 6%), and 25% (6% versus 4%) (p = 1.000). CONCLUSIONS: Both GAN and DDPM yielded promising results in low-dose image reconstruction. However, neither of them showed superior results over the other model for all dose levels and evaluation metrics. Further development is needed to counteract false-positives. RELEVANCE STATEMENT: For MRI-based breast cancer screening, reducing the contrast agent dose is desirable. Diffusion probabilistic models and generative adversarial networks were capable of retrospectively enhancing the signal of low-dose images. Hence, they may supplement imaging with reduced doses in the future. KEY POINTS: • Deep learning may help recover signal in low-dose contrast-enhanced breast MRI. • Two models (DDPM and GAN) were trained at different dose levels. • Radiologists preferred DDPM at 25%, and GAN images at 5% dose. • Lesion conspicuity between DDPM and GAN was similar, except at 5% dose. • GAN and DDPM yield promising results in low-dose image reconstruction.


Subject(s)
Breast Neoplasms , Contrast Media , Magnetic Resonance Imaging , Humans , Female , Retrospective Studies , Contrast Media/administration & dosage , Breast Neoplasms/diagnostic imaging , Magnetic Resonance Imaging/methods , Middle Aged , Models, Statistical , Adult , Aged
8.
JAMA ; 331(15): 1320-1321, 2024 04 16.
Article in English | MEDLINE | ID: mdl-38497956

ABSTRACT

This study compares 2 large language models and their performance vs that of competing open-source models.


Subject(s)
Artificial Intelligence , Diagnostic Imaging , Medical History Taking , Language
10.
Commun Med (Lond) ; 4(1): 46, 2024 Mar 14.
Article in English | MEDLINE | ID: mdl-38486100

ABSTRACT

BACKGROUND: Artificial intelligence (AI) models are increasingly used in the medical domain. However, as medical data is highly sensitive, special precautions to ensure its protection are required. The gold standard for privacy preservation is the introduction of differential privacy (DP) to model training. Prior work indicates that DP has negative implications on model accuracy and fairness, which are unacceptable in medicine and represent a main barrier to the widespread use of privacy-preserving techniques. In this work, we evaluated the effect of privacy-preserving training of AI models regarding accuracy and fairness compared to non-private training. METHODS: We used two datasets: (1) A large dataset (N = 193,311) of high quality clinical chest radiographs, and (2) a dataset (N = 1625) of 3D abdominal computed tomography (CT) images, with the task of classifying the presence of pancreatic ductal adenocarcinoma (PDAC). Both were retrospectively collected and manually labeled by experienced radiologists. We then compared non-private deep convolutional neural networks (CNNs) and privacy-preserving (DP) models with respect to privacy-utility trade-offs measured as area under the receiver operating characteristic curve (AUROC), and privacy-fairness trade-offs, measured as Pearson's r or Statistical Parity Difference. RESULTS: We find that, while the privacy-preserving training yields lower accuracy, it largely does not amplify discrimination against age, sex or co-morbidity. However, we find an indication that difficult diagnoses and subgroups suffer stronger performance hits in private training. CONCLUSIONS: Our study shows that - under the challenging realistic circumstances of a real-life clinical dataset - the privacy-preserving training of diagnostic deep learning models is possible with excellent diagnostic accuracy and fairness.


Artificial intelligence (AI), in which computers can learn to do tasks that normally require human intelligence, is particularly useful in medical imaging. However, AI should be used in a way that preserves patient privacy. We explored the balance between maintaining patient data privacy and AI performance in medical imaging. We use an approach called differential privacy to protect the privacy of patients' images. We show that, although training AI with differential privacy leads to a slight decrease in accuracy, it does not substantially increase bias against different age groups, genders, or patients with multiple health conditions. However, we notice that AI faces more challenges in accurately diagnosing complex cases and specific subgroups when trained under these privacy constraints. These findings highlight the importance of designing AI systems that are both privacy-conscious and capable of reliable diagnoses across patient groups.

11.
Diagnostics (Basel) ; 14(5)2024 Feb 23.
Article in English | MEDLINE | ID: mdl-38472955

ABSTRACT

Increased attention has been given to MRI in radiation-free screening for malignant nodules in recent years. Our objective was to compare the performance of human readers and radiomic feature analysis based on stand-alone and complementary CT and MRI imaging in classifying pulmonary nodules. This single-center study comprises patients with CT findings of pulmonary nodules who underwent additional lung MRI and whose nodules were classified as benign/malignant by resection. For radiomic features analysis, 2D segmentation was performed for each lung nodule on axial CT, T2-weighted (T2w), and diffusion (DWI) images. The 105 extracted features were reduced by iterative backward selection. The performance of radiomics and human readers was compared by calculating accuracy with Clopper-Pearson confidence intervals. Fifty patients (mean age 63 +/- 10 years) with 66 pulmonary nodules (40 malignant) were evaluated. ACC values for radiomic features analysis vs. radiologists based on CT alone (0.68; 95%CI: 0.56, 0.79 vs. 0.59; 95%CI: 0.46, 0.71), T2w alone (0.65; 95%CI: 0.52, 0.77 vs. 0.68; 95%CI: 0.54, 0.78), DWI alone (0.61; 95%CI:0.48, 0.72 vs. 0.73; 95%CI: 0.60, 0.83), combined T2w/DWI (0.73; 95%CI: 0.60, 0.83 vs. 0.70; 95%CI: 0.57, 0.80), and combined CT/T2w/DWI (0.83; 95%CI: 0.72, 0.91 vs. 0.64; 95%CI: 0.51, 0.75) were calculated. This study is the first to show that by combining quantitative image information from CT, T2w, and DWI datasets, pulmonary nodule assessment through radiomics analysis is superior to using one modality alone, even exceeding human readers' performance.

12.
NPJ Precis Oncol ; 8(1): 72, 2024 Mar 22.
Article in English | MEDLINE | ID: mdl-38519519

ABSTRACT

The technological progress in artificial intelligence (AI) has massively accelerated since 2022, with far-reaching implications for oncology and cancer research. Large language models (LLMs) now perform at human-level competency in text processing. Notably, both text and image processing networks are increasingly based on transformer neural networks. This convergence enables the development of multimodal AI models that take diverse types of data as an input simultaneously, marking a qualitative shift from specialized niche models which were prevalent in the 2010s. This editorial summarizes these developments, which are expected to impact precision oncology in the coming years.

13.
Eur Radiol Exp ; 8(1): 10, 2024 Feb 08.
Article in English | MEDLINE | ID: mdl-38326501

ABSTRACT

BACKGROUND: Pretraining labeled datasets, like ImageNet, have become a technical standard in advanced medical image analysis. However, the emergence of self-supervised learning (SSL), which leverages unlabeled data to learn robust features, presents an opportunity to bypass the intensive labeling process. In this study, we explored if SSL for pretraining on non-medical images can be applied to chest radiographs and how it compares to supervised pretraining on non-medical images and on medical images. METHODS: We utilized a vision transformer and initialized its weights based on the following: (i) SSL pretraining on non-medical images (DINOv2), (ii) supervised learning (SL) pretraining on non-medical images (ImageNet dataset), and (iii) SL pretraining on chest radiographs from the MIMIC-CXR database, the largest labeled public dataset of chest radiographs to date. We tested our approach on over 800,000 chest radiographs from 6 large global datasets, diagnosing more than 20 different imaging findings. Performance was quantified using the area under the receiver operating characteristic curve and evaluated for statistical significance using bootstrapping. RESULTS: SSL pretraining on non-medical images not only outperformed ImageNet-based pretraining (p < 0.001 for all datasets) but, in certain cases, also exceeded SL on the MIMIC-CXR dataset. Our findings suggest that selecting the right pretraining strategy, especially with SSL, can be pivotal for improving diagnostic accuracy of artificial intelligence in medical imaging. CONCLUSIONS: By demonstrating the promise of SSL in chest radiograph analysis, we underline a transformative shift towards more efficient and accurate AI models in medical imaging. RELEVANCE STATEMENT: Self-supervised learning highlights a paradigm shift towards the enhancement of AI-driven accuracy and efficiency in medical imaging. Given its promise, the broader application of self-supervised learning in medical imaging calls for deeper exploration, particularly in contexts where comprehensive annotated datasets are limited.


Subject(s)
Artificial Intelligence , Deep Learning , Databases, Factual
14.
Nat Commun ; 15(1): 1603, 2024 Feb 21.
Article in English | MEDLINE | ID: mdl-38383555

ABSTRACT

A knowledge gap persists between machine learning (ML) developers (e.g., data scientists) and practitioners (e.g., clinicians), hampering the full utilization of ML for clinical data analysis. We investigated the potential of the ChatGPT Advanced Data Analysis (ADA), an extension of GPT-4, to bridge this gap and perform ML analyses efficiently. Real-world clinical datasets and study details from large trials across various medical specialties were presented to ChatGPT ADA without specific guidance. ChatGPT ADA autonomously developed state-of-the-art ML models based on the original study's training data to predict clinical outcomes such as cancer development, cancer progression, disease complications, or biomarkers such as pathogenic gene sequences. Following the re-implementation and optimization of the published models, the head-to-head comparison of the ChatGPT ADA-crafted ML models and their respective manually crafted counterparts revealed no significant differences in traditional performance metrics (p ≥ 0.072). Strikingly, the ChatGPT ADA-crafted ML models often outperformed their counterparts. In conclusion, ChatGPT ADA offers a promising avenue to democratize ML in medicine by simplifying complex data analyses, yet should enhance, not replace, specialized training and resources, to promote broader applications in medical research and practice.


Subject(s)
Algorithms , Neoplasms , Humans , Benchmarking , Language , Machine Learning
15.
Radiologie (Heidelb) ; 64(4): 304-311, 2024 Apr.
Article in German | MEDLINE | ID: mdl-38170243

ABSTRACT

High-quality magnetic resonance (MR) imaging is essential for the precise assessment of the knee joint and plays a key role in the diagnostics, treatment and prognosis. Intact cartilage tissue is characterized by a smooth surface, uniform tissue thickness and an organized zonal structure, which are manifested as depth-dependent signal intensity variations. Cartilage pathologies are identifiable through alterations in signal intensity and morphology and should be communicated based on a precise terminology. Cartilage pathologies can show hyperintense and hypointense signal alterations. Cartilage defects are assessed based on their depth and should be described in terms of their location and extent. The following symptom constellations are of overarching clinical relevance in image reading and interpretation: symptom constellations associated with rapidly progressive forms of joint degeneration and unfavorable prognosis, accompanying symptom constellations mostly in connection with destabilizing meniscal lesions and subchondral insufficiency fractures (accelerated osteoarthritis) as well as symptoms beyond the "typical" degeneration, especially when a discrepancy is observed between (minor) structural changes and (major) synovitis and effusion (inflammatory arthropathy).


Subject(s)
Cartilage, Articular , Osteoarthritis, Knee , Humans , Osteoarthritis, Knee/complications , Osteoarthritis, Knee/pathology , Cartilage, Articular/pathology , Disease Progression , Knee Joint/pathology , Magnetic Resonance Imaging/methods
16.
Injury ; 55(2): 111254, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38070329

ABSTRACT

Delayed functional recovery after injury is associated with significant personal and socioeconomic burden. Identification of patients at risk for a prolonged recovery after a musculoskeletal injury is thus of high relevance. The aim of the current study was to show the feasibility of using a machine learning assisted model to predict functional recovery based on the pre- and immediate post injury patient activity as measured with wearable systems in trauma patients. Patients with a pre-existing wearable (smartphone and/or body-worn sensor), data availability of at least 7 days prior to their injury, and any musculoskeletal injury of the upper or lower extremity were included in this study. Patient age, sex, injured extremity, time off work and step count as activity data were recorded continuously both pre- and post-injury. Descriptive statistics were performed and a logistic regression machine learning model was used to predict the patient's functional recovery status after 6 weeks based on their pre- and post-injury activity characteristics. Overall 38 patients (7 upper extremity, 24 lower extremity, 5 pelvis, 2 combined) were included in this proof-of-concept study. The average follow-up with available wearable data was 85.4 days. Based on the activity data, a predictive model was constructed to determine the likelihood of having a recovery of at least 50 % of the pre-injury activity state by post injury week 6. Based on the individual activity by week 3 a predictive accuracy of over 80 % was achieved on an independent test set (F1=0,82; AUC=0,86; ACC=8,83). The employed model is feasible to assess the principal risk for a slower recovery based on readily available personal wearable activity data. The model has the potential to identify patients requiring additional aftercare attention early during the treatment course, thus optimizing return to the pre-injury status through focused interventions. Additional patient data is needed to adapt the model to more specifically focus on different fracture entities and patient groups.


Subject(s)
Fractures, Bone , Wearable Electronic Devices , Humans , Feasibility Studies , Machine Learning
17.
Arch Gynecol Obstet ; 309(4): 1543-1549, 2024 Apr.
Article in English | MEDLINE | ID: mdl-37975899

ABSTRACT

PURPOSE: The market and application possibilities for artificial intelligence are currently growing at high speed and are increasingly finding their way into gynecology. While the medical side is highly represented in the current literature, the patient's perspective is still lagging behind. Therefore, the aim of this study was to evaluate the recommendations of ChatGPT regarding patient inquiries about the possible therapy of gynecological leading symptoms in a palliative situation by experts. METHODS: Case vignettes were constructed for 10 common concomitant symptoms in gynecologic oncology tumors in a palliative setting, and patient queries regarding therapy of these symptoms were generated as prompts for ChatGPT. Five experts in palliative care and gynecologic oncology evaluated the responses with respect to guideline adherence and applicability and identified advantages and disadvantages. RESULTS: The overall rating of ChatGPT responses averaged 4.1 (5 = strongly agree; 1 = strongly disagree). The experts saw an average guideline conformity of the therapy recommendations with a value of 4.0. ChatGPT sometimes omits relevant therapies and does not provide an individual assessment of the suggested therapies, but does indicate that a physician consultation is additionally necessary. CONCLUSIONS: Language models, such as ChatGPT, can provide valid and largely guideline-compliant therapy recommendations in their freely available and thus in principle accessible version for our patients. For a complete therapy recommendation, an evaluation of the therapies, their individual adjustment as well as a filtering of possible wrong recommendations, a medical expert's opinion remains indispensable.


Subject(s)
Genital Neoplasms, Female , Gynecology , Humans , Female , Artificial Intelligence , Genital Neoplasms, Female/drug therapy , Patient Compliance , Guideline Adherence
18.
IEEE Trans Med Imaging ; 43(1): 542-557, 2024 Jan.
Article in English | MEDLINE | ID: mdl-37713220

ABSTRACT

The early detection of glaucoma is essential in preventing visual impairment. Artificial intelligence (AI) can be used to analyze color fundus photographs (CFPs) in a cost-effective manner, making glaucoma screening more accessible. While AI models for glaucoma screening from CFPs have shown promising results in laboratory settings, their performance decreases significantly in real-world scenarios due to the presence of out-of-distribution and low-quality images. To address this issue, we propose the Artificial Intelligence for Robust Glaucoma Screening (AIROGS) challenge. This challenge includes a large dataset of around 113,000 images from about 60,000 patients and 500 different screening centers, and encourages the development of algorithms that are robust to ungradable and unexpected input data. We evaluated solutions from 14 teams in this paper and found that the best teams performed similarly to a set of 20 expert ophthalmologists and optometrists. The highest-scoring team achieved an area under the receiver operating characteristic curve of 0.99 (95% CI: 0.98-0.99) for detecting ungradable images on-the-fly. Additionally, many of the algorithms showed robust performance when tested on three other publicly available datasets. These results demonstrate the feasibility of robust AI-enabled glaucoma screening.


Subject(s)
Artificial Intelligence , Glaucoma , Humans , Glaucoma/diagnostic imaging , Fundus Oculi , Diagnostic Techniques, Ophthalmological , Algorithms
19.
Br J Clin Pharmacol ; 90(3): 649-661, 2024 03.
Article in English | MEDLINE | ID: mdl-37728146

ABSTRACT

AIMS: To explore international undergraduate pharmacy students' views on integrating artificial intelligence (AI) into pharmacy education and practice. METHODS: This cross-sectional institutional review board-approved multinational, multicentre study comprised an anonymous online survey of 14 multiple-choice items to assess pharmacy students' preferences for AI events in the pharmacy curriculum, the current state of AI education, and students' AI knowledge and attitudes towards using AI in the pharmacy profession, supplemented by 8 demographic queries. Subgroup analyses were performed considering sex, study year, tech-savviness, and prior AI knowledge and AI events in the curriculum using the Mann-Whitney U-test. Variances were reported for responses in Likert scale format. RESULTS: The survey gathered 387 pharmacy student opinions across 16 faculties and 12 countries. Students showed predominantly positive attitudes towards AI in medicine (58%, n = 225) and expressed a strong desire for more AI education (72%, n = 276). However, they reported limited general knowledge of AI (63%, n = 242) and felt inadequately prepared to use AI in their future careers (51%, n = 197). Male students showed more positive attitudes towards increasing efficiency through AI (P = .011), while tech-savvy and advanced-year students expressed heightened concerns about potential legal and ethical issues related to AI (P < .001/P = .025, respectively). Students who had AI courses as part of their studies reported better AI knowledge (P < .001) and felt more prepared to apply it professionally (P < .001). CONCLUSIONS: Our findings underline the generally positive attitude of international pharmacy students towards AI application in medicine and highlight the necessity for a greater emphasis on AI education within pharmacy curricula.


Subject(s)
Students, Pharmacy , Humans , Male , Cross-Sectional Studies , Artificial Intelligence , Surveys and Questionnaires , Curriculum
20.
Eur Radiol ; 34(2): 1176-1178, 2024 Feb.
Article in English | MEDLINE | ID: mdl-37580599
SELECTION OF CITATIONS
SEARCH DETAIL
...