Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 2 de 2
Filtrar
Mais filtros

Base de dados
Tipo de documento
País/Região como assunto
Ano de publicação
Intervalo de ano de publicação
1.
J Med Internet Res ; 26: e53297, 2024 Jun 14.
Artigo em Inglês | MEDLINE | ID: mdl-38875696

RESUMO

BACKGROUND: Large language models (LLMs) have demonstrated impressive performances in various medical domains, prompting an exploration of their potential utility within the high-demand setting of emergency department (ED) triage. This study evaluated the triage proficiency of different LLMs and ChatGPT, an LLM-based chatbot, compared to professionally trained ED staff and untrained personnel. We further explored whether LLM responses could guide untrained staff in effective triage. OBJECTIVE: This study aimed to assess the efficacy of LLMs and the associated product ChatGPT in ED triage compared to personnel of varying training status and to investigate if the models' responses can enhance the triage proficiency of untrained personnel. METHODS: A total of 124 anonymized case vignettes were triaged by untrained doctors; different versions of currently available LLMs; ChatGPT; and professionally trained raters, who subsequently agreed on a consensus set according to the Manchester Triage System (MTS). The prototypical vignettes were adapted from cases at a tertiary ED in Germany. The main outcome was the level of agreement between raters' MTS level assignments, measured via quadratic-weighted Cohen κ. The extent of over- and undertriage was also determined. Notably, instances of ChatGPT were prompted using zero-shot approaches without extensive background information on the MTS. The tested LLMs included raw GPT-4, Llama 3 70B, Gemini 1.5, and Mixtral 8x7b. RESULTS: GPT-4-based ChatGPT and untrained doctors showed substantial agreement with the consensus triage of professional raters (κ=mean 0.67, SD 0.037 and κ=mean 0.68, SD 0.056, respectively), significantly exceeding the performance of GPT-3.5-based ChatGPT (κ=mean 0.54, SD 0.024; P<.001). When untrained doctors used this LLM for second-opinion triage, there was a slight but statistically insignificant performance increase (κ=mean 0.70, SD 0.047; P=.97). Other tested LLMs performed similar to or worse than GPT-4-based ChatGPT or showed odd triaging behavior with the used parameters. LLMs and ChatGPT models tended toward overtriage, whereas untrained doctors undertriaged. CONCLUSIONS: While LLMs and the LLM-based product ChatGPT do not yet match professionally trained raters, their best models' triage proficiency equals that of untrained ED doctors. In its current form, LLMs or ChatGPT thus did not demonstrate gold-standard performance in ED triage and, in the setting of this study, failed to significantly improve untrained doctors' triage when used as decision support. Notable performance enhancements in newer LLM versions over older ones hint at future improvements with further technological development and specific training.


Assuntos
Medicina de Emergência , Triagem , Triagem/métodos , Triagem/normas , Humanos , Medicina de Emergência/normas , Médicos/estatística & dados numéricos , Serviço Hospitalar de Emergência/normas , Idioma , Alemanha , Feminino
2.
NPJ Digit Med ; 7(1): 139, 2024 May 24.
Artigo em Inglês | MEDLINE | ID: mdl-38789620

RESUMO

The 2019 German Digital Healthcare Act introduced the Digital Health Application program, known in German as 'Digitale Gesundheitsanwendungen' (DiGA). The program has established a pioneering model for integrating Digital Therapeutics (DTx) into a healthcare system with scalable and effective reimbursement strategies. To date, the continuous upward trend enabled by this framework has resulted in more than 374,000 DiGA prescriptions, increasingly cementing its role in the German healthcare system. This perspective provides a synthesis of the DiGA program's evolution since its inception three years ago, highlighting trends regarding prescriptions and pricing as well as criticisms and identified shortcomings. It further discusses forthcoming legislative amendments, including the anticipated integration of higher-risk medical devices, which have the potential to significantly transform the program. Despite encountering challenges related to effectiveness, evidence requirements, and integration within the healthcare system, the DiGA program continues to evolve and serves as a seminal example for the integration of DTx, offering valuable insights for healthcare systems globally.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA