Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 50
Filter
Add more filters

Country/Region as subject
Publication year range
1.
J Biomed Inform ; 95: 103219, 2019 07.
Article in English | MEDLINE | ID: mdl-31150777

ABSTRACT

Clinical narratives are a valuable source of information for both patient care and biomedical research. Given the unstructured nature of medical reports, specific automatic techniques are required to extract relevant entities from such texts. In the natural language processing (NLP) community, this task is often addressed by using supervised methods. To develop such methods, both reliably-annotated corpora and elaborately designed features are needed. Despite the recent advances on corpora collection and annotation, research on multiple domains and languages is still limited. In addition, to compute the features required for supervised classification, suitable language- and domain-specific tools are needed. In this work, we propose a novel application of recurrent neural networks (RNNs) for event extraction from medical reports written in Italian. To train and evaluate the proposed approach, we annotated a corpus of 75 cardiology reports for a total of 4365 mentions of relevant events and their attributes (e.g., the polarity). For the annotation task, we developed specific annotation guidelines, which are provided together with this paper. The RNN-based classifier was trained on a training set including 3335 events (60 documents). The resulting model was integrated into an NLP pipeline that uses a dictionary lookup approach to search for relevant concepts inside the text. A test set of 1030 events (15 documents) was used to evaluate and compare different pipeline configurations. As a main result, using the RNN-based classifier instead of the dictionary lookup approach allowed increasing recall from 52.4% to 88.9%, and precision from 81.1% to 88.2%. Further, using the two methods in combination, we obtained final recall, precision, and F1 score of 91.7%, 88.6%, and 90.1%, respectively. These experiments indicate that integrating a well-performing RNN-based classifier with a standard knowledge-based approach can be a good strategy to extract information from clinical text in non-English languages.


Subject(s)
Data Mining/methods , Electronic Health Records , Natural Language Processing , Heart Diseases , Humans , Italy , Neural Networks, Computer , Semantics
2.
J Am Med Inform Assoc ; 31(4): 940-948, 2024 Apr 03.
Article in English | MEDLINE | ID: mdl-38261400

ABSTRACT

OBJECTIVE: Large language models (LLMs) have shown impressive ability in biomedical question-answering, but have not been adequately investigated for more specific biomedical applications. This study investigates ChatGPT family of models (GPT-3.5, GPT-4) in biomedical tasks beyond question-answering. MATERIALS AND METHODS: We evaluated model performance with 11 122 samples for two fundamental tasks in the biomedical domain-classification (n = 8676) and reasoning (n = 2446). The first task involves classifying health advice in scientific literature, while the second task is detecting causal relations in biomedical literature. We used 20% of the dataset for prompt development, including zero- and few-shot settings with and without chain-of-thought (CoT). We then evaluated the best prompts from each setting on the remaining dataset, comparing them to models using simple features (BoW with logistic regression) and fine-tuned BioBERT models. RESULTS: Fine-tuning BioBERT produced the best classification (F1: 0.800-0.902) and reasoning (F1: 0.851) results. Among LLM approaches, few-shot CoT achieved the best classification (F1: 0.671-0.770) and reasoning (F1: 0.682) results, comparable to the BoW model (F1: 0.602-0.753 and 0.675 for classification and reasoning, respectively). It took 78 h to obtain the best LLM results, compared to 0.078 and 0.008 h for the top-performing BioBERT and BoW models, respectively. DISCUSSION: The simple BoW model performed similarly to the most complex LLM prompting. Prompt engineering required significant investment. CONCLUSION: Despite the excitement around viral ChatGPT, fine-tuning for two fundamental biomedical natural language processing tasks remained the best strategy.


Subject(s)
Language , Natural Language Processing
3.
NPJ Digit Med ; 7(1): 6, 2024 Jan 11.
Article in English | MEDLINE | ID: mdl-38200151

ABSTRACT

Social determinants of health (SDoH) play a critical role in patient outcomes, yet their documentation is often missing or incomplete in the structured data of electronic health records (EHRs). Large language models (LLMs) could enable high-throughput extraction of SDoH from the EHR to support research and clinical care. However, class imbalance and data limitations present challenges for this sparsely documented yet critical information. Here, we investigated the optimal methods for using LLMs to extract six SDoH categories from narrative text in the EHR: employment, housing, transportation, parental status, relationship, and social support. The best-performing models were fine-tuned Flan-T5 XL for any SDoH mentions (macro-F1 0.71), and Flan-T5 XXL for adverse SDoH mentions (macro-F1 0.70). Adding LLM-generated synthetic data to training varied across models and architecture, but improved the performance of smaller Flan-T5 models (delta F1 + 0.12 to +0.23). Our best-fine-tuned models outperformed zero- and few-shot performance of ChatGPT-family models in the zero- and few-shot setting, except GPT4 with 10-shot prompting for adverse SDoH. Fine-tuned models were less likely than ChatGPT to change their prediction when race/ethnicity and gender descriptors were added to the text, suggesting less algorithmic bias (p < 0.05). Our models identified 93.8% of patients with adverse SDoH, while ICD-10 codes captured 2.0%. These results demonstrate the potential of LLMs in improving real-world evidence on SDoH and assisting in identifying patients who could benefit from resource support.

4.
JCO Clin Cancer Inform ; 7: e2300048, 2023 07.
Article in English | MEDLINE | ID: mdl-37506330

ABSTRACT

PURPOSE: Radiotherapy (RT) toxicities can impair survival and quality of life, yet remain understudied. Real-world evidence holds potential to improve our understanding of toxicities, but toxicity information is often only in clinical notes. We developed natural language processing (NLP) models to identify the presence and severity of esophagitis from notes of patients treated with thoracic RT. METHODS: Our corpus consisted of a gold-labeled data set of 1,524 clinical notes from 124 patients with lung cancer treated with RT, manually annotated for Common Terminology Criteria for Adverse Events (CTCAE) v5.0 esophagitis grade, and a silver-labeled data set of 2,420 notes from 1,832 patients from whom toxicity grades had been collected as structured data during clinical care. We fine-tuned statistical and pretrained Bidirectional Encoder Representations from Transformers-based models for three esophagitis classification tasks: task 1, no esophagitis versus grade 1-3; task 2, grade ≤1 versus >1; and task 3, no esophagitis versus grade 1 versus grade 2-3. Transferability was tested on 345 notes from patients with esophageal cancer undergoing RT. RESULTS: Fine-tuning of PubMedBERT yielded the best performance. The best macro-F1 was 0.92, 0.82, and 0.74 for tasks 1, 2, and 3, respectively. Selecting the most informative note sections during fine-tuning improved macro-F1 by ≥2% for all tasks. Silver-labeled data improved the macro-F1 by ≥3% across all tasks. For the esophageal cancer notes, the best macro-F1 was 0.73, 0.74, and 0.65 for tasks 1, 2, and 3, respectively, without additional fine-tuning. CONCLUSION: To our knowledge, this is the first effort to automatically extract esophagitis toxicity severity according to CTCAE guidelines from clinical notes. This provides proof of concept for NLP-based automated detailed toxicity monitoring in expanded domains.


Subject(s)
Esophageal Neoplasms , Esophagitis , Humans , Natural Language Processing , Quality of Life , Silver , Esophagitis/diagnosis , Esophagitis/etiology
5.
Int J Radiat Oncol Biol Phys ; 117(1): 262-273, 2023 09 01.
Article in English | MEDLINE | ID: mdl-36990288

ABSTRACT

PURPOSE: Real-world evidence for radiation therapy (RT) is limited because it is often documented only in the clinical narrative. We developed a natural language processing system for automated extraction of detailed RT events from text to support clinical phenotyping. METHODS AND MATERIALS: A multi-institutional data set of 96 clinician notes, 129 North American Association of Central Cancer Registries cancer abstracts, and 270 RT prescriptions from HemOnc.org was used and divided into train, development, and test sets. Documents were annotated for RT events and associated properties: dose, fraction frequency, fraction number, date, treatment site, and boost. Named entity recognition models for properties were developed by fine-tuning BioClinicalBERT and RoBERTa transformer models. A multiclass RoBERTa-based relation extraction model was developed to link each dose mention with each property in the same event. Models were combined with symbolic rules to create a hybrid end-to-end pipeline for comprehensive RT event extraction. RESULTS: Named entity recognition models were evaluated on the held-out test set with F1 results of 0.96, 0.88, 0.94, 0.88, 0.67, and 0.94 for dose, fraction frequency, fraction number, date, treatment site, and boost, respectively. The relation model achieved an average F1 of 0.86 when the input was gold-labeled entities. The end-to-end system F1 result was 0.81. The end-to-end system performed best on North American Association of Central Cancer Registries abstracts (average F1 0.90), which are mostly copy-paste content from clinician notes. CONCLUSIONS: We developed methods and a hybrid end-to-end system for RT event extraction, which is the first natural language processing system for this task. This system provides proof-of-concept for real-world RT data collection for research and is promising for the potential of natural language processing methods to support clinical care.


Subject(s)
Natural Language Processing , Neoplasms , Humans , Neoplasms/radiotherapy , Electronic Health Records
6.
JCO Clin Cancer Inform ; 7: e2200196, 2023 05.
Article in English | MEDLINE | ID: mdl-37235847

ABSTRACT

PURPOSE: There is an unmet need to empirically explore and understand drivers of cancer disparities, particularly social determinants of health. We explored natural language processing methods to automatically and empirically extract clinical documentation of social contexts and needs that may underlie disparities. METHODS: This was a retrospective analysis of 230,325 clinical notes from 5,285 patients treated with radiotherapy from 2007 to 2019. We compared linguistic features among White versus non-White, low-income insurance versus other insurance, and male versus female patients' notes. Log odds ratios with an informative Dirichlet prior were calculated to compare words over-represented in each group. A variational autoencoder topic model was applied, and topic probability was compared between groups. The presence of machine-learnable bias was explored by developing statistical and neural demographic group classifiers. RESULTS: Terms associated with varied social contexts and needs were identified for all demographic group comparisons. For example, notes of non-White and low-income insurance patients were over-represented with terms associated with housing and transportation, whereas notes of White and other insurance patients were over-represented with terms related to physical activity. Topic models identified a social history topic, and topic probability varied significantly between the demographic group comparisons. Classification models performed poorly at classifying notes of non-White and low-income insurance patients (F1 of 0.30 and 0.23, respectively). CONCLUSION: Exploration of linguistic differences in clinical notes between patients of different race/ethnicity, insurance status, and sex identified social contexts and needs in patients with cancer and revealed high-level differences in notes. Future work is needed to validate whether these findings may play a role in cancer disparities.


Subject(s)
Natural Language Processing , Neoplasms , Humans , Male , Female , Retrospective Studies , Social Environment , Neoplasms/diagnosis , Neoplasms/epidemiology , Neoplasms/therapy
7.
J Biomed Inform ; 45(3): 507-21, 2012 Jun.
Article in English | MEDLINE | ID: mdl-22343015

ABSTRACT

MOTIVATION: Expressions that refer to a real-world entity already mentioned in a narrative are often considered anaphoric. For example, in the sentence "The pain comes and goes," the expression "the pain" is probably referring to a previous mention of pain. Interpretation of meaning involves resolving the anaphoric reference: deciding which expression in the text is the correct antecedent of the referring expression, also called an anaphor. We annotated a set of 180 clinical reports (surgical pathology, radiology, discharge summaries, and emergency department) from two institutions to indicate all anaphor-antecedent pairs. OBJECTIVE: The objective of this study is to describe the characteristics of the corpus in terms of the frequency of anaphoric relations, the syntactic and semantic nature of the members of the pairs, and the types of anaphoric relations that occur. Understanding how anaphoric reference is exhibited in clinical reports is critical to developing reference resolution algorithms and to identifying peculiarities of clinical text that may alter the features and methodologies that will be successful for automated anaphora resolution. RESULTS: We found that anaphoric reference is prevalent in all types of clinical reports, that annotations of noun phrases, semantic type, and section headings may be especially important for automated resolution of anaphoric reference, and that separate modules for reference resolution may be required for different report types, different institutions, and different types of anaphors. Accurate resolution will probably require extensive domain knowledge-especially for pathology and radiology reports with more part/whole and set/subset relations. CONCLUSION: We hope researchers will leverage the annotations in this corpus to develop automated algorithms and will add to the annotations to generate a more extensive corpus.


Subject(s)
Electronic Health Records/standards , Semantics , Algorithms , Data Mining/methods , Humans
8.
J Biomed Inform ; 44(6): 1113-22, 2011 Dec.
Article in English | MEDLINE | ID: mdl-21856441

ABSTRACT

Coreference resolution is the task of determining linguistic expressions that refer to the same real-world entity in natural language. Research on coreference resolution in the general English domain dates back to 1960s and 1970s. However, research on coreference resolution in the clinical free text has not seen major development. The recent US government initiatives that promote the use of electronic health records (EHRs) provide opportunities to mine patient notes as more and more health care institutions adopt EHR. Our goal was to review recent advances in general purpose coreference resolution to lay the foundation for methodologies in the clinical domain, facilitated by the availability of a shared lexical resource of gold standard coreference annotations, the Ontology Development and Information Extraction (ODIE) corpus.


Subject(s)
Medical Informatics/methods , Natural Language Processing , Electronic Health Records , Humans , Information Storage and Retrieval , Linguistics
9.
Int J Radiat Oncol Biol Phys ; 110(3): 641-655, 2021 07 01.
Article in English | MEDLINE | ID: mdl-33545300

ABSTRACT

Natural language processing (NLP), which aims to convert human language into expressions that can be analyzed by computers, is one of the most rapidly developing and widely used technologies in the field of artificial intelligence. Natural language processing algorithms convert unstructured free text data into structured data that can be extracted and analyzed at scale. In medicine, this unlocking of the rich, expressive data within clinical free text in electronic medical records will help untap the full potential of big data for research and clinical purposes. Recent major NLP algorithmic advances have significantly improved the performance of these algorithms, leading to a surge in academic and industry interest in developing tools to automate information extraction and phenotyping from clinical texts. Thus, these technologies are poised to transform medical research and alter clinical practices in the future. Radiation oncology stands to benefit from NLP algorithms if they are appropriately developed and deployed, as they may enable advances such as automated inclusion of radiation therapy details into cancer registries, discovery of novel insights about cancer care, and improved patient data curation and presentation at the point of care. However, challenges remain before the full value of NLP is realized, such as the plethora of jargon specific to radiation oncology, nonstandard nomenclature, a lack of publicly available labeled data for model development, and interoperability limitations between radiation oncology data silos. Successful development and implementation of high quality and high value NLP models for radiation oncology will require close collaboration between computer scientists and the radiation oncology community. Here, we present a primer on artificial intelligence algorithms in general and NLP algorithms in particular; provide guidance on how to assess the performance of such algorithms; review prior research on NLP algorithms for oncology; and describe future avenues for NLP in radiation oncology research and clinics.


Subject(s)
Natural Language Processing , Radiation Oncology , Electronic Health Records , Humans
10.
J Digit Imaging ; 23(2): 119-32, 2010 Apr.
Article in English | MEDLINE | ID: mdl-19484309

ABSTRACT

Information in electronic medical records is often in an unstructured free-text format. This format presents challenges for expedient data retrieval and may fail to convey important findings. Natural language processing (NLP) is an emerging technique for rapid and efficient clinical data retrieval. While proven in disease detection, the utility of NLP in discerning disease progression from free-text reports is untested. We aimed to (1) assess whether unstructured radiology reports contained sufficient information for tumor status classification; (2) develop an NLP-based data extraction tool to determine tumor status from unstructured reports; and (3) compare NLP and human tumor status classification outcomes. Consecutive follow-up brain tumor magnetic resonance imaging reports (2000--2007) from a tertiary center were manually annotated using consensus guidelines on tumor status. Reports were randomized to NLP training (70%) or testing (30%) groups. The NLP tool utilized a support vector machines model with statistical and rule-based outcomes. Most reports had sufficient information for tumor status classification, although 0.8% did not describe status despite reference to prior examinations. Tumor size was unreported in 68.7% of documents, while 50.3% lacked data on change magnitude when there was detectable progression or regression. Using retrospective human classification as the gold standard, NLP achieved 80.6% sensitivity and 91.6% specificity for tumor status determination (mean positive predictive value, 82.4%; negative predictive value, 92.0%). In conclusion, most reports contained sufficient information for tumor status determination, though variable features were used to describe status. NLP demonstrated good accuracy for tumor status classification and may have novel application for automated disease status classification from electronic databases.


Subject(s)
Electronic Data Processing/statistics & numerical data , Information Storage and Retrieval/methods , Magnetic Resonance Imaging/standards , Natural Language Processing , Neoplasms/diagnosis , Radiology Information Systems , Female , Humans , Magnetic Resonance Imaging/methods , Male , Medical Records Systems, Computerized , Reproducibility of Results , Sensitivity and Specificity
11.
JAMIA Open ; 3(3): 413-421, 2020 Oct.
Article in English | MEDLINE | ID: mdl-33215076

ABSTRACT

OBJECTIVE: To advance use of real-world data (RWD) for pharmacovigilance, we sought to integrate a high-sensitivity natural language processing (NLP) pipeline for detecting potential adverse drug events (ADEs) with easily interpretable output for high-efficiency human review and adjudication of true ADEs. MATERIALS AND METHODS: The adverse drug event presentation and tracking (ADEPT) system employs an open source NLP pipeline to identify in clinical notes mentions of medications and signs and symptoms potentially indicative of ADEs. ADEPT presents the output to human reviewers by highlighting these drug-event pairs within the context of the clinical note. To measure incidence of seizures associated with sildenafil, we applied ADEPT to 149 029 notes for 982 patients with pediatric pulmonary hypertension. RESULTS: Of 416 patients identified as taking sildenafil, NLP found 72 [17%, 95% confidence interval (CI) 14-21] with seizures as a potential ADE. Upon human review and adjudication, only 4 (0.96%, 95% CI 0.37-2.4) patients with seizures were determined to have true ADEs. Reviewers using ADEPT required a median of 89 s (interquartile range 57-142 s) per patient to review potential ADEs. DISCUSSION: ADEPT combines high throughput NLP to increase sensitivity of ADE detection and human review, to increase specificity by differentiating true ADEs from signs and symptoms related to comorbidities, effects of other medications, or other confounders. CONCLUSION: ADEPT is a promising tool for creating gold standard, patient-level labels for advancing NLP-based pharmacovigilance. ADEPT is a potentially time savings platform for computer-assisted pharmacovigilance based on RWD.

12.
J Am Med Inform Assoc ; 27(2): 294-300, 2020 02 01.
Article in English | MEDLINE | ID: mdl-31769835

ABSTRACT

OBJECTIVE: Real-world data (RWD) are increasingly used for pharmacoepidemiology and regulatory innovation. Our objective was to compare adverse drug event (ADE) rates determined from two RWD sources, electronic health records and administrative claims data, among children treated with drugs for pulmonary hypertension. MATERIALS AND METHODS: Textual mentions of medications and signs/symptoms that may represent ADEs were identified in clinical notes using natural language processing. Diagnostic codes for the same signs/symptoms were identified in our electronic data warehouse for the patients with textual evidence of taking pulmonary hypertension-targeted drugs. We compared rates of ADEs identified in clinical notes to those identified from diagnostic code data. In addition, we compared putative ADE rates from clinical notes to those from a healthcare claims dataset from a large, national insurer. RESULTS: Analysis of clinical notes identified up to 7-fold higher ADE rates than those ascertained from diagnostic codes. However, certain ADEs (eg, hearing loss) were more often identified in diagnostic code data. Similar results were found when ADE rates ascertained from clinical notes and national claims data were compared. DISCUSSION: While administrative claims and clinical notes are both increasingly used for RWD-based pharmacovigilance, ADE rates substantially differ depending on data source. CONCLUSION: Pharmacovigilance based on RWD may lead to discrepant results depending on the data source analyzed. Further work is needed to confirm the validity of identified ADEs, to distinguish them from disease effects, and to understand tradeoffs in sensitivity and specificity between data sources.


Subject(s)
Current Procedural Terminology , Drug-Related Side Effects and Adverse Reactions , Electronic Health Records , Hypertension, Pulmonary/drug therapy , Natural Language Processing , Child , Child, Preschool , Female , Humans , Insurance, Health , Male , Pharmacovigilance , Regression Analysis , Retrospective Studies
13.
JAMA Oncol ; 10(4): 538-539, 2024 Apr 01.
Article in English | MEDLINE | ID: mdl-38358777
14.
Cancer Res ; 79(21): 5463-5470, 2019 11 01.
Article in English | MEDLINE | ID: mdl-31395609

ABSTRACT

Current models for correlating electronic medical records with -omics data largely ignore clinical text, which is an important source of phenotype information for patients with cancer. This data convergence has the potential to reveal new insights about cancer initiation, progression, metastasis, and response to treatment. Insights from this real-world data will catalyze clinical care, research, and regulatory activities. Natural language processing (NLP) methods are needed to extract these rich cancer phenotypes from clinical text. Here, we review the advances of NLP and information extraction methods relevant to oncology based on publications from PubMed as well as NLP and machine learning conference proceedings in the last 3 years. Given the interdisciplinary nature of the fields of oncology and information extraction, this analysis serves as a critical trail marker on the path to higher fidelity oncology phenotypes from real-world data.


Subject(s)
Data Mining/methods , Medical Oncology/methods , Electronic Health Records , Humans , Machine Learning , Natural Language Processing , Phenotype
15.
Acad Pediatr ; 19(5): 589-598, 2019 07.
Article in English | MEDLINE | ID: mdl-30470563

ABSTRACT

OBJECTIVE: Comparison of readmission rates requires adjustment for case-mix (ie, differences in patient populations), but previously only claims data were available for this purpose. We examined whether incorporation of relatively readily available clinical data improves prediction of pediatric readmissions and thus might enhance case-mix adjustment. METHODS: We examined 30-day readmissions using claims and electronic health record data for patients ≤18 years and 29 days of age who were admitted to 3 children's hospitals from February 2011 to February 2014. Using the Pediatric All-Condition Readmission Measure and starting with a model including age, gender, chronic conditions, and primary diagnosis, we examined whether the addition of initial vital sign and laboratory data improved model performance. We employed machine learning to evaluate the same variables, using the L2-regularized logistic regression with cost-sensitive learning and convolutional neural network. RESULTS: Controlling for the core model variables, low red blood cell count and mean corpuscular hemoglobin concentration and high red cell distribution width were associated with greater readmission risk, as were certain interactions between laboratory and chronic condition variables. However, the C-statistic (0.722 vs 0.713) and McFadden's pseudo R2 (0.085 vs 0.076) for this and the core model were similar, suggesting minimal improvement in performance. In machine learning analyses, the F-measure (harmonic mean of sensitivity and positive predictive value) was similar for the best-performing model (containing all variables) and core model (0.250 vs 0.243). CONCLUSIONS: Readily available clinical variables do not meaningfully improve the prediction of pediatric readmissions and would be unlikely to enhance case-mix adjustment unless their distributions varied widely across hospitals.


Subject(s)
Patient Readmission , Quality Indicators, Health Care , Adolescent , Child , Child, Preschool , Female , Humans , Male , Risk Adjustment , Risk Assessment , Risk Factors , Socioeconomic Factors , Time Factors
17.
J Am Med Inform Assoc ; 15(1): 25-8, 2008.
Article in English | MEDLINE | ID: mdl-17947622

ABSTRACT

This article describes our system entry for the 2006 I2B2 contest "Challenges in Natural Language Processing for Clinical Data" for the task of identifying the smoking status of patients. Our system makes the simplifying assumption that patient-level smoking status determination can be achieved by accurately classifying individual sentences from a patient's record. We created our system with reusable text analysis components built on the Unstructured Information Management Architecture and Weka. This reuse of code minimized the development effort related specifically to our smoking status classifier. We report precision, recall, F-score, and 95% exact confidence intervals for each metric. Recasting the classification task for the sentence level and reusing code from other text analysis projects allowed us to quickly build a classification system that performs with a system F-score of 92.64 based on held-out data tests and of 85.57 on the formal evaluation data. Our general medical natural language engine is easily adaptable to a real-world medical informatics application. Some of the limitations as applied to the use-case are negation detection and temporal resolution.


Subject(s)
Classification/methods , Medical Records Systems, Computerized , Natural Language Processing , Smoking , Databases, Factual , Humans
18.
J Biomed Inform ; 41(6): 1088-100, 2008 Dec.
Article in English | MEDLINE | ID: mdl-18375190

ABSTRACT

The aim of this study is to explore the word sense disambiguation (WSD) problem across two biomedical domains-biomedical literature and clinical notes. A supervised machine learning technique was used for the WSD task. One of the challenges addressed is the creation of a suitable clinical corpus with manual sense annotations. This corpus in conjunction with the WSD set from the National Library of Medicine provided the basis for the evaluation of our method across multiple domains and for the comparison of our results to published ones. Noteworthy is that only 20% of the most relevant ambiguous terms within a domain overlap between the two domains, having more senses associated with them in the clinical space than in the biomedical literature space. Experimentation with 28 different feature sets rendered a system achieving an average F-score of 0.82 on the clinical data and 0.86 on the biomedical literature.


Subject(s)
Language , Algorithms , Artificial Intelligence , Information Services
19.
Comput Inform Nurs ; 26(5): 282-9, 2008.
Article in English | MEDLINE | ID: mdl-18769183

ABSTRACT

Although an unambiguous and consistent representation is the foundation of data reuse, a locally developed documentation system such as nursing flowsheets often fails to meet the requirement. This article presents the domain modeling process of the ICU nursing flowsheet to clarify the meaning that its contents represent and the lessons learned during the activity. This study has been done as a first step toward reusing the data documented in a computerized nursing flowsheet for an algorithmic decision making. Following the ontology development processes proposed by other researchers, a conceptual model was developed using Protégé. Then, the existing information model was refined by fully specifying the embedded information structures and by establishing linkages to the conceptual model at the finest-grained concept level. Domain knowledge that the experienced nurses provided was critical to correctly interpret the meaning of the flowsheet contents as well as to verify the newly developed models. This study reassured the importance of the roles of a nurse informaticist to develop a computerized nursing documentation system that accurately represents the information needs in nursing practice.


Subject(s)
Critical Care , Data Collection/methods , Documentation/methods , Medical Records Systems, Computerized/organization & administration , Nursing Records , Vocabulary, Controlled , Algorithms , Artificial Intelligence , Computer Simulation , Critical Care/organization & administration , Decision Trees , Forms and Records Control , Humans , Minnesota , Models, Nursing , Numerical Analysis, Computer-Assisted , Nursing Assessment , Nursing Diagnosis , Nursing Evaluation Research , Nursing Research/organization & administration , Software Design , User-Computer Interface
20.
JAMA Oncol ; 9(10): 1459-1462, 2023 Oct 01.
Article in English | MEDLINE | ID: mdl-37615976

ABSTRACT

This survey study examines the performance of a large language model chatbot in providing cancer treatment recommendations that are concordant with National Comprehensive Cancer Network guidelines.


Subject(s)
Artificial Intelligence , Neoplasms , Humans , Neoplasms/therapy
SELECTION OF CITATIONS
SEARCH DETAIL