Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 137
Filter
1.
Article in English | MEDLINE | ID: mdl-39356433

ABSTRACT

Given the widespread presence of micropollutants in urban water systems, it is imperative to gain a comprehensive understanding of their degradation pathways. This paper focuses on sulfamethoxazole (SMX) as a model molecule due to its extensive study, aiming to elucidate its degradation pathways in biological (BIO) and oxidative (AOP) processes. Numerous reaction pathways are outlined in scientific papers. However, a significant deficiency in current methodologies has led to the development of a novel meta-analytical approach, seeking consensus among researchers by synthesizing data from studies characterized by their heterogeneity and contradictions. As an innovative alternative, probabilistic graphical models such as Bayesian networks (BNs) could illuminate the relationships and dependencies between various transformation products, providing a holistic view of the degradation process. Based on the analysis of an extensive bibliography gathering more than 45 articles for more than 140 molecules and 177 reaction pathways, this study proposes a meta-analysis methodology based on Bayesian networks.

2.
Brief Bioinform ; 25(6)2024 Sep 23.
Article in English | MEDLINE | ID: mdl-39325460

ABSTRACT

Drug repurposing has emerged as a effective and efficient strategy to identify new treatments for a variety of diseases. One of the most effective approaches for discovering potential new drug candidates involves the utilization of Knowledge Graphs (KGs). This review comprehensively explores some of the most prominent KGs, detailing their structure, data sources, and how they facilitate the repurposing of drugs. In addition to KGs, this paper delves into various artificial intelligence techniques that enhance the process of drug repurposing. These methods not only accelerate the identification of viable drug candidates but also improve the precision of predictions by leveraging complex datasets and advanced algorithms. Furthermore, the importance of explainability in drug repurposing is emphasized. Explainability methods are crucial as they provide insights into the reasoning behind AI-generated predictions, thereby increasing the trustworthiness and transparency of the repurposing process. We will discuss several techniques that can be employed to validate these predictions, ensuring that they are both reliable and understandable.


Subject(s)
Drug Repositioning , Drug Repositioning/methods , Humans , Algorithms , Artificial Intelligence , Databases, Factual , Computational Biology/methods
3.
bioRxiv ; 2024 Aug 19.
Article in English | MEDLINE | ID: mdl-39229008

ABSTRACT

The rapid expansion of multi-omics data has transformed biological research, offering unprecedented opportunities to explore complex genomic relationships across diverse organisms. However, the vast volume and heterogeneity of these datasets presents significant challenges for analyses. Here we introduce SocialGene, a comprehensive software suite designed to collect, analyze, and organize multi-omics data into structured knowledge graphs, with the ability to handle small projects to repository-scale analyses. Originally developed to enhance genome mining for natural product drug discovery, SocialGene has been effective across various applications, including functional genomics, evolutionary studies, and systems biology. SocialGene's concerted Python and Nextflow libraries streamline data ingestion, manipulation, aggregation, and analysis, culminating in a custom Neo4j database. The software not only facilitates the exploration of genomic synteny but also provides a foundational knowledge graph supporting the integration of additional diverse datasets and the development of advanced search engines and analyses. This manuscript introduces some of SocialGene's capabilities through brief case studies including targeted genome mining for drug discovery, accelerated searches for similar and distantly related biosynthetic gene clusters in biobank-available organisms, integration of chemical and analytical data, and more. SocialGene is free, open-source, MIT-licensed, designed for adaptability and extension, and available from github.com/socialgene.

4.
Neural Netw ; 179: 106583, 2024 Nov.
Article in English | MEDLINE | ID: mdl-39111163

ABSTRACT

Entity alignment is a crucial task in knowledge graphs, aiming to match corresponding entities from different knowledge graphs. Due to the scarcity of pre-aligned entities in real-world scenarios, research focused on unsupervised entity alignment has become more popular. However, current unsupervised entity alignment methods suffer from a lack of informative entity guidance, hindering their ability to accurately predict challenging entities with similar names and structures. To solve these problems, we present an unsupervised multi-view contrastive learning framework with an attention-based reranking strategy for entity alignment, named AR-Align. In AR-Align, two kinds of data augmentation methods are employed to provide a complementary view for neighborhood and attribute, respectively. Next, a multi-view contrastive learning method is introduced to reduce the semantic gap between different views of the augmented entities. Moreover, an attention-based reranking strategy is proposed to rerank the hard entities through calculating their weighted sum of embedding similarities on different structures. Experimental results indicate that AR-Align outperforms most both supervised and unsupervised state-of-the-art methods on three benchmark datasets.


Subject(s)
Unsupervised Machine Learning , Attention , Semantics , Neural Networks, Computer , Algorithms , Humans
5.
PeerJ Comput Sci ; 10: e2010, 2024.
Article in English | MEDLINE | ID: mdl-39145203

ABSTRACT

Personalized learning resource recommendations may help resolve the difficulties of online education that include learning mazes and information overload. However, existing personalized learning resource recommendation algorithms have shortcomings such as low accuracy and low efficiency. This study proposes a deep recommendation system algorithm based on a knowledge graph (D-KGR) that includes four data processing units. These units are the recommendation unit (RS unit), the knowledge graph feature representation unit (KGE unit), the cross compression unit (CC unit), and the feature extraction unit (FE unit). This model integrates technologies including the knowledge graph, deep learning, neural network, and data mining. It introduces cross compression in the feature learning process of the knowledge graph and predicts user attributes. Multimodal technology is used to optimize the process of project attribute processing; text type attributes, multivalued type attributes, and other type attributes are processed separately to reconstruct the knowledge graph. A convolutional neural network algorithm is introduced in the reconstruction process to optimize the data feature qualities. Experimental analysis was conducted from two aspects of algorithm efficiency and accuracy, and the particle swarm optimization, neural network, and knowledge graph algorithms were compared. Several tests showed that the deep recommendation system algorithm had obvious advantages when the number of learning resources and users exceeded 1,000. It has the ability to integrate systems such as the particle swarm optimization iterative classification, neural network intelligent simulation, and low resource consumption. It can quickly process massive amounts of information data, reduce algorithm complexity and requires less time and had lower costs. Our algorithm also has better efficiency and accuracy.

6.
Heliyon ; 10(12): e32479, 2024 Jun 30.
Article in English | MEDLINE | ID: mdl-39183851

ABSTRACT

Numerous methods and pipelines have recently emerged for the automatic extraction of knowledge graphs from documents such as scientific publications and patents. However, adapting these methods to incorporate alternative text sources like micro-blogging posts and news has proven challenging as they struggle to model open-domain entities and relations, typically found in these sources. In this paper, we propose an enhanced information extraction pipeline tailored to the extraction of a knowledge graph comprising open-domain entities from micro-blogging posts on social media platforms. Our pipeline leverages dependency parsing and classifies entity relations in an unsupervised manner through hierarchical clustering over word embeddings. We provide a use case on extracting semantic triples from a corpus of 100 thousand tweets about digital transformation and publicly release the generated knowledge graph. On the same dataset, we conduct two experimental evaluations, showing that the system produces triples with precision over 95% and outperforms similar pipelines of around 5% in terms of precision, while generating a comparatively higher number of triples.

7.
Stud Health Technol Inform ; 316: 1463-1464, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176479

ABSTRACT

This paper presents a versatile solution to formally represent the contents of electronic health records. It is based on the knowledge graph paradigm, and semantic web standards RDF and OWL. It employs the established semantic standards SNOMED CT and FHIR, which warrant international interoperability. A graph-based form is not only useful to feed different target visualizations, but it can also be subject to AI-powered services such as (fuzzy) retrieval and summarization.


Subject(s)
Electronic Health Records , Humans , Semantic Web , Systematized Nomenclature of Medicine , Precision Medicine , Semantics , Computer Graphics , Information Storage and Retrieval
8.
Stud Health Technol Inform ; 316: 1406-1410, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176643

ABSTRACT

Real-world data (RWD) (i.e., data from Electronic Healthcare Records - EHRs, ePrescription systems, patient registries, etc.) gain increasing attention as they could support observational studies on a large scale. OHDSI is one of the most prominent initiatives regarding the harmonization of RWD and the development of relevant tools via the use of a common data model, OMOP-CDM. OMOP-CDM is a crucial step towards syntactic and semantic data interoperability. Still, OMOP-CDM is based on a typical relational database format, and thus, the vision of a fully connected semantically enriched model is not fully realized. This work presents an open-source effort to map the OMOP-CDM model and the data it hosts, to an ontological model using RDF to support the FAIRness of RWD and their interlinking with Linked Open Data (LOD) towards the vision of the Semantic Web.


Subject(s)
Electronic Health Records , Semantic Web , Humans , Semantics , Medical Record Linkage/methods
9.
Stud Health Technol Inform ; 316: 575-579, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176807

ABSTRACT

Developing novel predictive models with complex biomedical information is challenging due to various idiosyncrasies related to heterogeneity, standardization or sparseness of the data. We previously introduced a person-centric ontology to organize information about individual patients, and a representation learning framework to extract person-centric knowledge graphs (PKGs) and to train Graph Neural Networks (GNNs). In this paper, we propose a systematic approach to examine the results of GNN models trained with both structured and unstructured information from the MIMIC-III dataset. Through ablation studies on different clinical, demographic, and social data, we show the robustness of this approach in identifying predictive features in PKGs for the task of readmission prediction.


Subject(s)
Neural Networks, Computer , Humans , Patient Readmission
10.
Stud Health Technol Inform ; 316: 1933-1937, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176870

ABSTRACT

Biomedical data analysis and visualization often demand data experts for each unique health event. There is a clear lack of automatic tools for semantic visualization of the spread of health risks through biomedical data. Illnesses such as coronavirus disease (COVID-19) and Monkeypox spread rampantly around the world before governments could make decisions based on the analysis of such data. We propose the design of a knowledge graph (KG) for spatio-temporal tracking of public health event propagation. To achieve this, we propose the specialization of the Core Propagation Phenomenon Ontology (PropaPhen) into a health-related propagation phenomenon domain ontology. Data from the UMLS and OpenStreetMaps are suggested for instantiating the proposed knowledge graph. Finally, the results of a use case on COVID-19 data from the World Health Organization are analyzed to evaluate the possibilities of our approach.


Subject(s)
COVID-19 , Humans , SARS-CoV-2 , Biological Ontologies , Unified Medical Language System
11.
Stud Health Technol Inform ; 316: 873-874, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176931

ABSTRACT

Artificial Intelligence (AI), particularly Machine Learning (ML), has gained attention for its potential in various domains. However, approaches integrating symbolic AI with ML on Knowledge Graphs have not gained significant focus yet. We argue that exploiting RDF/OWL semantics while conducting ML could provide useful insights. We present a use case using signaling pathways from the Reactome database to explore drug safety. Promising outcomes suggest the need for further investigation and collaboration with domain experts.


Subject(s)
Machine Learning , Humans , Drug-Related Side Effects and Adverse Reactions , Semantics , Signal Transduction , Databases, Factual , Adverse Drug Reaction Reporting Systems
12.
Stud Health Technol Inform ; 316: 1018-1022, 2024 Aug 22.
Article in English | MEDLINE | ID: mdl-39176963

ABSTRACT

Health literacy empowers people to access, understand and apply health information to effectively manage their own health and to be an active participant in healthcare decisions. In this paper we propose a conceptual model for cognitive factors affecting health literacy and related socioeconomic aspects. Then we develop the HEALIE Knowledge Graph to represent the model, drawing from various medical ontologies, resources, and insights from domain experts. Finally, we combine the Knowledge Graph with a Large Language Model to generate personalised medical content and showcase the results through an example.


Subject(s)
Health Literacy , Humans , Patient Participation , Precision Medicine , Natural Language Processing , Empowerment
13.
Clin Ther ; 46(7): 544-554, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38981792

ABSTRACT

PURPOSE: To critically assess the role and added value of knowledge graphs in pharmacovigilance, focusing on their ability to predict adverse drug reactions. METHODS: A systematic scoping review was conducted in which detailed information, including objectives, technology, data sources, methodology, and performance metrics, were extracted from a set of peer-reviewed publications reporting the use of knowledge graphs to support pharmacovigilance signal detection. FINDINGS: The review, which included 47 peer-reviewed articles, found knowledge graphs were utilized for detecting/predicting single-drug adverse reactions and drug-drug interactions, with variable reported performance and sparse comparisons to legacy methods. IMPLICATIONS: Research to date suggests that knowledge graphs have the potential to augment predictive signal detection in pharmacovigilance, but further research using more reliable reference sets of adverse drug reactions and comparison with legacy pharmacovigilance methods are needed to more clearly define best practices and to establish their place in holistic pharmacovigilance systems.


Subject(s)
Adverse Drug Reaction Reporting Systems , Drug-Related Side Effects and Adverse Reactions , Pharmacovigilance , Humans , Drug-Related Side Effects and Adverse Reactions/epidemiology , Adverse Drug Reaction Reporting Systems/statistics & numerical data , Drug Interactions
14.
Neural Netw ; 179: 106516, 2024 Nov.
Article in English | MEDLINE | ID: mdl-39003981

ABSTRACT

Temporal Knowledge Graphs (TKGs) enable effective modeling of knowledge dynamics and event evolution, facilitating deeper insights and analysis into temporal information. Recently, extrapolation of TKG reasoning has attracted great significance due to its remarkable ability to capture historical correlations and predict future events. Existing studies of extrapolation aim mainly at encoding the structural and temporal semantics based on snapshot sequences, which contain graph aggregators for the association within snapshots and recurrent units for the evolution. However, these methods are limited to modeling long-distance history, as they primarily focus on capturing temporal correlations over shorter periods. Besides, a few approaches rely on compiling historical repetitive statistics of TKGs for predicting future facts. But they often overlook explicit interactions in the graph structure among concurrent events. To address these issues, we propose a PotentiaL concurrEnt Aggregation and contraStive learnING (PLEASING) method for TKG extrapolation. PLEASING is a two-step reasoning framework that effectively leverages the historical and potential features of TKGs. It includes two encoders for historical and global events with an adaptive gated mechanism, acquiring predictions with appropriate weight of the two aspects. Specifically, PLEASING constructs two auxiliary graphs to capture temporal interaction among timestamps and correlations among potential concurrent events, respectively, enabling a holistic investigation of temporal characteristics and future potential possibilities in TKGs. Furthermore, PLEASING incorporates contrastive learning to strengthen its capacity to identify whether queries are related to history. Extensive experiments on seven benchmark datasets demonstrate the state-of-the-art performances of PLEASING and its comprehensive ability to model TKG semantics.


Subject(s)
Neural Networks, Computer , Humans , Semantics , Algorithms , Knowledge , Time Factors , Machine Learning
15.
Int J Digit Libr ; 25(2): 273-285, 2024.
Article in English | MEDLINE | ID: mdl-38948004

ABSTRACT

Due to the growing number of scholarly publications, finding relevant articles becomes increasingly difficult. Scholarly knowledge graphs can be used to organize the scholarly knowledge presented within those publications and represent them in machine-readable formats. Natural language processing (NLP) provides scalable methods to automatically extract knowledge from articles and populate scholarly knowledge graphs. However, NLP extraction is generally not sufficiently accurate and, thus, fails to generate high granularity quality data. In this work, we present TinyGenius, a methodology to validate NLP-extracted scholarly knowledge statements using microtasks performed with crowdsourcing. TinyGenius is employed to populate a paper-centric knowledge graph, using five distinct NLP methods. We extend our previous work of the TinyGenius methodology in various ways. Specifically, we discuss the NLP tasks in more detail and include an explanation of the data model. Moreover, we present a user evaluation where participants validate the generated NLP statements. The results indicate that employing microtasks for statement validation is a promising approach despite the varying participant agreement for different microtasks.

16.
J Med Internet Res ; 26: e54263, 2024 Jul 05.
Article in English | MEDLINE | ID: mdl-38968598

ABSTRACT

BACKGROUND: The medical knowledge graph provides explainable decision support, helping clinicians with prompt diagnosis and treatment suggestions. However, in real-world clinical practice, patients visit different hospitals seeking various medical services, resulting in fragmented patient data across hospitals. With data security issues, data fragmentation limits the application of knowledge graphs because single-hospital data cannot provide complete evidence for generating precise decision support and comprehensive explanations. It is important to study new methods for knowledge graph systems to integrate into multicenter, information-sensitive medical environments, using fragmented patient records for decision support while maintaining data privacy and security. OBJECTIVE: This study aims to propose an electronic health record (EHR)-oriented knowledge graph system for collaborative reasoning with multicenter fragmented patient medical data, all the while preserving data privacy. METHODS: The study introduced an EHR knowledge graph framework and a novel collaborative reasoning process for utilizing multicenter fragmented information. The system was deployed in each hospital and used a unified semantic structure and Observational Medical Outcomes Partnership (OMOP) vocabulary to standardize the local EHR data set. The system transforms local EHR data into semantic formats and performs semantic reasoning to generate intermediate reasoning findings. The generated intermediate findings used hypernym concepts to isolate original medical data. The intermediate findings and hash-encrypted patient identities were synchronized through a blockchain network. The multicenter intermediate findings were collaborated for final reasoning and clinical decision support without gathering original EHR data. RESULTS: The system underwent evaluation through an application study involving the utilization of multicenter fragmented EHR data to alert non-nephrology clinicians about overlooked patients with chronic kidney disease (CKD). The study covered 1185 patients in nonnephrology departments from 3 hospitals. The patients visited at least two of the hospitals. Of these, 124 patients were identified as meeting CKD diagnosis criteria through collaborative reasoning using multicenter EHR data, whereas the data from individual hospitals alone could not facilitate the identification of CKD in these patients. The assessment by clinicians indicated that 78/91 (86%) patients were CKD positive. CONCLUSIONS: The proposed system was able to effectively utilize multicenter fragmented EHR data for clinical application. The application study showed the clinical benefits of the system with prompt and comprehensive decision support.


Subject(s)
Decision Support Systems, Clinical , Electronic Health Records , Humans
17.
Article in English | MEDLINE | ID: mdl-38946554

ABSTRACT

BACKGROUND: Acute hepatic porphyria (AHP) is a group of rare but treatable conditions associated with diagnostic delays of 15 years on average. The advent of electronic health records (EHR) data and machine learning (ML) may improve the timely recognition of rare diseases like AHP. However, prediction models can be difficult to train given the limited case numbers, unstructured EHR data, and selection biases intrinsic to healthcare delivery. We sought to train and characterize models for identifying patients with AHP. METHODS: This diagnostic study used structured and notes-based EHR data from 2 centers at the University of California, UCSF (2012-2022) and UCLA (2019-2022). The data were split into 2 cohorts (referral and diagnosis) and used to develop models that predict (1) who will be referred for testing of acute porphyria, among those who presented with abdominal pain (a cardinal symptom of AHP), and (2) who will test positive, among those referred. The referral cohort consisted of 747 patients referred for testing and 99 849 contemporaneous patients who were not. The diagnosis cohort consisted of 72 confirmed AHP cases and 347 patients who tested negative. The case cohort was 81% female and 6-75 years old at the time of diagnosis. Candidate models used a range of architectures. Feature selection was semi-automated and incorporated publicly available data from knowledge graphs. Our primary outcome was the F-score on an outcome-stratified test set. RESULTS: The best center-specific referral models achieved an F-score of 86%-91%. The best diagnosis model achieved an F-score of 92%. To further test our model, we contacted 372 current patients who lack an AHP diagnosis but were predicted by our models as potentially having it (≥10% probability of referral, ≥50% of testing positive). However, we were only able to recruit 10 of these patients for biochemical testing, all of whom were negative. Nonetheless, post hoc evaluations suggested that these models could identify 71% of cases earlier than their diagnosis date, saving 1.2 years. CONCLUSIONS: ML can reduce diagnostic delays in AHP and other rare diseases. Robust recruitment strategies and multicenter coordination will be needed to validate these models before they can be deployed.

18.
Sci Rep ; 14(1): 16587, 2024 07 18.
Article in English | MEDLINE | ID: mdl-39025897

ABSTRACT

Drug repurposing aims to find new therapeutic applications for existing drugs in the pharmaceutical market, leading to significant savings in time and cost. The use of artificial intelligence and knowledge graphs to propose repurposing candidates facilitates the process, as large amounts of data can be processed. However, it is important to pay attention to the explainability needed to validate the predictions. We propose a general architecture to understand several explainable methods for graph completion based on knowledge graphs and design our own architecture for drug repurposing. We present XG4Repo (eXplainable Graphs for Repurposing), a framework that takes advantage of the connectivity of any biomedical knowledge graph to link compounds to the diseases they can treat. Our method allows methapaths of different types and lengths, which are automatically generated and optimised based on data. XG4Repo focuses on providing meaningful explanations to the predictions, which are based on paths from compounds to diseases. These paths include nodes such as genes, pathways, side effects, or anatomies, so they provide information about the targets and other characteristics of the biomedical mechanism that link compounds and diseases. Paths make predictions interpretable for experts who can validate them and use them in further research on drug repurposing. We also describe three use cases where we analyse new uses for Epirubicin, Paclitaxel, and Predinisone and present the paths that support the predictions.


Subject(s)
Drug Repositioning , Drug Repositioning/methods , Humans , Artificial Intelligence , Algorithms
19.
Heliyon ; 10(13): e33833, 2024 Jul 15.
Article in English | MEDLINE | ID: mdl-39050435

ABSTRACT

Major depressive disorder (MDD) is a debilitating mental health condition that poses significant risks and burdens. Resting-state functional magnetic resonance imaging (fMRI) has emerged as a promising tool in investigating the neural mechanisms underlying MDD. However, a comprehensive bibliometric analysis of resting-state fMRI in MDD is currently lacking. Here, we aimed to thoroughly explore the trends and frontiers of resting-state fMRI in MDD research. The relevant publications were retrieved from the Web of Science database for the period between 1998 and 2022, and the CiteSpace software was employed to identify the influence of authors, institutions, countries/regions, and the latest research trends. A total of 1501 publications met the search criteria, revealing a gradual increase in the number of annual publications over the years. China contributed the largest publication output, accounting for the highest percentage among all countries. Particularly, the University of Electronic Science and Technology of China, Capital Medical University, and Harvard Medical School were identified as key institutions that have made substantial contributions to this growth. Neuroimage, Biological Psychiatry, Journal of Affective Disorders, and Proceedings of the National Academy of Sciences of the United States of America are among the influential journals in the field of resting-state fMRI research in MDD. Burst keywords analysis suggest the emerging research frontiers in this field are characterized by prominent keywords such as dynamic functional connectivity, cognitive control network, transcranial brain stimulation, and childhood trauma. Overall, our study provides a systematic overview into the historical development, current status, and future trends of resting-state fMRI in MDD, thus offering a useful guide for researchers to plan their future research.

20.
Neural Netw ; 178: 106468, 2024 Oct.
Article in English | MEDLINE | ID: mdl-38943862

ABSTRACT

Knowledge graph reasoning, vital for addressing incompleteness and supporting applications, faces challenges with the continuous growth of graphs. To address this challenge, several inductive reasoning models for encoding emerging entities have been proposed. However, they do not consider the multi-batch emergence scenario, where new entities and new facts are usually added to knowledge graphs (KGs) in multiple batches in the order of their emergence. To simulate the continuous growth of knowledge graphs, a novel multi-batch emergence (MBE) scenario has recently been proposed. We propose a path-based inductive model to handle multi-batch entity growth, enhancing entity encoding with type information. Specifically, we observe a noteworthy pattern in which entity types at the head and tail of the same relation exhibit relative regularity. To utilize this regularity, we introduce a pair of learnable parameters for each relation, representing entity type features linked to the relation. The type features are dedicated to encoding and updating the features of entities. Meanwhile, our model incorporates a novel attention mechanism, combining statistical co-occurrence and semantic similarity of relations effectively for contextual information capture. After generating embeddings, we employ reinforcement learning for path reasoning. To reduce sparsity and expand the action space, our model generates soft candidate facts by grounding a set of soft path rules. Meanwhile, we incorporate the confidence scores of these facts in the action space to facilitate the agent to better distinguish between original facts and rule-generated soft facts. Performances on three multi-batch entity growth datasets demonstrate robust performance, consistently outperforming state-of-the-art models.


Subject(s)
Neural Networks, Computer , Algorithms , Semantics , Knowledge , Humans , Machine Learning , Computer Simulation
SELECTION OF CITATIONS
SEARCH DETAIL