Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 20
Filtrar
1.
J Med Internet Res ; 26: e45209, 2024 Jan 30.
Artigo em Inglês | MEDLINE | ID: mdl-38289660

RESUMO

BACKGROUND: The increasing use of electronic health records and the Internet of Things has led to interoperability issues at different levels (structural and semantic). Standards are important not only for successfully exchanging data but also for appropriately interpreting them (semantic interoperability). Thus, to facilitate the semantic interoperability of data exchanged in health care, considerable resources have been deployed to improve the quality of shared clinical data by structuring and mapping them to the Fast Healthcare Interoperability Resources (FHIR) standard. OBJECTIVE: The aims of this study are 2-fold: to inventory the studies on FHIR semantic interoperability resources and terminologies and to identify and classify the approaches and contributions proposed in these studies. METHODS: A systematic mapping review (SMR) was conducted using 10 electronic databases as sources of information for inventory and review studies published during 2012 to 2022 on the development and improvement of semantic interoperability using the FHIR standard. RESULTS: A total of 70 FHIR studies were selected and analyzed to identify FHIR resource types and terminologies from a semantic perspective. The proposed semantic approaches were classified into 6 categories, namely mapping (31/126, 24.6%), terminology services (18/126, 14.3%), resource description framework or web ontology language-based proposals (24/126, 19%), annotation proposals (18/126, 14.3%), machine learning (ML) and natural language processing (NLP) proposals (20/126, 15.9%), and ontology-based proposals (15/126, 11.9%). From 2012 to 2022, there has been continued research in 6 categories of approaches as well as in new and emerging annotations and ML and NLP proposals. This SMR also classifies the contributions of the selected studies into 5 categories: framework or architecture proposals, model proposals, technique proposals, comparison services, and tool proposals. The most frequent type of contribution is the proposal of a framework or architecture to enable semantic interoperability. CONCLUSIONS: This SMR provides a classification of the different solutions proposed to address semantic interoperability using FHIR at different levels: collecting, extracting and annotating data, modeling electronic health record data from legacy systems, and applying transformation and mapping to FHIR models and terminologies. The use of ML and NLP for unstructured data is promising and has been applied to specific use case scenarios. In addition, terminology services are needed to accelerate their use and adoption; furthermore, techniques and tools to automate annotation and ontology comparison should help reduce human interaction.


Assuntos
Registros Eletrônicos de Saúde , Semântica , Humanos , Idioma , Bases de Dados Factuais , Atenção à Saúde
2.
Sensors (Basel) ; 24(9)2024 Apr 27.
Artigo em Inglês | MEDLINE | ID: mdl-38732910

RESUMO

IoT has seen remarkable growth, particularly in healthcare, leading to the rise of IoMT. IoMT integrates medical devices for real-time data analysis and transmission but faces challenges in data security and interoperability. This research identifies a significant gap in the existing literature regarding a comprehensive ontology for vulnerabilities in medical IoT devices. This paper proposes a fundamental domain ontology named MIoT (Medical Internet of Things) ontology, focusing on cybersecurity in IoMT (Internet of Medical Things), particularly in remote patient monitoring settings. This research will refer to similar-looking acronyms, IoMT and MIoT ontology. It is important to distinguish between the two. IoMT is a collection of various medical devices and their applications within the research domain. On the other hand, MIoT ontology refers to the proposed ontology that defines various concepts, roles, and individuals. MIoT ontology utilizes the knowledge engineering methodology outlined in Ontology Development 101, along with the structured life cycle, and establishes semantic interoperability among medical devices to secure IoMT assets from vulnerabilities and cyberattacks. By defining key concepts and relationships, it becomes easier to understand and analyze the complex network of information within the IoMT. The MIoT ontology captures essential key terms and security-related entities for future extensions. A conceptual model is derived from the MIoT ontology and validated through a case study. Furthermore, this paper outlines a roadmap for future research, highlighting potential impacts on security automation in healthcare applications.


Assuntos
Segurança Computacional , Internet das Coisas , Humanos , Monitorização Fisiológica/métodos , Monitorização Fisiológica/instrumentação , Telemedicina/métodos
3.
BMC Med Inform Decis Mak ; 22(1): 16, 2022 01 19.
Artigo em Inglês | MEDLINE | ID: mdl-35042480

RESUMO

BACKGROUND: For standardization of terms in the reports of medical device adverse events, 89 Japanese medical device adverse event terminologies were published in March 2015. The 89 terminologies were developed independently by 13 industry associations, suggesting that there may be inconsistencies among the terms proposed. The purpose of this study was to integrate the 89 sets of terminologies and evaluate inconsistencies among them using SPARQL. METHODS: In order to evaluate the inconsistencies among the integrated terminology, the following six items were evaluated: (1) whether the two-layer structure between category term and preferred term is consistent, (2) whether synonyms of a preferred term are involved. Reversing the layer-category order of matching was also performed, (3) whether each preferred term is subordinate to only one category term, (4) whether the definitions of terms are uniquely determined, (5) whether CDRH-NCIt terms corresponding to preferred terms are uniquely determined, (6) whether a term in a medical device problem is used for patient problems. RESULTS: About 60% of the total number of duplicated terms were found. This is because industry associations that created multiple terminologies adopted the same terms in terminologies of similar medical device groups. In the case that all terms with the same spelling have the same concept, efficient integration can be achieved automatically using RDF. Furthermore, we evaluated six matters of inconsistency in this study, terms that need to be reviewed accounted for about 10% or less than 10% in each item. CONCLUSIONS: The RDF and SPARQL were useful tools to explore inconsistencies of hierarchies, definition statements, and synonyms when integrating terminolgy by term notation, and these had the advantage of reducing the physical and time burden.


Assuntos
Idioma , Humanos , Japão
4.
J Biomed Inform ; 117: 103755, 2021 05.
Artigo em Inglês | MEDLINE | ID: mdl-33781919

RESUMO

Resource Description Framework (RDF) is one of the three standardized data formats in the HL7 Fast Healthcare Interoperability Resources (FHIR) specification and is being used by healthcare and research organizations to join FHIR and non-FHIR data. However, RDF previously had not been integrated into popular FHIR tooling packages, hindering the adoption of FHIR RDF in the semantic web and other communities. The objective of the study is to develop and evaluate a Java based FHIR RDF data transformation toolkit to facilitate the use and validation of FHIR RDF data. We extended the popular HAPI FHIR tooling to add RDF support, thus enabling FHIR data in XML or JSON to be transformed to or from RDF. We also developed an RDF Shape Expression (ShEx)-based validation framework to verify conformance of FHIR RDF data to the ShEx schemas provided in the FHIR specification for FHIR versions R4 and R5. The effectiveness of ShEx validation was demonstrated by testing it against 2693 FHIR R4 examples and 2197 FHIR R5 examples that are included in the FHIR specification. A total of 5 types of errors including missing properties, unknown element, missing resource Type, invalid attribute value, and unknown resource name in the R5 examples were revealed, demonstrating the value of the ShEx in the quality assurance of the evolving R5 development. This FHIR RDF data transformation and validation framework, based on HAPI and ShEx, is robust and ready for community use in adopting FHIR RDF, improving FHIR data quality, and evolving the FHIR specification.


Assuntos
Atenção à Saúde , Registros Eletrônicos de Saúde
5.
J Biomed Inform ; 67: 90-100, 2017 03.
Artigo em Inglês | MEDLINE | ID: mdl-28213144

RESUMO

BACKGROUND: HL7 Fast Healthcare Interoperability Resources (FHIR) is an emerging open standard for the exchange of electronic healthcare information. FHIR resources are defined in a specialized modeling language. FHIR instances can currently be represented in either XML or JSON. The FHIR and Semantic Web communities are developing a third FHIR instance representation format in Resource Description Framework (RDF). Shape Expressions (ShEx), a formal RDF data constraint language, is a candidate for describing and validating the FHIR RDF representation. OBJECTIVE: Create a FHIR to ShEx model transformation and assess its ability to describe and validate FHIR RDF data. METHODS: We created the methods and tools that generate the ShEx schemas modeling the FHIR to RDF specification being developed by HL7 ITS/W3C RDF Task Force, and evaluated the applicability of ShEx in the description and validation of FHIR to RDF transformations. RESULTS: The ShEx models contributed significantly to workgroup consensus. Algorithmic transformations from the FHIR model to ShEx schemas and FHIR example data to RDF transformations were incorporated into the FHIR build process. ShEx schemas representing 109 FHIR resources were used to validate 511 FHIR RDF data examples from the Standards for Trial Use (STU 3) Ballot version. We were able to uncover unresolved issues in the FHIR to RDF specification and detect 10 types of errors and root causes in the actual implementation. The FHIR ShEx representations have been included in the official FHIR web pages for the STU 3 Ballot version since September 2016. DISCUSSION: ShEx can be used to define and validate the syntax of a FHIR resource, which is complementary to the use of RDF Schema (RDFS) and Web Ontology Language (OWL) for semantic validation. CONCLUSION: ShEx proved useful for describing a standard model of FHIR RDF data. The combination of a formal model and a succinct format enabled comprehensive review and automated validation.


Assuntos
Algoritmos , Internet , Semântica , Registros Eletrônicos de Saúde , Humanos
6.
Biochim Biophys Acta ; 1844(1 Pt A): 98-107, 2014 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-23429179

RESUMO

This paper focuses on the use of controlled vocabularies (CVs) and ontologies especially in the area of proteomics, primarily related to the work of the Proteomics Standards Initiative (PSI). It describes the relevant proteomics standard formats and the ontologies used within them. Software and tools for working with these ontology files are also discussed. The article also examines the "mapping files" used to ensure correct controlled vocabulary terms that are placed within PSI standards and the fulfillment of the MIAPE (Minimum Information about a Proteomics Experiment) requirements. This article is part of a Special Issue entitled: Computational Proteomics in the Post-Identification Era. Guest Editors: Martin Eisenacher and Christian Stephan.


Assuntos
Proteômica , Vocabulário Controlado , Linguagens de Programação , Software
7.
J Cheminform ; 16(1): 49, 2024 May 01.
Artigo em Inglês | MEDLINE | ID: mdl-38693555

RESUMO

Adverse Outcome Pathways (AOPs) have been proposed to facilitate mechanistic understanding of interactions of chemicals/materials with biological systems. Each AOP starts with a molecular initiating event (MIE) and possibly ends with adverse outcome(s) (AOs) via a series of key events (KEs). So far, the interaction of engineered nanomaterials (ENMs) with biomolecules, biomembranes, cells, and biological structures, in general, is not yet fully elucidated. There is also a huge lack of information on which AOPs are ENMs-relevant or -specific, despite numerous published data on toxicological endpoints they trigger, such as oxidative stress and inflammation. We propose to integrate related data and knowledge recently collected. Our approach combines the annotation of nanomaterials and their MIEs with ontology annotation to demonstrate how we can then query AOPs and biological pathway information for these materials. We conclude that a FAIR (Findable, Accessible, Interoperable, Reusable) representation of the ENM-MIE knowledge simplifies integration with other knowledge. SCIENTIFIC CONTRIBUTION: This study introduces a new database linking nanomaterial stressors to the first known MIE or KE. Second, it presents a reproducible workflow to analyze and summarize this knowledge. Third, this work extends the use of semantic web technologies to the field of nanoinformatics and nanosafety.

8.
J Cheminform ; 15(1): 61, 2023 Jun 20.
Artigo em Inglês | MEDLINE | ID: mdl-37340506

RESUMO

Current biological and chemical research is increasingly dependent on the reusability of previously acquired data, which typically come from various sources. Consequently, there is a growing need for database systems and databases stored in them to be interoperable with each other. One of the possible solutions to address this issue is to use systems based on Semantic Web technologies, namely on the Resource Description Framework (RDF) to express data and on the SPARQL query language to retrieve the data. Many existing biological and chemical databases are stored in the form of a relational database (RDB). Converting a relational database into the RDF form and storing it in a native RDF database system may not be desirable in many cases. It may be necessary to preserve the original database form, and having two versions of the same data may not be convenient. A solution may be to use a system mapping the relational database to the RDF form. Such a system keeps data in their original relational form and translates incoming SPARQL queries to equivalent SQL queries, which are evaluated by a relational-database system. This review compares different RDB-to-RDF mapping systems with a primary focus on those that can be used free of charge. In addition, it compares different approaches to expressing RDB-to-RDF mappings. The review shows that these systems represent a viable method providing sufficient performance. Their real-life performance is demonstrated on data and queries coming from the neXtProt project.

9.
BMC Med Genomics ; 15(1): 167, 2022 07 30.
Artigo em Inglês | MEDLINE | ID: mdl-35907849

RESUMO

BACKGROUND: Next-generation sequencing provides comprehensive information about individuals' genetic makeup and is commonplace in precision oncology practice. Due to the heterogeneity of individual patient's disease conditions and treatment journeys, not all targeted therapies were initiated despite actionable mutations. To better understand and support the clinical decision-making process in precision oncology, there is a need to examine real-world associations between patients' genetic information and treatment choices. METHODS: To fill the gap of insufficient use of real-world data (RWD) in electronic health records (EHRs), we generated a single Resource Description Framework (RDF) resource, called PO2RDF (precision oncology to RDF), by integrating information regarding genes, variants, diseases, and drugs from genetic reports and EHRs. RESULTS: There are a total 2,309,014 triples contained in the PO2RDF. Among them, 32,815 triples are related to Gene, 34,695 triples are related to Variant, 8,787 triples are related to Disease, 26,154 triples are related to Drug. We performed two use case analyses to demonstrate the usability of the PO2RDF: (1) we examined real-world associations between EGFR mutations and targeted therapies to confirm existing knowledge and detect off-label use. (2) We examined differences in prognosis for lung cancer patients with/without TP53 mutations. CONCLUSIONS: In conclusion, our work proposed to use RDF to organize and distribute clinical RWD that is otherwise inaccessible externally. Our work serves as a pilot study that will lead to new clinical applications and could ultimately stimulate progress in the field of precision oncology.


Assuntos
Neoplasias , Sequenciamento de Nucleotídeos em Larga Escala , Humanos , Oncologia , Neoplasias/tratamento farmacológico , Neoplasias/genética , Projetos Piloto , Medicina de Precisão
10.
Appl In Vitro Toxicol ; 8(1): 2-13, 2022 Mar 01.
Artigo em Inglês | MEDLINE | ID: mdl-35388368

RESUMO

Introduction: The AOP-Wiki is the main platform for the development and storage of adverse outcome pathways (AOPs). These AOPs describe mechanistic information about toxicodynamic processes and can be used to develop effective risk assessment strategies. However, it is challenging to automatically and systematically parse, filter, and use its contents. We explored solutions to better structure the AOP-Wiki content, and to link it with chemical and biological resources. Together, this allows more detailed exploration, which can be automated. Materials and Methods: We converted the complete AOP-Wiki content into resource description framework (RDF) triples. We used >20 ontologies for the semantic annotation of property-object relations, including the Chemical Information Ontology, Dublin Core, and the AOP Ontology. Results: The resulting RDF contains >122,000 triples describing 158 unique properties of >15,000 unique subjects. Furthermore, >3500 link-outs were added to 12 chemical databases, and >7500 link-outs to 4 gene and protein databases. The AOP-Wiki RDF has been made available at https://aopwiki.rdf.bigcat-bioinformatics.org. Discussion: SPARQL queries can be used to answer biological and toxicological questions, such as listing measurement methods for all Key Events leading to an Adverse Outcome of interest. The full power that the use of this new resource provides becomes apparent when combining the content with external databases using federated queries. Conclusion: Overall, the AOP-Wiki RDF allows new ways to explore the rapidly growing AOP knowledge and makes the integration of this database in automated workflows possible, making the AOP-Wiki more FAIR.

11.
HRB Open Res ; 4: 101, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-38433955

RESUMO

Background: The benefits of having high-quality healthcare data are well established. However, high-dimensionality and irregularity of healthcare data pose challenges in their management. Knowledge graphs have gained increasing popularity in many domains, as a method for representing data to overcome such challenges. However, little is known about their suitability for use with healthcare data. One important factor in representing data is "time". Data with time related attributes are considered, temporal data. Temporal data are frequently observed in healthcare and the management of rapidly changing patient data is an ongoing challenge. Traditionally, data models have focused on presenting static data and do not account for temporal data. Temporal data models ensure time consistency in data models and assist analysing the history of data and predicting the future trends in data. Knowledge graphs can include temporal data models and are therefore of interest to the field of healthcare data management. As such, the herein aim is to outline a protocol for an inter-disciplinary systematic review of approaches, applications and challenges in modelling temporal data in knowledge graphs so that we can inform the application of knowledge graphs to healthcare data. Method: The research questions is, what are the existing approaches in modelling temporal data in RDF based knowledge graphs. Two sub-questions on applications, and challenges will also be evaluated. ACM digital library, IEEE Xplore and Scopus will be searched for this review. The search will be limited to peer-reviewed literature referring to knowledge graphs based on Resource Description Framework (RDF). A narrative synthesis of the papers will be conducted. Conclusion: The findings of this systematic review will be useful for data engineers to better represent data and perform analytics through temporal data modelling. They can be applied in the context of healthcare data and the current challenges faced in managing rapidly changing patient data.

12.
JMIR Med Inform ; 9(5): e23586, 2021 May 25.
Artigo em Inglês | MEDLINE | ID: mdl-34032581

RESUMO

BACKGROUND: Precision oncology has the potential to leverage clinical and genomic data in advancing disease prevention, diagnosis, and treatment. A key research area focuses on the early detection of primary cancers and potential prediction of cancers of unknown primary in order to facilitate optimal treatment decisions. OBJECTIVE: This study presents a methodology to harmonize phenotypic and genetic data features to classify primary cancer types and predict cancers of unknown primaries. METHODS: We extracted genetic data elements from oncology genetic reports of 1011 patients with cancer and their corresponding phenotypical data from Mayo Clinic's electronic health records. We modeled both genetic and electronic health record data with HL7 Fast Healthcare Interoperability Resources. The semantic web Resource Description Framework was employed to generate the network-based data representation (ie, patient-phenotypic-genetic network). Based on the Resource Description Framework data graph, Node2vec graph-embedding algorithm was applied to generate features. Multiple machine learning and deep learning backbone models were compared for cancer prediction performance. RESULTS: With 6 machine learning tasks designed in the experiment, we demonstrated the proposed method achieved favorable results in classifying primary cancer types (area under the receiver operating characteristic curve [AUROC] 96.56% for all 9 cancer predictions on average based on the cross-validation) and predicting unknown primaries (AUROC 80.77% for all 8 cancer predictions on average for real-patient validation). To demonstrate the interpretability, 17 phenotypic and genetic features that contributed the most to the prediction of each cancer were identified and validated based on a literature review. CONCLUSIONS: Accurate prediction of cancer types can be achieved with existing electronic health record data with satisfactory precision. The integration of genetic reports improves prediction, illustrating the translational values of incorporating genetic tests early at the diagnosis stage for patients with cancer.

13.
F1000Res ; 8: 1822, 2019.
Artigo em Inglês | MEDLINE | ID: mdl-32612807

RESUMO

The increasing use of Semantic Web technologies in the life sciences, in particular the use of the Resource Description Framework (RDF) and the RDF query language SPARQL, opens the path for novel integrative analyses, combining information from multiple data sources. However, analyzing evolutionary data in RDF is not trivial, due to the steep learning curve required to understand both the data models adopted by different RDF data sources, as well as the equivalent SPARQL constructs required to benefit from this data - in particular, recursive property paths. In this article, we provide a hands-on introduction to querying evolutionary data across several data sources that publish orthology information in RDF, namely: The Orthologous MAtrix (OMA), the European Bioinformatics Institute (EBI) RDF platform, the Database of Orthologous Groups (OrthoDB) and the Microbial Genome Database (MBGD). We present four protocols in increasing order of complexity. In these protocols, we demonstrate through SPARQL queries how to retrieve pairwise orthologs, homologous groups, and hierarchical orthologous groups. Finally, we show how orthology information in different data sources can be compared, through the use of federated SPARQL queries.


Assuntos
Evolução Biológica , Biologia Computacional , Armazenamento e Recuperação da Informação , Linguagens de Programação , Bases de Dados Factuais , Genoma Microbiano
14.
HRB Open Res ; 1: 20, 2018.
Artigo em Inglês | MEDLINE | ID: mdl-32002509

RESUMO

There is an ongoing challenge as to how best manage and understand 'big data' in precision medicine settings. This paper describes the potential for a Linked Data approach, using a Resource Description Framework (RDF) model, to combine multiple datasets with temporal and spatial elements of varying dimensionality. This "AVERT model" provides a framework for converting multiple standalone files of various formats, from both clinical and environmental settings, into a single data source. This data source can thereafter be queried effectively, shared with outside parties, more easily understood by multiple stakeholders using standardized vocabularies, incorporating provenance metadata and supporting temporo-spatial reasoning. The approach has further advantages in terms of data sharing, security and subsequent analysis. We use a case study relating to anti-Glomerular Basement Membrane (GBM) disease, a rare autoimmune condition, to illustrate a technical proof of concept for the AVERT model.

15.
Artigo em Inglês | MEDLINE | ID: mdl-29629236

RESUMO

Obesity has been linked to several types of cancer. Access to adequate health information activates people's participation in managing their own health, which ultimately improves their health outcomes. Nevertheless, the existing online information about the relationship between obesity and cancer is heterogeneous and poorly organized. A formal knowledge representation can help better organize and deliver quality health information. Currently, there are several efforts in the biomedical domain to convert unstructured data to structured data and store them in Semantic Web knowledge bases (KB). In this demo paper, we present, OC-2-KB (Obesity and Cancer to Knowledge Base), a system that is tailored to guide the automatic KB construction for managing obesity and cancer knowledge from free-text scientific literature (i.e., PubMed abstracts) in a systematic way. OC-2-KB has two important modules which perform the acquisition of entities and the extraction then classification of relationships among these entities. We tested the OC-2-KB system on a data set with 23 manually annotated obesity and cancer PubMed abstracts and created a preliminary KB with 765 triples. We conducted a preliminary evaluation on this sample of triples and reported our evaluation results.

16.
IEEE Trans Biomed Eng ; 64(11): 2531-2545, 2017 11.
Artigo em Inglês | MEDLINE | ID: mdl-28885147

RESUMO

Quantitative susceptibility mapping (QSM) solves the magnetic field-to-magnetization (tissue susceptibility) inverse problem under conditions of noisy and incomplete field data acquired using magnetic resonance imaging. Therefore, sophisticated algorithms are necessary to treat the ill-posed nature of the problem and are reviewed here. The forward problem is typically presented as an integral form, where the field is the convolution of the dipole kernel and tissue susceptibility distribution. This integral form can be equivalently written as a partial differential equation (PDE). Algorithmic challenges are to reduce streaking and shadow artifacts characterized by the fundamental solution of the PDE. Bayesian maximum a posteriori estimation can be employed to solve the inverse problem, where morphological and relevant biomedical knowledge (specific to the imaging situation) are used as priors. As the cost functions in Bayesian QSM framework are typically convex, solutions can be robustly computed using a gradient-based optimization algorithm. Moreover, one can not only accelerate Bayesian QSM, but also increase its effectiveness at reducing shadows using prior knowledge based preconditioners. Improving the efficiency of QSM is under active development, and a rigorous analysis of preconditioning needs to be carried out for further investigation.Quantitative susceptibility mapping (QSM) solves the magnetic field-to-magnetization (tissue susceptibility) inverse problem under conditions of noisy and incomplete field data acquired using magnetic resonance imaging. Therefore, sophisticated algorithms are necessary to treat the ill-posed nature of the problem and are reviewed here. The forward problem is typically presented as an integral form, where the field is the convolution of the dipole kernel and tissue susceptibility distribution. This integral form can be equivalently written as a partial differential equation (PDE). Algorithmic challenges are to reduce streaking and shadow artifacts characterized by the fundamental solution of the PDE. Bayesian maximum a posteriori estimation can be employed to solve the inverse problem, where morphological and relevant biomedical knowledge (specific to the imaging situation) are used as priors. As the cost functions in Bayesian QSM framework are typically convex, solutions can be robustly computed using a gradient-based optimization algorithm. Moreover, one can not only accelerate Bayesian QSM, but also increase its effectiveness at reducing shadows using prior knowledge based preconditioners. Improving the efficiency of QSM is under active development, and a rigorous analysis of preconditioning needs to be carried out for further investigation.


Assuntos
Algoritmos , Processamento de Imagem Assistida por Computador/métodos , Imageamento por Ressonância Magnética/métodos , Artefatos , Teorema de Bayes , Encéfalo/diagnóstico por imagem , Simulação por Computador , Humanos
17.
J Cheminform ; 8: 31, 2016.
Artigo em Inglês | MEDLINE | ID: mdl-27275187

RESUMO

BACKGROUND: In recent years, the Resource Description Framework (RDF) and the SPARQL query language have become more widely used in the area of cheminformatics and bioinformatics databases. These technologies allow better interoperability of various data sources and powerful searching facilities. However, we identified several deficiencies that make usage of such RDF databases restrictive or challenging for common users. RESULTS: We extended a SPARQL engine to be able to use special procedures inside SPARQL queries. This allows the user to work with data that cannot be simply precomputed and thus cannot be directly stored in the database. We designed an algorithm that checks a query against data ontology to identify possible user errors. This greatly improves query debugging. We also introduced an approach to visualize retrieved data in a user-friendly way, based on templates describing visualizations of resource classes. To integrate all of our approaches, we developed a simple web application. CONCLUSIONS: Our system was implemented successfully, and we demonstrated its usability on the ChEBI database transformed into RDF form. To demonstrate procedure call functions, we employed compound similarity searching based on OrChem. The application is publicly available at https://bioinfo.uochb.cas.cz/projects/chemRDF.

18.
J Libr Metadata ; 15(3-4): 157-176, 2015.
Artigo em Inglês | MEDLINE | ID: mdl-26877832

RESUMO

In February 2014 the National Library of Medicine formed the Linked Data Infrastructure Working Group to investigate the potential for publishing linked data, determine best practices for publishing linked data, and prioritize linked data projects, beginning with transforming the Medical Subject Headings as a linked data pilot. This article will review the pilot project to convert the Medical Subject Headings from XML to RDF. It will discuss the collaborative process, the technical and organizational issues tackled, and the future of linked data at the library.

19.
J Cheminform ; 6: 18, 2014.
Artigo em Inglês | MEDLINE | ID: mdl-24855494

RESUMO

BACKGROUND: The drug discovery process is now highly dependent on the management, curation and integration of large amounts of potentially useful data. Semantics are necessary in order to interpret the information and derive knowledge. Advances in recent years have mitigated concerns that the lack of robust, usable tools has inhibited the adoption of methodologies based on semantics. RESULTS: THIS PAPER PRESENTS THREE EXAMPLES OF HOW SEMANTIC WEB TECHNIQUES AND TECHNOLOGIES CAN BE USED IN ORDER TO SUPPORT CHEMISTRY RESEARCH: a controlled vocabulary for quantities, units and symbols in physical chemistry; a controlled vocabulary for the classification and labelling of chemical substances and mixtures; and, a database of chemical identifiers. This paper also presents a Web-based service that uses the datasets in order to assist with the completion of risk assessment forms, along with a discussion of the legal implications and value-proposition for the use of such a service. CONCLUSIONS: We have introduced the Semantic Web concepts, technologies, and methodologies that can be used to support chemistry research, and have demonstrated the application of those techniques in three areas very relevant to modern chemistry research, generating three new datasets that we offer as exemplars of an extensible portfolio of advanced data integration facilities. We have thereby established the importance of Semantic Web techniques and technologies for meeting Wild's fourth "grand challenge".

20.
E-Cienc. inf ; 6(2)dic. 2016.
Artigo em Espanhol | LILACS-Express | LILACS | ID: biblio-1506095

RESUMO

Por medio de esta revisión, se busca realizar un recuento y una explicación sobre el entorno de intercambio bibliográfico BibFrame a partir del conocimiento que ha sido diseminado, en su mayor parte, por un grupo de profesionales de la Biblioteca del Congreso de los Estados Unidos de América. En este trabajo, se muestran aspectos clave de esta iniciativa en y se enfocan cuestiones medulares de su funcionamiento como entorno ubicado en la web. Se discute su posible normalización partiendo del traslado de MARC21 a esta herramienta.


This review aims to report on and explain the bibliographic exchange environment BibFrame taking as starting point the knowledge that has been disseminated, mainly, by a group of professionals from the Library Of Congress of the United States of America. This work shows key aspects of this initiative and focus on medular points regarding its functioning as a web environment. The work discusses the possible standardization of this environment, starting with transferring from MARC21 to this tool.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA