Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 21
Filter
Add more filters










Publication year range
1.
Clin Res Cardiol ; 113(5): 672-679, 2024 May.
Article in English | MEDLINE | ID: mdl-37847314

ABSTRACT

The sharing and documentation of cardiovascular research data are essential for efficient use and reuse of data, thereby aiding scientific transparency, accelerating the progress of cardiovascular research and healthcare, and contributing to the reproducibility of research results. However, challenges remain. This position paper, written on behalf of and approved by the German Cardiac Society and German Centre for Cardiovascular Research, summarizes our current understanding of the challenges in cardiovascular research data management (RDM). These challenges include lack of time, awareness, incentives, and funding for implementing effective RDM; lack of standardization in RDM processes; a need to better identify meaningful and actionable data among the increasing volume and complexity of data being acquired; and a lack of understanding of the legal aspects of data sharing. While several tools exist to increase the degree to which data are findable, accessible, interoperable, and reusable (FAIR), more work is needed to lower the threshold for effective RDM not just in cardiovascular research but in all biomedical research, with data sharing and reuse being factored in at every stage of the scientific process. A culture of open science with FAIR research data should be fostered through education and training of early-career and established research professionals. Ultimately, FAIR RDM requires permanent, long-term effort at all levels. If outcomes can be shown to be superior and to promote better (and better value) science, modern RDM will make a positive difference to cardiovascular science and practice. The full position paper is available in the supplementary materials.


Subject(s)
Biomedical Research , Cardiovascular System , Humans , Data Management , Reproducibility of Results , Heart
2.
EMBO J ; 42(23): e115008, 2023 Dec 01.
Article in English | MEDLINE | ID: mdl-37964598

ABSTRACT

The main goals and challenges for the life science communities in the Open Science framework are to increase reuse and sustainability of data resources, software tools, and workflows, especially in large-scale data-driven research and computational analyses. Here, we present key findings, procedures, effective measures and recommendations for generating and establishing sustainable life science resources based on the collaborative, cross-disciplinary work done within the EOSC-Life (European Open Science Cloud for Life Sciences) consortium. Bringing together 13 European life science research infrastructures, it has laid the foundation for an open, digital space to support biological and medical research. Using lessons learned from 27 selected projects, we describe the organisational, technical, financial and legal/ethical challenges that represent the main barriers to sustainability in the life sciences. We show how EOSC-Life provides a model for sustainable data management according to FAIR (findability, accessibility, interoperability, and reusability) principles, including solutions for sensitive- and industry-related resources, by means of cross-disciplinary training and best practices sharing. Finally, we illustrate how data harmonisation and collaborative work facilitate interoperability of tools, data, solutions and lead to a better understanding of concepts, semantics and functionalities in the life sciences.


Subject(s)
Biological Science Disciplines , Biomedical Research , Software , Workflow
3.
Sci Data ; 10(1): 292, 2023 05 19.
Article in English | MEDLINE | ID: mdl-37208467

ABSTRACT

The notion that data should be Findable, Accessible, Interoperable and Reusable, according to the FAIR Principles, has become a global norm for good data stewardship and a prerequisite for reproducibility. Nowadays, FAIR guides data policy actions and professional practices in the public and private sectors. Despite such global endorsements, however, the FAIR Principles are aspirational, remaining elusive at best, and intimidating at worst. To address the lack of practical guidance, and help with capability gaps, we developed the FAIR Cookbook, an open, online resource of hands-on recipes for "FAIR doers" in the Life Sciences. Created by researchers and data managers professionals in academia, (bio)pharmaceutical companies and information service industries, the FAIR Cookbook covers the key steps in a FAIRification journey, the levels and indicators of FAIRness, the maturity model, the technologies, the tools and the standards available, as well as the skills required, and the challenges to achieve and improve data FAIRness. Part of the ELIXIR ecosystem, and recommended by funders, the FAIR Cookbook is open to contributions of new recipes.

4.
Mol Syst Biol ; 16(8): e9110, 2020 08.
Article in English | MEDLINE | ID: mdl-32845085

ABSTRACT

Systems biology has experienced dramatic growth in the number, size, and complexity of computational models. To reproduce simulation results and reuse models, researchers must exchange unambiguous model descriptions. We review the latest edition of the Systems Biology Markup Language (SBML), a format designed for this purpose. A community of modelers and software authors developed SBML Level 3 over the past decade. Its modular form consists of a core suited to representing reaction-based models and packages that extend the core with features suited to other model types including constraint-based models, reaction-diffusion models, logical network models, and rule-based models. The format leverages two decades of SBML and a rich software ecosystem that transformed how systems biologists build and interact with models. More recently, the rise of multiscale models of whole cells and organs, and new data sources such as single-cell measurements and live imaging, has precipitated new ways of integrating data with models. We provide our perspectives on the challenges presented by these developments and how SBML Level 3 provides the foundation needed to support this evolution.


Subject(s)
Systems Biology/methods , Animals , Humans , Logistic Models , Models, Biological , Software
6.
Article in English | MEDLINE | ID: mdl-27189610

ABSTRACT

BioSharing (http://www.biosharing.org) is a manually curated, searchable portal of three linked registries. These resources cover standards (terminologies, formats and models, and reporting guidelines), databases, and data policies in the life sciences, broadly encompassing the biological, environmental and biomedical sciences. Launched in 2011 and built by the same core team as the successful MIBBI portal, BioSharing harnesses community curation to collate and cross-reference resources across the life sciences from around the world. BioSharing makes these resources findable and accessible (the core of the FAIR principle). Every record is designed to be interlinked, providing a detailed description not only on the resource itself, but also on its relations with other life science infrastructures. Serving a variety of stakeholders, BioSharing cultivates a growing community, to which it offers diverse benefits. It is a resource for funding bodies and journal publishers to navigate the metadata landscape of the biological sciences; an educational resource for librarians and information advisors; a publicising platform for standard and database developers/curators; and a research tool for bench and computer scientists to plan their work. BioSharing is working with an increasing number of journals and other registries, for example linking standards and databases to training material and tools. Driven by an international Advisory Board, the BioSharing user-base has grown by over 40% (by unique IP address), in the last year thanks to successful engagement with researchers, publishers, librarians, developers and other stakeholders via several routes, including a joint RDA/Force11 working group and a collaboration with the International Society for Biocuration. In this article, we describe BioSharing, with a particular focus on community-led curation.Database URL: https://www.biosharing.org.


Subject(s)
Biological Science Disciplines , Crowdsourcing/standards , Database Management Systems , Databases, Factual , Metadata/standards , Biological Science Disciplines/legislation & jurisprudence , Biological Science Disciplines/standards , Computational Biology , Database Management Systems/legislation & jurisprudence , Database Management Systems/standards , Databases, Factual/legislation & jurisprudence , Databases, Factual/standards , Humans , Internet , Registries/standards , User-Computer Interface
7.
PLoS One ; 11(4): e0154556, 2016.
Article in English | MEDLINE | ID: mdl-27128319

ABSTRACT

The Ontology for Biomedical Investigations (OBI) is an ontology that provides terms with precisely defined meanings to describe all aspects of how investigations in the biological and medical domains are conducted. OBI re-uses ontologies that provide a representation of biomedical knowledge from the Open Biological and Biomedical Ontologies (OBO) project and adds the ability to describe how this knowledge was derived. We here describe the state of OBI and several applications that are using it, such as adding semantic expressivity to existing databases, building data entry forms, and enabling interoperability between knowledge resources. OBI covers all phases of the investigation process, such as planning, execution and reporting. It represents information and material entities that participate in these processes, as well as roles and functions. Prior to OBI, it was not possible to use a single internally consistent resource that could be applied to multiple types of experiments for these applications. OBI has made this possible by creating terms for entities involved in biological and medical investigations and by importing parts of other biomedical ontologies such as GO, Chemical Entities of Biological Interest (ChEBI) and Phenotype Attribute and Trait Ontology (PATO) without altering their meaning. OBI is being used in a wide range of projects covering genomics, multi-omics, immunology, and catalogs of services. OBI has also spawned other ontologies (Information Artifact Ontology) and methods for importing parts of ontologies (Minimum information to reference an external ontology term (MIREOT)). The OBI project is an open cross-disciplinary collaborative effort, encompassing multiple research communities from around the globe. To date, OBI has created 2366 classes and 40 relations along with textual and formal definitions. The OBI Consortium maintains a web resource (http://obi-ontology.org) providing details on the people, policies, and issues being addressed in association with OBI. The current release of OBI is available at http://purl.obolibrary.org/obo/obi.owl.


Subject(s)
Biological Ontologies , Animals , Biological Ontologies/organization & administration , Biological Ontologies/statistics & numerical data , Biological Ontologies/trends , Computational Biology , Databases, Factual , Humans , Internet , Metadata , Semantics , Software
9.
J Biomed Semantics ; 5: 25, 2014.
Article in English | MEDLINE | ID: mdl-25068035

ABSTRACT

MOTIVATION: Biomedical ontologists to date have concentrated on ontological descriptions of biomedical entities such as gene products and their attributes, phenotypes and so on. Recently, effort has diversified to descriptions of the laboratory investigations by which these entities were produced. However, much biological insight is gained from the analysis of the data produced from these investigations, and there is a lack of adequate descriptions of the wide range of software that are central to bioinformatics. We need to describe how data are analyzed for discovery, audit trails, provenance and reproducibility. RESULTS: The Software Ontology (SWO) is a description of software used to store, manage and analyze data. Input to the SWO has come from beyond the life sciences, but its main focus is the life sciences. We used agile techniques to gather input for the SWO and keep engagement with our users. The result is an ontology that meets the needs of a broad range of users by describing software, its information processing tasks, data inputs and outputs, data formats versions and so on. Recently, the SWO has incorporated EDAM, a vocabulary for describing data and related concepts in bioinformatics. The SWO is currently being used to describe software used in multiple biomedical applications. CONCLUSION: The SWO is another element of the biomedical ontology landscape that is necessary for the description of biomedical entities and how they were discovered. An ontology of software used to analyze data produced by investigations in the life sciences can be made in such a way that it covers the important features requested and prioritized by its users. The SWO thus fits into the landscape of biomedical ontologies and is produced using techniques designed to keep it in line with user's needs. AVAILABILITY: The Software Ontology is available under an Apache 2.0 license at http://theswo.sourceforge.net/; the Software Ontology blog can be read at http://softwareontology.wordpress.com.

10.
Mol Syst Biol ; 7: 543, 2011 Oct 25.
Article in English | MEDLINE | ID: mdl-22027554

ABSTRACT

The use of computational modeling to describe and analyze biological systems is at the heart of systems biology. Model structures, simulation descriptions and numerical results can be encoded in structured formats, but there is an increasing need to provide an additional semantic layer. Semantic information adds meaning to components of structured descriptions to help identify and interpret them unambiguously. Ontologies are one of the tools frequently used for this purpose. We describe here three ontologies created specifically to address the needs of the systems biology community. The Systems Biology Ontology (SBO) provides semantic information about the model components. The Kinetic Simulation Algorithm Ontology (KiSAO) supplies information about existing algorithms available for the simulation of systems biology models, their characterization and interrelationships. The Terminology for the Description of Dynamics (TEDDY) categorizes dynamical features of the simulation results and general systems behavior. The provision of semantic information extends a model's longevity and facilitates its reuse. It provides useful insight into the biology of modeled processes, and may be used to make informed decisions on subsequent simulation experiments.


Subject(s)
Computational Biology , Semantics , Systems Biology , Vocabulary, Controlled , Algorithms , Computer Simulation , Information Storage and Retrieval , Models, Biological
11.
PLoS Genet ; 7(4): e1001362, 2011 Apr.
Article in English | MEDLINE | ID: mdl-21490951

ABSTRACT

To better understand telomere biology in budding yeast, we have performed systematic suppressor/enhancer analyses on yeast strains containing a point mutation in the essential telomere capping gene CDC13 (cdc13-1) or containing a null mutation in the DNA damage response and telomere capping gene YKU70 (yku70Δ). We performed Quantitative Fitness Analysis (QFA) on thousands of yeast strains containing mutations affecting telomere-capping proteins in combination with a library of systematic gene deletion mutations. To perform QFA, we typically inoculate 384 separate cultures onto solid agar plates and monitor growth of each culture by photography over time. The data are fitted to a logistic population growth model; and growth parameters, such as maximum growth rate and maximum doubling potential, are deduced. QFA reveals that as many as 5% of systematic gene deletions, affecting numerous functional classes, strongly interact with telomere capping defects. We show that, while Cdc13 and Yku70 perform complementary roles in telomere capping, their genetic interaction profiles differ significantly. At least 19 different classes of functionally or physically related proteins can be identified as interacting with cdc13-1, yku70Δ, or both. Each specific genetic interaction informs the roles of individual gene products in telomere biology. One striking example is with genes of the nonsense-mediated RNA decay (NMD) pathway which, when disabled, suppress the conditional cdc13-1 mutation but enhance the null yku70Δ mutation. We show that the suppressing/enhancing role of the NMD pathway at uncapped telomeres is mediated through the levels of Stn1, an essential telomere capping protein, which interacts with Cdc13 and recruitment of telomerase to telomeres. We show that increased Stn1 levels affect growth of cells with telomere capping defects due to cdc13-1 and yku70Δ. QFA is a sensitive, high-throughput method that will also be useful to understand other aspects of microbial cell biology.


Subject(s)
Saccharomyces cerevisiae Proteins/metabolism , Telomere-Binding Proteins/metabolism , Telomere/metabolism , Telomere/pathology , Transcription Factors/metabolism , Gene Expression Regulation, Fungal , Models, Biological , Mutation/genetics , RNA Stability/genetics , Saccharomyces cerevisiae/genetics , Saccharomyces cerevisiae/metabolism , Telomere/genetics , Telomere-Binding Proteins/genetics , Temperature
12.
J Biomed Semantics ; 1 Suppl 1: S3, 2010 Jun 22.
Article in English | MEDLINE | ID: mdl-20626923

ABSTRACT

BACKGROUND: The creation of accurate quantitative Systems Biology Markup Language (SBML) models is a time-intensive, manual process often complicated by the many data sources and formats required to annotate even a small and well-scoped model. Ideally, the retrieval and integration of biological knowledge for model annotation should be performed quickly, precisely, and with a minimum of manual effort. RESULTS: Here we present rule-based mediation, a method of semantic data integration applied to systems biology model annotation. The heterogeneous data sources are first syntactically converted into ontologies, which are then aligned to a small domain ontology by applying a rule base. We demonstrate proof-of-principle of this application of rule-based mediation using off-the-shelf semantic web technology through two use cases for SBML model annotation. Existing tools and technology provide a framework around which the system is built, reducing development time and increasing usability. CONCLUSIONS: Integrating resources in this way accommodates multiple formats with different semantics, and provides richly-modelled biological knowledge suitable for annotation of SBML models. This initial work establishes the feasibility of rule-based mediation as part of an automated SBML model annotation system. AVAILABILITY: Detailed information on the project files as well as further information on and comparisons with similar projects is available from the project page at http://cisban-silico.cs.ncl.ac.uk/RBM/.

15.
Methods Mol Biol ; 604: 333-43, 2010.
Article in English | MEDLINE | ID: mdl-20013382

ABSTRACT

Data management and sharing in omics science is highly challenging due to the constant evolution of experimental techniques, the range of instrument types and software used for analysis, and the high volumes of data produced. The Functional Genomics Experiment (FuGE) Model was created to provide a model for capturing descriptions of sample processing, experimental protocols and multidimensional data for any kind of omics experiment. FuGE has two modes of action: (a) as a storage architecture for experimental workflows and (b) as a framework for building new technology-specific data standards.FuGE is an object model that is converted into an XML implementation for data exchange. Software toolkits have been developed for data handling and for bridging between XML data files and relational database implementations. FuGE has been adopted by the Proteomics Standards Initiative (PSI, http://www.psidev.info ) for building several new data formats, and it is being used in a variety of other experimental contexts, thus allowing data to be integrated across a range of experimental types to support Systems Biology approaches. This chapter provides a practical guide for laboratories or groups wishing to manage their data, and for developers wishing to create new data formats using FuGE.


Subject(s)
Computational Biology/standards , Database Management Systems/standards , Databases, Genetic/standards , Software/standards , Computational Biology/methods , Genomics/methods , Genomics/standards , Proteomics/methods , Proteomics/standards , Workflow
16.
Bioinformatics ; 25(22): 3026-7, 2009 Nov 15.
Article in English | MEDLINE | ID: mdl-19734151

ABSTRACT

UNLABELLED: Saint is a web application which provides a lightweight annotation integration environment for quantitative biological models. The system enables modellers to rapidly mark up models with biological information derived from a range of data sources. AVAILABILITY AND IMPLEMENTATION: Saint is freely available for use on the web at http://www.cisban.ac.uk/saint. The web application is implemented in Google Web Toolkit and Tomcat, with all major browsers supported. The Java source code is freely available for download at http://saint-annotate.sourceforge.net. The Saint web server requires an installation of libSBML and has been tested on Linux (32-bit Ubuntu 8.10 and 9.04).


Subject(s)
Computational Biology/methods , Software , Databases, Factual , Internet , User-Computer Interface
17.
Genome Biol ; 10(6): 309, 2009.
Article in English | MEDLINE | ID: mdl-19591648

ABSTRACT

A report of BioSysBio 2009, the IET conference on Synthetic Biology, Systems Biology and Bioinformatics, Cambridge, UK, 23-25 March 2009.


Subject(s)
Genes, Synthetic , Systems Biology , Automation , Computational Biology , Ethics, Research
18.
OMICS ; 13(3): 239-51, 2009 Jun.
Article in English | MEDLINE | ID: mdl-19441879

ABSTRACT

The Functional Genomics Experiment data model (FuGE) has been developed to increase the consistency and efficiency of experimental data modeling in the life sciences, and it has been adopted by a number of high-profile standardization organizations. FuGE can be used: (1) directly, whereby generic modeling constructs are used to represent concepts from specific experimental activities; or (2) as a framework within which method-specific models can be developed. FuGE is both rich and flexible, providing a considerable number of modeling constructs, which can be used in a range of different ways. However, such richness and flexibility also mean that modelers and application developers have choices to make when applying FuGE in a given context. This paper captures emerging best practice in the use of FuGE in the light of the experience of several groups by: (1) proposing guidelines for the use and extension of the FuGE data model; (2) presenting design patterns that reflect recurring requirements in experimental data modeling; and (3) describing a community software tool kit (STK) that supports application development using FuGE. We anticipate that these guidelines will encourage consistent usage of FuGE, and as such, will contribute to the development of convergent data standards in omics research.


Subject(s)
Computational Biology/methods , Genomics/methods , Models, Theoretical , Computer Simulation , Flow Cytometry/instrumentation , Flow Cytometry/methods , Reproducibility of Results , Software , User-Computer Interface
19.
OMICS ; 12(2): 101-8, 2008 Jun.
Article in English | MEDLINE | ID: mdl-18564914

ABSTRACT

This meeting report summarizes the proceedings of the "eGenomics: Cataloguing our Complete Genome Collection IV" workshop held June 6-8, 2007, at the National Institute for Environmental eScience (NIEeS), Cambridge, United Kingdom. This fourth workshop of the Genomic Standards Consortium (GSC) was a mix of short presentations, strategy discussions, and technical sessions. Speakers provided progress reports on the development of the "Minimum Information about a Genome Sequence" (MIGS) specification and the closely integrated "Minimum Information about a Metagenome Sequence" (MIMS) specification. The key outcome of the workshop was consensus on the next version of the MIGS/MIMS specification (v1.2). This drove further definition and restructuring of the MIGS/MIMS XML schema (syntax). With respect to semantics, a term vetting group was established to ensure that terms are properly defined and submitted to the appropriate ontology projects. Perhaps the single most important outcome of the workshop was a proposal to move beyond the concept of "minimum" to create a far richer XML schema that would define a "Genomic Contextual Data Markup Language" (GCDML) suitable for wider semantic integration across databases. GCDML will contain not only curated information (e.g., compliant with MIGS/MIMS), but also be extended to include a variety of data processing and calculations. Further information about the Genomic Standards Consortium and its range of activities can be found at http://gensc.org.


Subject(s)
Databases, Genetic , Genomics , Education , Programming Languages , Reference Standards
20.
OMICS ; 12(2): 143-9, 2008 Jun.
Article in English | MEDLINE | ID: mdl-18447634

ABSTRACT

This article summarizes the motivation for, and the proceedings of, the first ISA-TAB workshop held December 6-8, 2007, at the EBI, Cambridge, UK. This exploratory workshop, organized by members of the Microarray Gene Expression Data (MGED) Society's Reporting Structure for Biological Investigations (RSBI) working group, brought together a group of developers of a range of collaborative systems to discuss the use of a common format to address the pressing need of reporting and communicating data and metadata from biological, biomedical, and environmental studies employing combinations of genomics, transcriptomics, proteomics, and metabolomics technologies along with more conventional methodologies. The expertise of the participants comprised database development, data management, and hands-on experience in the development of data communication standards. The workshop's outcomes are set to help formalize the proposed Investigation, Study, Assay (ISA)-TAB tab-delimited format for representing and communicating experimental metadata. This article is part of the special issue of OMICS on the activities of the Genomics Standards Consortium (GSC).


Subject(s)
Computational Biology , Database Management Systems , Education , Genomics , Proteomics , RNA, Messenger/genetics , United Kingdom
SELECTION OF CITATIONS
SEARCH DETAIL