Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 14 de 14
Filter
1.
Bioinformatics ; 38(4): 1171-1172, 2022 01 27.
Article in English | MEDLINE | ID: mdl-34791064

ABSTRACT

SUMMARY: COBREXA.jl is a Julia package for scalable, high-performance constraint-based reconstruction and analysis of very large-scale biological models. Its primary purpose is to facilitate the integration of modern high performance computing environments with the processing and analysis of large-scale metabolic models of challenging complexity. We report the architecture of the package, and demonstrate how the design promotes analysis scalability on several use-cases with multi-organism community models. AVAILABILITY AND IMPLEMENTATION: https://doi.org/10.17881/ZKCR-BT30. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Subject(s)
Computing Methodologies , Software , Models, Biological
2.
Nucleic Acids Res ; 49(W1): W535-W540, 2021 07 02.
Article in English | MEDLINE | ID: mdl-33999203

ABSTRACT

Since 1992 PredictProtein (https://predictprotein.org) is a one-stop online resource for protein sequence analysis with its main site hosted at the Luxembourg Centre for Systems Biomedicine (LCSB) and queried monthly by over 3,000 users in 2020. PredictProtein was the first Internet server for protein predictions. It pioneered combining evolutionary information and machine learning. Given a protein sequence as input, the server outputs multiple sequence alignments, predictions of protein structure in 1D and 2D (secondary structure, solvent accessibility, transmembrane segments, disordered regions, protein flexibility, and disulfide bridges) and predictions of protein function (functional effects of sequence variation or point mutations, Gene Ontology (GO) terms, subcellular localization, and protein-, RNA-, and DNA binding). PredictProtein's infrastructure has moved to the LCSB increasing throughput; the use of MMseqs2 sequence search reduced runtime five-fold (apparently without lowering performance of prediction methods); user interface elements improved usability, and new prediction methods were added. PredictProtein recently included predictions from deep learning embeddings (GO and secondary structure) and a method for the prediction of proteins and residues binding DNA, RNA, or other proteins. PredictProtein.org aspires to provide reliable predictions to computational and experimental biologists alike. All scripts and methods are freely available for offline execution in high-throughput settings.


Subject(s)
Protein Conformation , Software , Binding Sites , Coronavirus Nucleocapsid Proteins/chemistry , DNA-Binding Proteins/chemistry , Phosphoproteins/chemistry , Protein Structure, Secondary , Proteins/chemistry , Proteins/physiology , RNA-Binding Proteins/chemistry , Sequence Alignment , Sequence Analysis, Protein
3.
Metab Eng ; 74: 72-82, 2022 11.
Article in English | MEDLINE | ID: mdl-36152931

ABSTRACT

Metabolic models are typically characterized by a large number of parameters. Traditionally, metabolic control analysis is applied to differential equation-based models to investigate the sensitivity of predictions to parameters. A corresponding theory for constraint-based models is lacking, due to their formulation as optimization problems. Here, we show that optimal solutions of optimization problems can be efficiently differentiated using constrained optimization duality and implicit differentiation. We use this to calculate the sensitivities of predicted reaction fluxes and enzyme concentrations to turnover numbers in an enzyme-constrained metabolic model of Escherichia coli. The sensitivities quantitatively identify rate limiting enzymes and are mathematically precise, unlike current finite difference based approaches used for sensitivity analysis. Further, efficient differentiation of constraint-based models unlocks the ability to use gradient information for parameter estimation. We demonstrate this by improving, genome-wide, the state-of-the-art turnover number estimates for E. coli. Finally, we show that this technique can be generalized to arbitrarily complex models. By differentiating the optimal solution of a model incorporating both thermodynamic and kinetic rate equations, the effect of metabolite concentrations on biomass growth can be elucidated. We benchmark these metabolite sensitivities against a large experimental gene knockdown study, and find good alignment between the predicted sensitivities and in vivo metabolome changes. In sum, we demonstrate several applications of differentiating optimal solutions of constraint-based metabolic models, and show how it connects to classic metabolic control analysis.


Subject(s)
Escherichia coli , Models, Biological , Kinetics , Escherichia coli/genetics , Escherichia coli/metabolism , Thermodynamics , Metabolome , Metabolic Networks and Pathways
4.
BMC Genomics ; 15: 1154, 2014 Dec 20.
Article in English | MEDLINE | ID: mdl-25528190

ABSTRACT

BACKGROUND: The human neuroblastoma cell line, SH-SY5Y, is a commonly used cell line in studies related to neurotoxicity, oxidative stress, and neurodegenerative diseases. Although this cell line is often used as a cellular model for Parkinson's disease, the relevance of this cellular model in the context of Parkinson's disease (PD) and other neurodegenerative diseases has not yet been systematically evaluated. RESULTS: We have used a systems genomics approach to characterize the SH-SY5Y cell line using whole-genome sequencing to determine the genetic content of the cell line and used transcriptomics and proteomics data to determine molecular correlations. Further, we integrated genomic variants using a network analysis approach to evaluate the suitability of the SH-SY5Y cell line for perturbation experiments in the context of neurodegenerative diseases, including PD. CONCLUSIONS: The systems genomics approach showed consistency across different biological levels (DNA, RNA and protein concentrations). Most of the genes belonging to the major Parkinson's disease pathways and modules were intact in the SH-SY5Y genome. Specifically, each analysed gene related to PD has at least one intact copy in SH-SY5Y. The disease-specific network analysis approach ranked the genetic integrity of SH-SY5Y as higher for PD than for Alzheimer's disease but lower than for Huntington's disease and Amyotrophic Lateral Sclerosis for loss of function perturbation experiments.


Subject(s)
Genomics , Neuroblastoma/pathology , Parkinson Disease/genetics , Cell Line, Tumor , DNA Copy Number Variations , DNA Transposable Elements/genetics , Gene Expression Profiling , Genetic Variation , Humans , INDEL Mutation , Proteomics
5.
Sci Data ; 11(1): 501, 2024 May 15.
Article in English | MEDLINE | ID: mdl-38750048

ABSTRACT

The EU General Data Protection Regulation (GDPR) requirements have prompted a shift from centralised controlled access genome-phenome archives to federated models for sharing sensitive human data. In a data-sharing federation, a central node facilitates data discovery; meanwhile, distributed nodes are responsible for handling data access requests, concluding agreements with data users and providing secure access to the data. Research institutions that want to become part of such federations often lack the resources to set up the required controlled access processes. The DS-PACK tool assembly is a reusable, open-source middleware solution that semi-automates controlled access processes end-to-end, from data submission to access. Data protection principles are engraved into all components of the DS-PACK assembly. DS-PACK centralises access control management and distributes access control enforcement with support for data access via cloud-based applications. DS-PACK is in production use at the ELIXIR Luxembourg data hosting platform, combined with an operational model including legal facilitation and data stewardship.


Subject(s)
Information Dissemination , Humans , Access to Information , Computer Security , Software
6.
Cell Commun Signal ; 11(1): 24, 2013 Apr 11.
Article in English | MEDLINE | ID: mdl-23578051

ABSTRACT

Biological systems present multiple scales of complexity, ranging from molecules to entire populations. Light microscopy is one of the least invasive techniques used to access information from various biological scales in living cells. The combination of molecular biology and imaging provides a bottom-up tool for direct insight into how molecular processes work on a cellular scale. However, imaging can also be used as a top-down approach to study the behavior of a system without detailed prior knowledge about its underlying molecular mechanisms. In this review, we highlight the recent developments on microscopy-based systems analyses and discuss the complementary opportunities and different challenges with high-content screening and high-throughput imaging. Furthermore, we provide a comprehensive overview of the available platforms that can be used for image analysis, which enable community-driven efforts in the development of image-based systems biology.

7.
Cell Rep Methods ; 1(3)2021 07 26.
Article in English | MEDLINE | ID: mdl-34761247

ABSTRACT

Omics experiments are ubiquitous in biological studies, leading to a deluge of data. However, it is still challenging to connect changes in these data to changes in cell functions because of complex interdependencies between genes, proteins, and metabolites. Here, we present a framework allowing researchers to infer how metabolic functions change on the basis of omics data. To enable this, we curated and standardized lists of metabolic tasks that mammalian cells can accomplish. Genome-scale metabolic networks were used to define gene sets associated with each metabolic task. We further developed a framework to overlay omics data on these sets and predict pathway usage for each metabolic task. We demonstrated how this approach can be used to quantify metabolic functions of diverse biological samples from the single cell to whole tissues and organs by using multiple transcriptomic datasets. To facilitate its adoption, we integrated the approach into GenePattern (www.genepattern.org-CellFie).


Subject(s)
Genome , Metabolic Networks and Pathways , Animals , Metabolic Networks and Pathways/genetics , Cell Physiological Phenomena , Gene Expression Profiling , Transcriptome/genetics , Mammals/genetics
8.
Gigascience ; 9(11)2020 11 18.
Article in English | MEDLINE | ID: mdl-33205814

ABSTRACT

BACKGROUND: The amount of data generated in large clinical and phenotyping studies that use single-cell cytometry is constantly growing. Recent technological advances allow the easy generation of data with hundreds of millions of single-cell data points with >40 parameters, originating from thousands of individual samples. The analysis of that amount of high-dimensional data becomes demanding in both hardware and software of high-performance computational resources. Current software tools often do not scale to the datasets of such size; users are thus forced to downsample the data to bearable sizes, in turn losing accuracy and ability to detect many underlying complex phenomena. RESULTS: We present GigaSOM.jl, a fast and scalable implementation of clustering and dimensionality reduction for flow and mass cytometry data. The implementation of GigaSOM.jl in the high-level and high-performance programming language Julia makes it accessible to the scientific community and allows for efficient handling and processing of datasets with billions of data points using distributed computing infrastructures. We describe the design of GigaSOM.jl, measure its performance and horizontal scaling capability, and showcase the functionality on a large dataset from a recent study. CONCLUSIONS: GigaSOM.jl facilitates the use of commonly available high-performance computing resources to process the largest available datasets within minutes, while producing results of the same quality as the current state-of-art software. Measurements indicate that the performance scales to much larger datasets. The example use on the data from a massive mouse phenotyping effort confirms the applicability of GigaSOM.jl to huge-scale studies.


Subject(s)
Algorithms , Programming Languages , Animals , Cluster Analysis , Mice , Software
9.
F1000Res ; 9: 1398, 2020.
Article in English | MEDLINE | ID: mdl-33604028

ABSTRACT

Today, academic researchers benefit from the changes driven by digital technologies and the enormous growth of knowledge and data, on globalisation, enlargement of the scientific community, and the linkage between different scientific communities and the society. To fully benefit from this development, however, information needs to be shared openly and transparently. Digitalisation plays a major role here because it permeates all areas of business, science and society and is one of the key drivers for innovation and international cooperation. To address the resulting opportunities, the EU promotes the development and use of collaborative ways to produce and share knowledge and data as early as possible in the research process, but also to appropriately secure results with the European strategy for Open Science (OS). It is now widely recognised that making research results more accessible to all societal actors contributes to more effective and efficient science; it also serves as a boost for innovation in the public and private sectors. However  for research data to be findable, accessible, interoperable and reusable the use of standards is essential. At the metadata level, considerable efforts in standardisation have already been made (e.g. Data Management Plan and FAIR Principle etc.), whereas in context with the raw data these fundamental efforts are still fragmented and in some cases completely missing. The CHARME consortium, funded by the European Cooperation in Science and Technology (COST) Agency, has identified needs and gaps in the field of standardisation in the life sciences and also discussed potential hurdles for implementation of standards in current practice. Here, the authors suggest four measures in response to current challenges to ensure a high quality of life science research data and their re-usability for research and innovation.


Subject(s)
Biological Science Disciplines , Trust , International Cooperation , Metadata , Quality of Life
10.
Gigascience ; 8(12)2019 12 01.
Article in English | MEDLINE | ID: mdl-31800037

ABSTRACT

BACKGROUND: The new European legislation on data protection, namely, the General Data Protection Regulation (GDPR), has introduced comprehensive requirements for the documentation about the processing of personal data as well as informing the data subjects of its use. GDPR's accountability principle requires institutions, projects, and data hubs to document their data processings and demonstrate compliance with the GDPR. In response to this requirement, we see the emergence of commercial data-mapping tools, and institutions creating GDPR data register with such tools. One shortcoming of this approach is the genericity of tools, and their process-based model not capturing the project-based, collaborative nature of data processing in biomedical research. FINDINGS: We have developed a software tool to allow research institutions to comply with the GDPR accountability requirement and map the sometimes very complex data flows in biomedical research. By analysing the transparency and record-keeping obligations of each GDPR principle, we observe that our tool effectively meets the accountability requirement. CONCLUSIONS: The GDPR is bringing data protection to center stage in research data management, necessitating dedicated tools, personnel, and processes. Our tool, DAISY, is tailored specifically for biomedical research and can help institutions in tackling the documentation challenge brought about by the GDPR. DAISY is made available as a free and open source tool on Github. DAISY is actively being used at the Luxembourg Centre for Systems Biomedicine and the ELIXIR-Luxembourg data hub.


Subject(s)
Computer Security/legislation & jurisprudence , Electronic Health Records , Europe , Humans , Social Responsibility
11.
Gigascience ; 7(9)2018 09 01.
Article in English | MEDLINE | ID: mdl-30165440

ABSTRACT

Background: Translational research platforms share the aim of promoting a deeper understanding of stored data by providing visualization and analysis tools for data exploration and hypothesis generation. However, such tools are usually platform bound and are not easily reusable by other systems. Furthermore, they rarely address access restriction issues when direct data transfer is not permitted. In this article, we present an analytical service that works in tandem with a visualization library to address these problems. Findings: Using a combination of existing technologies and a platform-specific data abstraction layer, we developed a service that is capable of providing existing web-based data warehouses and repositories with platform-independent visual analytical capabilities. The design of this service also allows for federated data analysis by eliminating the need to move the data directly to the researcher. Instead, all operations are based on statistics and interactive charts without direct access to the dataset. Conclusions: The software presented in this article has a potential to help translational researchers achieve a better understanding of a given dataset and quickly generate new hypotheses. Furthermore, it provides a framework that can be used to share and reuse explorative analysis tools within the community.


Subject(s)
Databases, Factual , Electronic Data Processing/methods , Software , Translational Research, Biomedical/methods
12.
Curr Opin Biotechnol ; 34: 48-55, 2015 Aug.
Article in English | MEDLINE | ID: mdl-25498477

ABSTRACT

Ecosystems and biological systems are known to be inherently complex and to exhibit nonlinear dynamics. Diseases such as microbiome dysregulation or depression can be seen as complex systems as well and were shown to exhibit patterns of nonlinearity in their response to perturbations. These nonlinearities can be revealed by a sudden shift in system states, for instance from health to disease. The identification and characterization of early warning signals which could predict upcoming critical transitions is of primordial interest as prevention of disease onset is a major aim in health care. In this review, we focus on recent evidence for critical transitions in diseases and discuss the potential of such studies for therapeutic applications.


Subject(s)
Ecosystem , Animals , Chronic Disease , Cross-Sectional Studies , Humans , Longitudinal Studies
13.
Ann Clin Transl Neurol ; 2(1): 67-73, 2015 Jan.
Article in English | MEDLINE | ID: mdl-25642436

ABSTRACT

OBJECTIVE: Mitochondrial dysfunction is a hallmark of idiopathic Parkinson's disease (IPD), which has been reported not to be restricted to striatal neurons. However, studies that analyzed mitochondrial function at the level of selected enzymatic activities in peripheral tissues have produced conflicting data. We considered the electron transport chain as a complex system with mitochondrial membrane potential as an integrative indicator for mitochondrial fitness. METHODS: Twenty-five IPD patients (nine females; mean disease duration, 6.2 years) and 16 healthy age-matched controls (12 females) were recruited. Live platelets were purified using magnetic-activated cell sorting (MACS) and single-cell data on mitochondrial membrane potential (Δψ) were measured by cytometry and challenged with a protonophore agent. RESULTS: Functional mitochondrial membrane potential was detected in all participants. The challenge test reduced the membrane potential in all IPD patients and controls (P < 0.001). However, the response to the challenge was not significantly different between patients and controls. INTERPRETATION: While the reported protonophore challenge assay is a valid marker of overall mitochondrial function in live platelets, intact mitochondrial membrane potential in platelets derived from IPD patients suggests that presumed mitochondrial enzymatic deficiencies are compensable in this cell type. In consequence, mitochondrial membrane potential in platelets cannot be used as a diagnostic biomarker for nonstratified IPD but should be further explored in potential Parkinson's disease subtypes and tissues with higher energy demands.

14.
Mol Neurobiol ; 49(1): 88-102, 2014 Feb.
Article in English | MEDLINE | ID: mdl-23832570

ABSTRACT

Parkinson's disease (PD) is a major neurodegenerative chronic disease, most likely caused by a complex interplay of genetic and environmental factors. Information on various aspects of PD pathogenesis is rapidly increasing and needs to be efficiently organized, so that the resulting data is available for exploration and analysis. Here we introduce a computationally tractable, comprehensive molecular interaction map of PD. This map integrates pathways implicated in PD pathogenesis such as synaptic and mitochondrial dysfunction, impaired protein degradation, alpha-synuclein pathobiology and neuroinflammation. We also present bioinformatics tools for the analysis, enrichment and annotation of the map, allowing the research community to open new avenues in PD research. The PD map is accessible at http://minerva.uni.lu/pd_map .


Subject(s)
Computational Biology/methods , Nerve Net/metabolism , Parkinson Disease/physiopathology , Proteolysis , Signal Transduction/physiology , Animals , Computational Biology/trends , Humans , Mesencephalon/metabolism , Mesencephalon/pathology , Mesencephalon/physiopathology , Nerve Net/pathology , Parkinson Disease/metabolism , Parkinson Disease/pathology
SELECTION OF CITATIONS
SEARCH DETAIL