Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 64
Filtrar
Más filtros

Banco de datos
Tipo del documento
Intervalo de año de publicación
1.
Nat Methods ; 20(6): 824-835, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-37069271

RESUMEN

BigNeuron is an open community bench-testing platform with the goal of setting open standards for accurate and fast automatic neuron tracing. We gathered a diverse set of image volumes across several species that is representative of the data obtained in many neuroscience laboratories interested in neuron tracing. Here, we report generated gold standard manual annotations for a subset of the available imaging datasets and quantified tracing quality for 35 automatic tracing algorithms. The goal of generating such a hand-curated diverse dataset is to advance the development of tracing algorithms and enable generalizable benchmarking. Together with image quality features, we pooled the data in an interactive web application that enables users and developers to perform principal component analysis, t-distributed stochastic neighbor embedding, correlation and clustering, visualization of imaging and tracing data, and benchmarking of automatic tracing algorithms in user-defined data subsets. The image quality metrics explain most of the variance in the data, followed by neuromorphological features related to neuron size. We observed that diverse algorithms can provide complementary information to obtain accurate results and developed a method to iteratively combine methods and generate consensus reconstructions. The consensus trees obtained provide estimates of the neuron structure ground truth that typically outperform single algorithms in noisy datasets. However, specific algorithms may outperform the consensus tree strategy in specific imaging conditions. Finally, to aid users in predicting the most accurate automatic tracing results without manual annotations for comparison, we used support vector machine regression to predict reconstruction quality given an image volume and a set of automatic tracings.


Asunto(s)
Benchmarking , Microscopía , Microscopía/métodos , Imagenología Tridimensional/métodos , Neuronas/fisiología , Algoritmos
2.
Angew Chem Int Ed Engl ; 63(12): e202319925, 2024 Mar 18.
Artículo en Inglés | MEDLINE | ID: mdl-38286754

RESUMEN

Anaerobes dominate the microbiota of the gastrointestinal (GI) tract, where a significant portion of small molecules can be degraded or modified. However, the enormous metabolic capacity of gut anaerobes remains largely elusive in contrast to aerobic bacteria, mainly due to the requirement of sophisticated laboratory settings. In this study, we employed an in silico machine learning platform, MoleculeX, to predict the metabolic capacity of a gut anaerobe, Clostridium sporogenes, against small molecules. Experiments revealed that among the top seven candidates predicted as unstable, six indeed exhibited instability in C. sporogenes culture. We further identified several metabolites resulting from the supplementation of everolimus in the bacterial culture for the first time. By utilizing bioinformatics and in vitro biochemical assays, we successfully identified an enzyme encoded in the genome of C. sporogenes responsible for everolimus transformation. Our framework thus can potentially facilitate future understanding of small molecules metabolism in the gut, further improve patient care through personalized medicine, and guide the development of new small molecule drugs and therapeutic approaches.


Asunto(s)
Clostridium , Everolimus , Humanos , Everolimus/metabolismo , Clostridium/metabolismo , Bacterias Anaerobias
3.
Bioinformatics ; 38(9): 2579-2586, 2022 04 28.
Artículo en Inglés | MEDLINE | ID: mdl-35179547

RESUMEN

MOTIVATION: Properties of molecules are indicative of their functions and thus are useful in many applications. With the advances of deep-learning methods, computational approaches for predicting molecular properties are gaining increasing momentum. However, there lacks customized and advanced methods and comprehensive tools for this task currently. RESULTS: Here, we develop a suite of comprehensive machine-learning methods and tools spanning different computational models, molecular representations and loss functions for molecular property prediction and drug discovery. Specifically, we represent molecules as both graphs and sequences. Built on these representations, we develop novel deep models for learning from molecular graphs and sequences. In order to learn effectively from highly imbalanced datasets, we develop advanced loss functions that optimize areas under precision-recall curves (PRCs) and receiver operating characteristic (ROC) curves. Altogether, our work not only serves as a comprehensive tool, but also contributes toward developing novel and advanced graph and sequence-learning methodologies. Results on both online and offline antibiotics discovery and molecular property prediction tasks show that our methods achieve consistent improvements over prior methods. In particular, our methods achieve #1 ranking in terms of both ROC-AUC (area under curve) and PRC-AUC on the AI Cures open challenge for drug discovery related to COVID-19. AVAILABILITY AND IMPLEMENTATION: Our source code is released as part of the MoleculeX library (https://github.com/divelab/MoleculeX) under AdvProp. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Asunto(s)
Tratamiento Farmacológico de COVID-19 , Humanos , Redes Neurales de la Computación , Programas Informáticos , Descubrimiento de Drogas , Aprendizaje Automático
4.
Bioinformatics ; 35(12): 2141-2149, 2019 06 01.
Artículo en Inglés | MEDLINE | ID: mdl-30398548

RESUMEN

MOTIVATION: Cellular function is closely related to the localizations of its sub-structures. It is, however, challenging to experimentally label all sub-cellular structures simultaneously in the same cell. This raises the need of building a computational model to learn the relationships among these sub-cellular structures and use reference structures to infer the localizations of other structures. RESULTS: We formulate such a task as a conditional image generation problem and propose to use conditional generative adversarial networks for tackling it. We employ an encoder-decoder network as the generator and propose to use skip connections between the encoder and decoder to provide spatial information to the decoder. To incorporate the conditional information in a variety of different ways, we develop three different types of skip connections, known as the self-gated connection, encoder-gated connection and label-gated connection. The proposed skip connections are built based on the conditional information using gating mechanisms. By learning a gating function, the network is able to control what information should be passed through the skip connections from the encoder to the decoder. Since the gate parameters are also learned automatically, we expect that only useful spatial information is transmitted to the decoder to help image generation. We perform both qualitative and quantitative evaluations to assess the effectiveness of our proposed approaches. Experimental results show that our cGAN-based approaches have the ability to generate the desired sub-cellular structures correctly. Our results also demonstrate that the proposed approaches outperform the existing approach based on adversarial auto-encoders, and the new skip connections lead to improved performance. In addition, the localizations of generated sub-cellular structures by our approaches are consistent with observations in biological experiments. AVAILABILITY AND IMPLEMENTATION: The source code and more results are available at https://github.com/divelab/cgan/.


Asunto(s)
Estructuras Celulares , Programas Informáticos
5.
Bioinformatics ; 33(16): 2555-2562, 2017 Aug 15.
Artículo en Inglés | MEDLINE | ID: mdl-28379412

RESUMEN

MOTIVATION: Progress in 3D electron microscopy (EM) imaging has greatly facilitated neuroscience research in high-throughput data acquisition. Correspondingly, high-throughput automated image analysis methods are necessary to work on par with the speed of data being produced. One such example is the need for automated EM image segmentation for neurite reconstruction. However, the efficiency and reliability of current methods are still lagging far behind human performance. RESULTS: Here, we propose DeepEM3D, a deep learning method for segmenting 3D anisotropic brain electron microscopy images. In this method, the deep learning model can efficiently build feature representation and incorporate sufficient multi-scale contextual information. We propose employing a combination of novel boundary map generation methods with optimized model ensembles to address the inherent challenges of segmenting anisotropic images. We evaluated our method by participating in the 3D segmentation of neurites in EM images (SNEMI3D) challenge. Our submission is ranked #1 on the current leaderboard as of Oct 15, 2016. More importantly, our result was very close to human-level performance in terms of the challenge evaluation metric: namely, a Rand error of 0.06015 versus the human value of 0.05998. AVAILABILITY AND IMPLEMENTATION: The code is available at https://github.com/divelab/deepem3d/. CONTACT: sji@eecs.wsu.edu. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Asunto(s)
Procesamiento de Imagen Asistido por Computador/métodos , Imagenología Tridimensional/métodos , Microscopía Electrónica/métodos , Neuritas/ultraestructura , Programas Informáticos , Algoritmos , Animales , Humanos , Neurociencias/métodos , Reproducibilidad de los Resultados
6.
Bioinformatics ; 32(15): 2352-8, 2016 08 01.
Artículo en Inglés | MEDLINE | ID: mdl-27153603

RESUMEN

MOTIVATION: Accurate segmentation of brain electron microscopy (EM) images is a critical step in dense circuit reconstruction. Although deep neural networks (DNNs) have been widely used in a number of applications in computer vision, most of these models that proved to be effective on image classification tasks cannot be applied directly to EM image segmentation, due to the different objectives of these tasks. As a result, it is desirable to develop an optimized architecture that uses the full power of DNNs and tailored specifically for EM image segmentation. RESULTS: In this work, we proposed a novel design of DNNs for this task. We trained a pixel classifier that operates on raw pixel intensities with no preprocessing to generate probability values for each pixel being a membrane or not. Although the use of neural networks in image segmentation is not completely new, we developed novel insights and model architectures that allow us to achieve superior performance on EM image segmentation tasks. Our submission based on these insights to the 2D EM Image Segmentation Challenge achieved the best performance consistently across all the three evaluation metrics. This challenge is still ongoing and the results in this paper are as of June 5, 2015. AVAILABILITY AND IMPLEMENTATION: https://github.com/ahmed-fakhry/dive CONTACT: : sji@eecs.wsu.edu.


Asunto(s)
Encéfalo , Microscopía Electrónica , Redes Neurales de la Computación , Algoritmos , Modelos Teóricos
7.
Methods ; 73: 71-8, 2015 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-25109429

RESUMEN

The brain is a multi-level system in which the high-level functions are generated by low-level genetic mechanisms. Thus, elucidating the relationship among multiple brain levels via correlative and predictive analytics is an important area in brain research. Currently, studies in multiple species have indicated that the spatiotemporal gene expression patterns are predictive of brain wiring. Specifically, results on the worm Caenorhabditis elegans have shown that the prediction of neuronal connectivity using gene expression signatures yielded statistically significant results. Recent studies on the mammalian brain produced similar results at the coarse regional level. In this study, we provide the first high-resolution, large-scale integrative analysis of the transcriptome and connectome in a single mammalian brain at a fine voxel level. By using the Allen Brain Atlas data, we predict voxel-level brain connectivity based on the gene expressions in the adult mouse brain. We employ regularized models to show that gene expression is predictive of connectivity at the voxel-level with an accuracy of 93%. We also identify a set of genes playing the most important role in connectivity prediction. We use only this small number of genes to predict the brain wiring with an accuracy over 80%. We discover that these important genes are enriched in neurons as compared to glia, and they perform connectivity-related functions. We perform several interesting correlative studies to further elucidate the transcriptome-connectome relationship.


Asunto(s)
Atlas como Asunto , Mapeo Encefálico/métodos , Encéfalo/anatomía & histología , Encéfalo/fisiología , Conectoma/métodos , Regulación de la Expresión Génica , Animales , Predicción , Masculino , Ratones , Ratones Endogámicos C57BL
8.
BMC Bioinformatics ; 16: 147, 2015 May 07.
Artículo en Inglés | MEDLINE | ID: mdl-25948335

RESUMEN

BACKGROUND: Profiling gene expression in brain structures at various spatial and temporal scales is essential to understanding how genes regulate the development of brain structures. The Allen Developing Mouse Brain Atlas provides high-resolution 3-D in situ hybridization (ISH) gene expression patterns in multiple developing stages of the mouse brain. Currently, the ISH images are annotated with anatomical terms manually. In this paper, we propose a computational approach to annotate gene expression pattern images in the mouse brain at various structural levels over the course of development. RESULTS: We applied deep convolutional neural network that was trained on a large set of natural images to extract features from the ISH images of developing mouse brain. As a baseline representation, we applied invariant image feature descriptors to capture local statistics from ISH images and used the bag-of-words approach to build image-level representations. Both types of features from multiple ISH image sections of the entire brain were then combined to build 3-D, brain-wide gene expression representations. We employed regularized learning methods for discriminating gene expression patterns in different brain structures. Results show that our approach of using convolutional model as feature extractors achieved superior performance in annotating gene expression patterns at multiple levels of brain structures throughout four developing ages. Overall, we achieved average AUC of 0.894 ± 0.014, as compared with 0.820 ± 0.046 yielded by the bag-of-words approach. CONCLUSIONS: Deep convolutional neural network model trained on natural image sets and applied to gene expression pattern annotation tasks yielded superior performance, demonstrating its transfer learning property is applicable to such biological image sets.


Asunto(s)
Encéfalo/metabolismo , Regulación del Desarrollo de la Expresión Génica , Anotación de Secuencia Molecular , Redes Neurales de la Computación , Reconocimiento de Normas Patrones Automatizadas , Animales , Encéfalo/crecimiento & desarrollo , Perfilación de la Expresión Génica/métodos , Procesamiento de Imagen Asistido por Computador , Hibridación in Situ , Ratones
9.
Neuroimage ; 108: 214-24, 2015 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-25562829

RESUMEN

The segmentation of infant brain tissue images into white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) plays an important role in studying early brain development in health and disease. In the isointense stage (approximately 6-8 months of age), WM and GM exhibit similar levels of intensity in both T1 and T2 MR images, making the tissue segmentation very challenging. Only a small number of existing methods have been designed for tissue segmentation in this isointense stage; however, they only used a single T1 or T2 images, or the combination of T1 and T2 images. In this paper, we propose to use deep convolutional neural networks (CNNs) for segmenting isointense stage brain tissues using multi-modality MR images. CNNs are a type of deep models in which trainable filters and local neighborhood pooling operations are applied alternatingly on the raw input images, resulting in a hierarchy of increasingly complex features. Specifically, we used multi-modality information from T1, T2, and fractional anisotropy (FA) images as inputs and then generated the segmentation maps as outputs. The multiple intermediate layers applied convolution, pooling, normalization, and other operations to capture the highly nonlinear mappings between inputs and outputs. We compared the performance of our approach with that of the commonly used segmentation methods on a set of manually segmented isointense stage brain images. Results showed that our proposed model significantly outperformed prior methods on infant brain tissue segmentation. In addition, our results indicated that integration of multi-modality images led to significant performance improvement.


Asunto(s)
Mapeo Encefálico/métodos , Encéfalo/anatomía & histología , Procesamiento de Imagen Asistido por Computador/métodos , Imagen por Resonancia Magnética/métodos , Redes Neurales de la Computación , Anisotropía , Sustancia Gris , Humanos , Lactante , Sustancia Blanca
10.
Bioinformatics ; 30(2): 266-73, 2014 Jan 15.
Artículo en Inglés | MEDLINE | ID: mdl-24300439

RESUMEN

MOTIVATION: Drosophila melanogaster is a major model organism for investigating the function and interconnection of animal genes in the earliest stages of embryogenesis. Today, images capturing Drosophila gene expression patterns are being produced at a higher throughput than ever before. The analysis of spatial patterns of gene expression is most biologically meaningful when images from a similar time point during development are compared. Thus, the critical first step is to determine the developmental stage of an embryo. This information is also needed to observe and analyze expression changes over developmental time. Currently, developmental stages (time) of embryos in images capturing spatial expression pattern are annotated manually, which is time- and labor-intensive. Embryos are often designated into stage ranges, making the information on developmental time course. This makes downstream analyses inefficient and biological interpretations of similarities and differences in spatial expression patterns challenging, particularly when using automated tools for analyzing expression patterns of large number of images. RESULTS: Here, we present a new computational approach to annotate developmental stage for Drosophila embryos in the gene expression images. In an analysis of 3724 images, the new approach shows high accuracy in predicting the developmental stage correctly (79%). In addition, it provides a stage score that enables one to more finely annotate each embryo so that they are divided into early and late periods of development within standard stage demarcations. Stage scores for all images containing expression patterns of the same gene enable a direct way to view expression changes over developmental time for any gene. We show that the genomewide-expression-maps generated using images from embryos in refined stages illuminate global gene activities and changes much better, and more refined stage annotations improve our ability to better interpret results when expression pattern matches are discovered between genes. AVAILABILITY AND IMPLEMENTATION: The software package is availablefor download at: http://www.public.asu.edu/*jye02/Software/Fly-Project/.


Asunto(s)
Biología Computacional , Proteínas de Drosophila/genética , Drosophila melanogaster/genética , Embrión no Mamífero/citología , Perfilación de la Expresión Génica , Regulación del Desarrollo de la Expresión Génica , Procesamiento de Imagen Asistido por Computador , Algoritmos , Animales , Drosophila melanogaster/embriología , Embrión no Mamífero/metabolismo , Desarrollo Embrionario/genética , Reconocimiento de Normas Patrones Automatizadas
11.
BMC Bioinformatics ; 15: 209, 2014 Jun 20.
Artículo en Inglés | MEDLINE | ID: mdl-24947138

RESUMEN

BACKGROUND: Differential gene expression patterns in cells of the mammalian brain result in the morphological, connectional, and functional diversity of cells. A wide variety of studies have shown that certain genes are expressed only in specific cell-types. Analysis of cell-type-specific gene expression patterns can provide insights into the relationship between genes, connectivity, brain regions, and cell-types. However, automated methods for identifying cell-type-specific genes are lacking to date. RESULTS: Here, we describe a set of computational methods for identifying cell-type-specific genes in the mouse brain by automated image computing of in situ hybridization (ISH) expression patterns. We applied invariant image feature descriptors to capture local gene expression information from cellular-resolution ISH images. We then built image-level representations by applying vector quantization on the image descriptors. We employed regularized learning methods for classifying genes specifically expressed in different brain cell-types. These methods can also rank image features based on their discriminative power. We used a data set of 2,872 genes from the Allen Brain Atlas in the experiments. Results showed that our methods are predictive of cell-type-specificity of genes. Our classifiers achieved AUC values of approximately 87% when the enrichment level is set to 20. In addition, we showed that the highly-ranked image features captured the relationship between cell-types. CONCLUSIONS: Overall, our results showed that automated image computing methods could potentially be used to identify cell-type-specific genes in the mouse brain.


Asunto(s)
Automatización de Laboratorios/métodos , Encéfalo , Perfilación de la Expresión Génica/métodos , Animales , Encéfalo/metabolismo , Expresión Génica , Hibridación in Situ , Ratones , Especificidad de Órganos
12.
Neuroimage ; 84: 245-53, 2014 Jan 01.
Artículo en Inglés | MEDLINE | ID: mdl-24004696

RESUMEN

Brain function is the result of interneuron signal transmission controlled by the fundamental biochemistry of each neuron. The biochemical content of a neuron is in turn determined by spatiotemporal gene expression and regulation encoded into the genomic regulatory networks. It is thus of particular interest to elucidate the relationship between gene expression patterns and connectivity in the brain. However, systematic studies of this relationship in a single mammalian brain are lacking to date. Here, we investigate this relationship in the mouse brain using the Allen Brain Atlas data. We employ computational models for predicting brain connectivity from gene expression data. In addition to giving competitive predictive performance, these models can rank the genes according to their predictive power. We show that gene expression is predictive of connectivity in the mouse brain when the connectivity signals are discretized. When the expression patterns of 4084 genes are used, we obtain a predictive accuracy of 93%. Our results also show that a small number of genes can almost give the full predictive power of using thousands of genes. We can achieve a prediction accuracy of 91% by using only 25 genes. Gene ontology analysis of the highly ranked genes shows that they are enriched for connectivity related processes.


Asunto(s)
Encéfalo/anatomía & histología , Encéfalo/metabolismo , Conectoma/métodos , Modelos Anatómicos , Modelos Neurológicos , Proteínas del Tejido Nervioso/metabolismo , Animales , Simulación por Computador , Perfilación de la Expresión Génica/métodos , Regulación de la Expresión Génica/fisiología , Masculino , Ratones , Integración de Sistemas , Distribución Tisular
13.
IEEE Trans Pattern Anal Mach Intell ; 46(7): 4567-4578, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38147422

RESUMEN

We investigate the explainability of graph neural networks (GNNs) as a step toward elucidating their working mechanisms. While most current methods focus on explaining graph nodes, edges, or features, we argue that, as the inherent functional mechanism of GNNs, message flows are more natural for performing explainability. To this end, we propose a novel method here, known as FlowX, to explain GNNs by identifying important message flows. To quantify the importance of flows, we propose to follow the philosophy of Shapley values from cooperative game theory. To tackle the complexity of computing all coalitions' marginal contributions, we propose a flow sampling scheme to compute Shapley value approximations as initial assessments of further training. We then propose an information-controlled learning algorithm to train flow scores toward diverse explanation targets: necessary or sufficient explanations. Experimental studies on both synthetic and real-world datasets demonstrate that our proposed FlowX and its variants lead to improved explainability of GNNs.

14.
Commun Biol ; 7(1): 414, 2024 Apr 05.
Artículo en Inglés | MEDLINE | ID: mdl-38580839

RESUMEN

Understanding the genetic architecture of brain structure is challenging, partly due to difficulties in designing robust, non-biased descriptors of brain morphology. Until recently, brain measures for genome-wide association studies (GWAS) consisted of traditionally expert-defined or software-derived image-derived phenotypes (IDPs) that are often based on theoretical preconceptions or computed from limited amounts of data. Here, we present an approach to derive brain imaging phenotypes using unsupervised deep representation learning. We train a 3-D convolutional autoencoder model with reconstruction loss on 6130 UK Biobank (UKBB) participants' T1 or T2-FLAIR (T2) brain MRIs to create a 128-dimensional representation known as Unsupervised Deep learning derived Imaging Phenotypes (UDIPs). GWAS of these UDIPs in held-out UKBB subjects (n = 22,880 discovery and n = 12,359/11,265 replication cohorts for T1/T2) identified 9457 significant SNPs organized into 97 independent genetic loci of which 60 loci were replicated. Twenty-six loci were not reported in earlier T1 and T2 IDP-based UK Biobank GWAS. We developed a perturbation-based decoder interpretation approach to show that these loci are associated with UDIPs mapped to multiple relevant brain regions. Our results established unsupervised deep learning can derive robust, unbiased, heritable, and interpretable brain imaging phenotypes.


Asunto(s)
Sitios Genéticos , Estudio de Asociación del Genoma Completo , Humanos , Estudio de Asociación del Genoma Completo/métodos , Fenotipo , Encéfalo/diagnóstico por imagen , Neuroimagen
15.
BMC Bioinformatics ; 14: 222, 2013 Jul 11.
Artículo en Inglés | MEDLINE | ID: mdl-23845024

RESUMEN

BACKGROUND: The structured organization of cells in the brain plays a key role in its functional efficiency. This delicate organization is the consequence of unique molecular identity of each cell gradually established by precise spatiotemporal gene expression control during development. Currently, studies on the molecular-structural association are beginning to reveal how the spatiotemporal gene expression patterns are related to cellular differentiation and structural development. RESULTS: In this article, we aim at a global, data-driven study of the relationship between gene expressions and neuroanatomy in the developing mouse brain. To enable visual explorations of the high-dimensional data, we map the in situ hybridization gene expression data to a two-dimensional space by preserving both the global and the local structures. Our results show that the developing brain anatomy is largely preserved in the reduced gene expression space. To provide a quantitative analysis, we cluster the reduced data into groups and measure the consistency with neuroanatomy at multiple levels. Our results show that the clusters in the low-dimensional space are more consistent with neuroanatomy than those in the original space. CONCLUSIONS: Gene expression patterns and developing brain anatomy are closely related. Dimensionality reduction and visual exploration facilitate the study of this relationship.


Asunto(s)
Encéfalo/metabolismo , Perfilación de la Expresión Génica/métodos , Animales , Encéfalo/anatomía & histología , Encéfalo/embriología , Encéfalo/crecimiento & desarrollo , Análisis por Conglomerados , Biología Computacional/métodos , Gráficos por Computador , Hibridación in Situ , Ratones
16.
BMC Bioinformatics ; 14: 350, 2013 Dec 03.
Artículo en Inglés | MEDLINE | ID: mdl-24299119

RESUMEN

BACKGROUND: Drosophila melanogaster has been established as a model organism for investigating the developmental gene interactions. The spatio-temporal gene expression patterns of Drosophila melanogaster can be visualized by in situ hybridization and documented as digital images. Automated and efficient tools for analyzing these expression images will provide biological insights into the gene functions, interactions, and networks. To facilitate pattern recognition and comparison, many web-based resources have been created to conduct comparative analysis based on the body part keywords and the associated images. With the fast accumulation of images from high-throughput techniques, manual inspection of images will impose a serious impediment on the pace of biological discovery. It is thus imperative to design an automated system for efficient image annotation and comparison. RESULTS: We present a computational framework to perform anatomical keywords annotation for Drosophila gene expression images. The spatial sparse coding approach is used to represent local patches of images in comparison with the well-known bag-of-words (BoW) method. Three pooling functions including max pooling, average pooling and Sqrt (square root of mean squared statistics) pooling are employed to transform the sparse codes to image features. Based on the constructed features, we develop both an image-level scheme and a group-level scheme to tackle the key challenges in annotating Drosophila gene expression pattern images automatically. To deal with the imbalanced data distribution inherent in image annotation tasks, the undersampling method is applied together with majority vote. Results on Drosophila embryonic expression pattern images verify the efficacy of our approach. CONCLUSION: In our experiment, the three pooling functions perform comparably well in feature dimension reduction. The undersampling with majority vote is shown to be effective in tackling the problem of imbalanced data. Moreover, combining sparse coding and image-level scheme leads to consistent performance improvement in keywords annotation.


Asunto(s)
Drosophila melanogaster/citología , Drosophila melanogaster/genética , Regulación del Desarrollo de la Expresión Génica , Genoma de los Insectos/genética , Modelos Genéticos , Anotación de Secuencia Molecular/métodos , Animales , Diferenciación Celular/genética , División Celular/genética , Biología Computacional/clasificación , Biología Computacional/métodos , Drosophila melanogaster/embriología , Perfilación de la Expresión Génica/clasificación , Perfilación de la Expresión Génica/métodos , Ensayos Analíticos de Alto Rendimiento , Anotación de Secuencia Molecular/clasificación , Valor Predictivo de las Pruebas , Máquina de Vectores de Soporte
17.
BMC Bioinformatics ; 14: 372, 2013 Dec 28.
Artículo en Inglés | MEDLINE | ID: mdl-24373308

RESUMEN

BACKGROUND: Multicellular organisms consist of cells of many different types that are established during development. Each type of cell is characterized by the unique combination of expressed gene products as a result of spatiotemporal gene regulation. Currently, a fundamental challenge in regulatory biology is to elucidate the gene expression controls that generate the complex body plans during development. Recent advances in high-throughput biotechnologies have generated spatiotemporal expression patterns for thousands of genes in the model organism fruit fly Drosophila melanogaster. Existing qualitative methods enhanced by a quantitative analysis based on computational tools we present in this paper would provide promising ways for addressing key scientific questions. RESULTS: We develop a set of computational methods and open source tools for identifying co-expressed embryonic domains and the associated genes simultaneously. To map the expression patterns of many genes into the same coordinate space and account for the embryonic shape variations, we develop a mesh generation method to deform a meshed generic ellipse to each individual embryo. We then develop a co-clustering formulation to cluster the genes and the mesh elements, thereby identifying co-expressed embryonic domains and the associated genes simultaneously. Experimental results indicate that the gene and mesh co-clusters can be correlated to key developmental events during the stages of embryogenesis we study. The open source software tool has been made available at http://compbio.cs.odu.edu/fly/. CONCLUSIONS: Our mesh generation and machine learning methods and tools improve upon the flexibility, ease-of-use and accuracy of existing methods.


Asunto(s)
Inteligencia Artificial , Biología Computacional/métodos , Regulación del Desarrollo de la Expresión Génica , Procesamiento de Imagen Asistido por Computador/métodos , Procesamiento de Imagen Asistido por Computador/normas , Máquina de Vectores de Soporte , Animales , Inteligencia Artificial/normas , Análisis por Conglomerados , Biología Computacional/normas , Drosophila/embriología , Drosophila/genética , Perfilación de la Expresión Génica/métodos , Procesamiento de Imagen Asistido por Computador/estadística & datos numéricos , Programas Informáticos
18.
IEEE Trans Pattern Anal Mach Intell ; 45(6): 6870-6880, 2023 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-32750778

RESUMEN

Graph neural networks have achieved great success in learning node representations for graph tasks such as node classification and link prediction. Graph representation learning requires graph pooling to obtain graph representations from node representations. It is challenging to develop graph pooling methods due to the variable sizes and isomorphic structures of graphs. In this work, we propose to use second-order pooling as graph pooling, which naturally solves the above challenges. In addition, compared to existing graph pooling methods, second-order pooling is able to use information from all nodes and collect second-order statistics, making it more powerful. We show that direct use of second-order pooling with graph neural networks leads to practical problems. To overcome these problems, we propose two novel global graph pooling methods based on second-order pooling; namely, bilinear mapping and attentional second-order pooling. In addition, we extend attentional second-order pooling to hierarchical graph pooling for more flexible use in GNNs. We perform thorough experiments on graph classification tasks to demonstrate the effectiveness and superiority of our proposed methods. Experimental results show that our methods improve the performance significantly and consistently.

19.
IEEE Trans Pattern Anal Mach Intell ; 45(3): 3169-3180, 2023 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-35604976

RESUMEN

We study self-supervised learning on graphs using contrastive methods. A general scheme of prior methods is to optimize two-view representations of input graphs. In many studies, a single graph-level representation is computed as one of the contrastive objectives, capturing limited characteristics of graphs. We argue that contrasting graphs in multiple subspaces enables graph encoders to capture more abundant characteristics. To this end, we propose a group contrastive learning framework in this work. Our framework embeds the given graph into multiple subspaces, of which each representation is prompted to encode specific characteristics of graphs. To learn diverse and informative representations, we develop principled objectives that enable us to capture the relations among both intra-space and inter-space representations in groups. Under the proposed framework, we further develop an attention-based group generator to compute representations that capture different substructures of a given graph. Built upon our framework, we extend two current methods into GroupCL and GroupIG, equipped with the proposed objective. Comprehensive experimental results show our framework achieves a promising boost in performance on a variety of datasets. In addition, our qualitative results show that features generated from our representor successfully capture various specific characteristics of graphs.

20.
IEEE Trans Pattern Anal Mach Intell ; 45(5): 5782-5799, 2023 05.
Artículo en Inglés | MEDLINE | ID: mdl-36063508

RESUMEN

Deep learning methods are achieving ever-increasing performance on many artificial intelligence tasks. A major limitation of deep models is that they are not amenable to interpretability. This limitation can be circumvented by developing post hoc techniques to explain predictions, giving rise to the area of explainability. Recently, explainability of deep models on images and texts has achieved significant progress. In the area of graph data, graph neural networks (GNNs) and their explainability are experiencing rapid developments. However, there is neither a unified treatment of GNN explainability methods, nor a standard benchmark and testbed for evaluations. In this survey, we provide a unified and taxonomic view of current GNN explainability methods. Our unified and taxonomic treatments of this subject shed lights on the commonalities and differences of existing methods and set the stage for further methodological developments. To facilitate evaluations, we provide a testbed for GNN explainability, including datasets, common algorithms and evaluation metrics. Furthermore, we conduct comprehensive experiments to compare and analyze the performance of many techniques. Altogether, this work provides a unified methodological treatment of GNN explainability and a standardized testbed for evaluations.


Asunto(s)
Algoritmos , Inteligencia Artificial , Redes Neurales de la Computación , Benchmarking
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA