Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 8 de 8
Filter
Add more filters










Database
Language
Publication year range
1.
J Neurosci ; 43(10): 1731-1741, 2023 03 08.
Article in English | MEDLINE | ID: mdl-36759190

ABSTRACT

Deep neural networks (DNNs) are promising models of the cortical computations supporting human object recognition. However, despite their ability to explain a significant portion of variance in neural data, the agreement between models and brain representational dynamics is far from perfect. We address this issue by asking which representational features are currently unaccounted for in neural time series data, estimated for multiple areas of the ventral stream via source-reconstructed magnetoencephalography data acquired in human participants (nine females, six males) during object viewing. We focus on the ability of visuo-semantic models, consisting of human-generated labels of object features and categories, to explain variance beyond the explanatory power of DNNs alone. We report a gradual reversal in the relative importance of DNN versus visuo-semantic features as ventral-stream object representations unfold over space and time. Although lower-level visual areas are better explained by DNN features starting early in time (at 66 ms after stimulus onset), higher-level cortical dynamics are best accounted for by visuo-semantic features starting later in time (at 146 ms after stimulus onset). Among the visuo-semantic features, object parts and basic categories drive the advantage over DNNs. These results show that a significant component of the variance unexplained by DNNs in higher-level cortical dynamics is structured and can be explained by readily nameable aspects of the objects. We conclude that current DNNs fail to fully capture dynamic representations in higher-level human visual cortex and suggest a path toward more accurate models of ventral-stream computations.SIGNIFICANCE STATEMENT When we view objects such as faces and cars in our visual environment, their neural representations dynamically unfold over time at a millisecond scale. These dynamics reflect the cortical computations that support fast and robust object recognition. DNNs have emerged as a promising framework for modeling these computations but cannot yet fully account for the neural dynamics. Using magnetoencephalography data acquired in human observers during object viewing, we show that readily nameable aspects of objects, such as 'eye', 'wheel', and 'face', can account for variance in the neural dynamics over and above DNNs. These findings suggest that DNNs and humans may in part rely on different object features for visual recognition and provide guidelines for model improvement.


Subject(s)
Pattern Recognition, Visual , Semantics , Male , Female , Humans , Neural Networks, Computer , Visual Perception , Brain , Brain Mapping/methods , Magnetic Resonance Imaging/methods
2.
Commun Biol ; 5(1): 1247, 2022 11 14.
Article in English | MEDLINE | ID: mdl-36376446

ABSTRACT

Distinguishing animate from inanimate things is of great behavioural importance. Despite distinct brain and behavioural responses to animate and inanimate things, it remains unclear which object properties drive these responses. Here, we investigate the importance of five object dimensions related to animacy ("being alive", "looking like an animal", "having agency", "having mobility", and "being unpredictable") in brain (fMRI, EEG) and behaviour (property and similarity judgements) of 19 participants. We used a stimulus set of 128 images, optimized by a genetic algorithm to disentangle these five dimensions. The five dimensions explained much variance in the similarity judgments. Each dimension explained significant variance in the brain representations (except, surprisingly, "being alive"), however, to a lesser extent than in behaviour. Different brain regions sensitive to animacy may represent distinct dimensions, either as accessible perceptual stepping stones toward detecting whether something is alive or because they are of behavioural importance in their own right.


Subject(s)
Brain , Pattern Recognition, Visual , Humans , Pattern Recognition, Visual/physiology , Brain/diagnostic imaging , Brain/physiology , Brain Mapping , Magnetic Resonance Imaging/methods , Judgment/physiology
3.
Proc Natl Acad Sci U S A ; 119(27): e2115047119, 2022 07 05.
Article in English | MEDLINE | ID: mdl-35767642

ABSTRACT

Human vision is attuned to the subtle differences between individual faces. Yet we lack a quantitative way of predicting how similar two face images look and whether they appear to show the same person. Principal component-based three-dimensional (3D) morphable models are widely used to generate stimuli in face perception research. These models capture the distribution of real human faces in terms of dimensions of physical shape and texture. How well does a "face space" based on these dimensions capture the similarity relationships humans perceive among faces? To answer this, we designed a behavioral task to collect dissimilarity and same/different identity judgments for 232 pairs of realistic faces. Stimuli sampled geometric relationships in a face space derived from principal components of 3D shape and texture (Basel face model [BFM]). We then compared a wide range of models in their ability to predict the data, including the BFM from which faces were generated, an active appearance model derived from face photographs, and image-computable models of visual perception. Euclidean distance in the BFM explained both dissimilarity and identity judgments surprisingly well. In a comparison against 16 diverse models, BFM distance was competitive with representational distances in state-of-the-art deep neural networks (DNNs), including novel DNNs trained on BFM synthetic identities or BFM latents. Models capturing the distribution of face shape and texture across individuals are not only useful tools for stimulus generation. They also capture important information about how faces are perceived, suggesting that human face representations are tuned to the statistical distribution of faces.


Subject(s)
Facial Recognition , Judgment , Visual Perception , Humans , Neural Networks, Computer
4.
Neuroimage ; 194: 12-24, 2019 07 01.
Article in English | MEDLINE | ID: mdl-30894333

ABSTRACT

The degree to which we perceive real-world objects as similar or dissimilar structures our perception and guides categorization behavior. Here, we investigated the neural representations enabling perceived similarity using behavioral judgments, fMRI and MEG. As different object dimensions co-occur and partly correlate, to understand the relationship between perceived similarity and brain activity it is necessary to assess the unique role of multiple object dimensions. We thus behaviorally assessed perceived object similarity in relation to shape, function, color and background. We then used representational similarity analyses to relate these behavioral judgments to brain activity. We observed a link between each object dimension and representations in visual cortex. These representations emerged rapidly within 200 ms of stimulus onset. Assessing the unique role of each object dimension revealed partly overlapping and distributed representations: while color-related representations distinctly preceded shape-related representations both in the processing hierarchy of the ventral visual pathway and in time, several dimensions were linked to high-level ventral visual cortex. Further analysis singled out the shape dimension as neither fully accounted for by supra-category membership, nor a deep neural network trained on object categorization. Together our results comprehensively characterize the relationship between perceived similarity of key object dimensions and neural activity.


Subject(s)
Pattern Recognition, Visual/physiology , Visual Cortex/physiology , Adult , Brain Mapping/methods , Female , Humans , Male
5.
Front Psychol ; 8: 1726, 2017.
Article in English | MEDLINE | ID: mdl-29062291

ABSTRACT

Recent advances in Deep convolutional Neural Networks (DNNs) have enabled unprecedentedly accurate computational models of brain representations, and present an exciting opportunity to model diverse cognitive functions. State-of-the-art DNNs achieve human-level performance on object categorisation, but it is unclear how well they capture human behavior on complex cognitive tasks. Recent reports suggest that DNNs can explain significant variance in one such task, judging object similarity. Here, we extend these findings by replicating them for a rich set of object images, comparing performance across layers within two DNNs of different depths, and examining how the DNNs' performance compares to that of non-computational "conceptual" models. Human observers performed similarity judgments for a set of 92 images of real-world objects. Representations of the same images were obtained in each of the layers of two DNNs of different depths (8-layer AlexNet and 16-layer VGG-16). To create conceptual models, other human observers generated visual-feature labels (e.g., "eye") and category labels (e.g., "animal") for the same image set. Feature labels were divided into parts, colors, textures and contours, while category labels were divided into subordinate, basic, and superordinate categories. We fitted models derived from the features, categories, and from each layer of each DNN to the similarity judgments, using representational similarity analysis to evaluate model performance. In both DNNs, similarity within the last layer explains most of the explainable variance in human similarity judgments. The last layer outperforms almost all feature-based models. Late and mid-level layers outperform some but not all feature-based models. Importantly, categorical models predict similarity judgments significantly better than any DNN layer. Our results provide further evidence for commonalities between DNNs and brain representations. Models derived from visual features other than object parts perform relatively poorly, perhaps because DNNs more comprehensively capture the colors, textures and contours which matter to human object perception. However, categorical models outperform DNNs, suggesting that further work may be needed to bring high-level semantic representations in DNNs closer to those extracted by humans. Modern DNNs explain similarity judgments remarkably well considering they were not trained on this task, and are promising models for many aspects of human cognition.

6.
Cell Rep ; 17(10): 2715-2723, 2016 12 06.
Article in English | MEDLINE | ID: mdl-27926873

ABSTRACT

FOXA1 is a pioneer factor that binds to enhancer regions that are enriched in H3K4 mono- and dimethylation (H3K4me1 and H3K4me2). We performed a FOXA1 rapid immunoprecipitation mass spectrometry of endogenous proteins (RIME) screen in ERα-positive MCF-7 breast cancer cells and found histone-lysine N-methyltransferase (MLL3) as the top FOXA1-interacting protein. MLL3 is typically thought to induce H3K4me3 at promoter regions, but recent findings suggest it may contribute to H3K4me1 deposition. We performed MLL3 chromatin immunoprecipitation sequencing (ChIP-seq) in breast cancer cells, and MLL3 was shown to occupy regions marked by FOXA1 occupancy and H3K4me1 and H3K4me2. MLL3 binding was dependent on FOXA1, indicating that FOXA1 recruits MLL3 to chromatin. MLL3 silencing decreased H3K4me1 at enhancer elements but had no appreciable impact on H3K4me3 at enhancer elements. We propose a mechanism whereby the pioneer factor FOXA1 recruits the chromatin modifier MLL3 to facilitate the deposition of H3K4me1 histone marks, subsequently demarcating active enhancer elements.


Subject(s)
Breast Neoplasms/genetics , Chromatin/genetics , DNA-Binding Proteins/genetics , Hepatocyte Nuclear Factor 3-alpha/genetics , Breast Neoplasms/pathology , DNA Methylation/genetics , Enhancer Elements, Genetic , Estrogen Receptor alpha/genetics , Female , Gene Expression Regulation, Neoplastic , Histones/genetics , Humans , MCF-7 Cells , Promoter Regions, Genetic
7.
Neuropsychologia ; 83: 201-226, 2016 Mar.
Article in English | MEDLINE | ID: mdl-26493748

ABSTRACT

Object similarity, in brain representations and conscious perception, must reflect a combination of the visual appearance of the objects on the one hand and the categories the objects belong to on the other. Indeed, visual object features and category membership have each been shown to contribute to the object representation in human inferior temporal (IT) cortex, as well as to object-similarity judgments. However, the explanatory power of features and categories has not been directly compared. Here, we investigate whether the IT object representation and similarity judgments are best explained by a categorical or a feature-based model. We use rich models (>100 dimensions) generated by human observers for a set of 96 real-world object images. The categorical model consists of a hierarchically nested set of category labels (such as "human", "mammal", and "animal"). The feature-based model includes both object parts (such as "eye", "tail", and "handle") and other descriptive features (such as "circular", "green", and "stubbly"). We used non-negative least squares to fit the models to the brain representations (estimated from functional magnetic resonance imaging data) and to similarity judgments. Model performance was estimated on held-out images not used in fitting. Both models explained significant variance in IT and the amounts explained were not significantly different. The combined model did not explain significant additional IT variance, suggesting that it is the shared model variance (features correlated with categories, categories correlated with features) that best explains IT. The similarity judgments were almost fully explained by the categorical model, which explained significantly more variance than the feature-based model. The combined model did not explain significant additional variance in the similarity judgments. Our findings suggest that IT uses features that help to distinguish categories as stepping stones toward a semantic representation. Similarity judgments contain additional categorical variance that is not explained by visual features, reflecting a higher-level more purely semantic representation.


Subject(s)
Brain Mapping , Brain/physiology , Judgment/physiology , Least-Squares Analysis , Pattern Recognition, Visual/physiology , Semantics , Adult , Brain/blood supply , Concept Formation , Female , Humans , Image Processing, Computer-Assisted , Magnetic Resonance Imaging , Male , Models, Theoretical , Oxygen/blood , Photic Stimulation
8.
Nat Rev Cancer ; 12(6): 381-5, 2012 May 04.
Article in English | MEDLINE | ID: mdl-22555282

ABSTRACT

Pioneer factors are a special class of transcription factor that can associate with compacted chromatin to facilitate the binding of additional transcription factors. The function of pioneer factors was originally described during development; more recently, they have been implicated in hormone-dependent cancers, such as oestrogen receptor-positive breast cancer and androgen receptor-positive prostate cancer. We discuss the importance of pioneer factors in these specific cancers, the discovery of new putative pioneer factors and the interplay between these proteins in mediating nuclear receptor function in cancer.


Subject(s)
Hormones/metabolism , Neoplasms/metabolism , Transcription Factors/genetics , Transcription Factors/metabolism , Breast Neoplasms/genetics , Breast Neoplasms/metabolism , DNA-Binding Proteins/genetics , DNA-Binding Proteins/metabolism , Drug Design , Female , Gene Expression Regulation , Hepatocyte Nuclear Factor 3-alpha/genetics , Hepatocyte Nuclear Factor 3-alpha/metabolism , Humans , Male , Neoplasms/genetics , Pre-B-Cell Leukemia Transcription Factor 1 , Prostatic Neoplasms/genetics , Prostatic Neoplasms/metabolism , Proto-Oncogene Proteins/genetics , Proto-Oncogene Proteins/metabolism , Receptors, Androgen/metabolism , Receptors, Estrogen/metabolism
SELECTION OF CITATIONS
SEARCH DETAIL
...