Your browser doesn't support javascript.
loading
Hand-Selective Visual Regions Represent How to Grasp 3D Tools: Brain Decoding during Real Actions.
Knights, Ethan; Mansfield, Courtney; Tonin, Diana; Saada, Janak; Smith, Fraser W; Rossit, Stéphanie.
Afiliação
  • Knights E; Medical Research Council Cognition and Brain Sciences Unit, University of Cambridge, Cambridge CB2 7EF, United Kingdom.
  • Mansfield C; School of Psychology, University of East Anglia, Norwich NR4 7TJ, United Kingdom.
  • Tonin D; School of Psychology, University of East Anglia, Norwich NR4 7TJ, United Kingdom.
  • Saada J; Department of Radiology, Norfolk and Norwich University Hospitals NHS Foundation Trust, Norwich NR4 7UY, United Kingdom.
  • Smith FW; School of Psychology, University of East Anglia, Norwich NR4 7TJ, United Kingdom.
  • Rossit S; School of Psychology, University of East Anglia, Norwich NR4 7TJ, United Kingdom s.rossit@uea.ac.uk.
J Neurosci ; 41(24): 5263-5273, 2021 06 16.
Article em En | MEDLINE | ID: mdl-33972399
ABSTRACT
Most neuroimaging experiments that investigate how tools and their actions are represented in the brain use visual paradigms where tools or hands are displayed as 2D images and no real movements are performed. These studies discovered selective visual responses in occipitotemporal and parietal cortices for viewing pictures of hands or tools, which are assumed to reflect action processing, but this has rarely been directly investigated. Here, we examined the responses of independently visually defined category-selective brain areas when participants grasped 3D tools (N = 20; 9 females). Using real-action fMRI and multivoxel pattern analysis, we found that grasp typicality representations (i.e., whether a tool is grasped appropriately for use) were decodable from hand-selective areas in occipitotemporal and parietal cortices, but not from tool-, object-, or body-selective areas, even if partially overlapping. Importantly, these effects were exclusive for actions with tools, but not for biomechanically matched actions with control nontools. In addition, grasp typicality decoding was significantly higher in hand than tool-selective parietal regions. Notably, grasp typicality representations were automatically evoked even when there was no requirement for tool use and participants were naive to object category (tool vs nontools). Finding a specificity for typical tool grasping in hand-selective, rather than tool-selective, regions challenges the long-standing assumption that activation for viewing tool images reflects sensorimotor processing linked to tool manipulation. Instead, our results show that typicality representations for tool grasping are automatically evoked in visual regions specialized for representing the human hand, the primary tool of the brain for interacting with the world.
Assuntos
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Desempenho Psicomotor / Mapeamento Encefálico / Imageamento Tridimensional / Mãos Limite: Adolescent / Adult / Female / Humans / Male Idioma: En Revista: J Neurosci Ano de publicação: 2021 Tipo de documento: Article País de afiliação: Reino Unido

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Desempenho Psicomotor / Mapeamento Encefálico / Imageamento Tridimensional / Mãos Limite: Adolescent / Adult / Female / Humans / Male Idioma: En Revista: J Neurosci Ano de publicação: 2021 Tipo de documento: Article País de afiliação: Reino Unido