Your browser doesn't support javascript.
loading
Driving and suppressing the human language network using large language models.
Tuckute, Greta; Sathe, Aalok; Srikant, Shashank; Taliaferro, Maya; Wang, Mingye; Schrimpf, Martin; Kay, Kendrick; Fedorenko, Evelina.
Affiliation
  • Tuckute G; Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA. gretatu@mit.edu.
  • Sathe A; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA. gretatu@mit.edu.
  • Srikant S; Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA.
  • Taliaferro M; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA.
  • Wang M; Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, USA.
  • Schrimpf M; MIT-IBM Watson AI Lab, Cambridge, MA, USA.
  • Kay K; Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA, USA.
  • Fedorenko E; McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA, USA.
Nat Hum Behav ; 8(3): 544-561, 2024 Mar.
Article in En | MEDLINE | ID: mdl-38172630
ABSTRACT
Transformer models such as GPT generate human-like language and are predictive of human brain responses to language. Here, using functional-MRI-measured brain responses to 1,000 diverse sentences, we first show that a GPT-based encoding model can predict the magnitude of the brain response associated with each sentence. We then use the model to identify new sentences that are predicted to drive or suppress responses in the human language network. We show that these model-selected novel sentences indeed strongly drive and suppress the activity of human language areas in new individuals. A systematic analysis of the model-selected sentences reveals that surprisal and well-formedness of linguistic input are key determinants of response strength in the language network. These results establish the ability of neural network models to not only mimic human language but also non-invasively control neural activity in higher-level cortical areas, such as the language network.
Subject(s)

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Comprehension / Language Type of study: Prognostic_studies Limits: Humans Language: En Journal: Nat Hum Behav Year: 2024 Document type: Article Affiliation country: United States

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Comprehension / Language Type of study: Prognostic_studies Limits: Humans Language: En Journal: Nat Hum Behav Year: 2024 Document type: Article Affiliation country: United States