Skip to main content
bioRxiv
  • Home
  • About
  • Submit
  • ALERTS / RSS
Advanced Search
New Results

Driving and suppressing the human language network using large language models

View ORCID ProfileGreta Tuckute, View ORCID ProfileAalok Sathe, View ORCID ProfileShashank Srikant, View ORCID ProfileMaya Taliaferro, Mingye Wang, View ORCID ProfileMartin Schrimpf, View ORCID ProfileKendrick Kay, View ORCID ProfileEvelina Fedorenko
doi: https://doi.org/10.1101/2023.04.16.537080
Greta Tuckute
1Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Greta Tuckute
  • For correspondence: gretatu@mit.edu evelina9@mit.edu
Aalok Sathe
1Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Aalok Sathe
Shashank Srikant
3Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
4MIT-IBM Watson AI Lab, Cambridge, MA 02142, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Shashank Srikant
Maya Taliaferro
1Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Maya Taliaferro
Mingye Wang
1Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Martin Schrimpf
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
5Quest for Intelligence, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Martin Schrimpf
Kendrick Kay
6Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN 55455 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Kendrick Kay
Evelina Fedorenko
1Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
2McGovern Institute for Brain Research, Massachusetts Institute of Technology, Cambridge, MA 02139 USA
7The Program in Speech and Hearing Bioscience and Technology, Harvard University, Cambridge, MA 02138 USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Evelina Fedorenko
  • For correspondence: gretatu@mit.edu evelina9@mit.edu
  • Abstract
  • Full Text
  • Info/History
  • Metrics
  • Supplementary material
  • Preview PDF
Loading

Abstract

Transformer language models are today’s most accurate models of language processing in the brain. Here, using fMRI-measured brain responses to 1,000 diverse sentences, we develop a GPT-based encoding model and use this model to identify new sentences that are predicted to drive or suppress responses in the human language network. We demonstrate that these model-selected novel sentences indeed drive and suppress activity of human language areas in new individuals (86% increase and 98% decrease relative to the average response to diverse naturalistic sentences). A systematic analysis of the model-selected sentences reveals that surprisal and well-formedness of linguistic input are key determinants of response strength in the language network. These results establish the ability of brain-aligned models to noninvasively control neural activity in higher-level cortical areas, like the language network.

Competing Interest Statement

The authors have declared no competing interest.

Footnotes

  • Updated Figure 4 panel B and C. Added 7 new supplementary figures.

  • ↵i https://huggingface.co/gpt2-xl

Copyright 
The copyright holder for this preprint is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY-NC-ND 4.0 International license.
Back to top
PreviousNext
Posted August 01, 2023.
Download PDF

Supplementary Material

Email

Thank you for your interest in spreading the word about bioRxiv.

NOTE: Your email address is requested solely to identify you as the sender of this article.

Enter multiple addresses on separate lines or separate them with commas.
Driving and suppressing the human language network using large language models
(Your Name) has forwarded a page to you from bioRxiv
(Your Name) thought you would like to see this page from the bioRxiv website.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Share
Driving and suppressing the human language network using large language models
Greta Tuckute, Aalok Sathe, Shashank Srikant, Maya Taliaferro, Mingye Wang, Martin Schrimpf, Kendrick Kay, Evelina Fedorenko
bioRxiv 2023.04.16.537080; doi: https://doi.org/10.1101/2023.04.16.537080
Reddit logo Twitter logo Facebook logo LinkedIn logo Mendeley logo
Citation Tools
Driving and suppressing the human language network using large language models
Greta Tuckute, Aalok Sathe, Shashank Srikant, Maya Taliaferro, Mingye Wang, Martin Schrimpf, Kendrick Kay, Evelina Fedorenko
bioRxiv 2023.04.16.537080; doi: https://doi.org/10.1101/2023.04.16.537080

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Subject Area

  • Neuroscience
Subject Areas
All Articles
  • Animal Behavior and Cognition (4654)
  • Biochemistry (10300)
  • Bioengineering (7615)
  • Bioinformatics (26192)
  • Biophysics (13448)
  • Cancer Biology (10620)
  • Cell Biology (15345)
  • Clinical Trials (138)
  • Developmental Biology (8453)
  • Ecology (12755)
  • Epidemiology (2067)
  • Evolutionary Biology (16765)
  • Genetics (11356)
  • Genomics (15400)
  • Immunology (10548)
  • Microbiology (25042)
  • Molecular Biology (10152)
  • Neuroscience (54101)
  • Paleontology (398)
  • Pathology (1655)
  • Pharmacology and Toxicology (2877)
  • Physiology (4314)
  • Plant Biology (9197)
  • Scientific Communication and Education (1581)
  • Synthetic Biology (2541)
  • Systems Biology (6752)
  • Zoology (1452)