Skip to main content
bioRxiv
  • Home
  • About
  • Submit
  • ALERTS / RSS
Advanced Search
Confirmatory Results

Narratives: fMRI data for evaluating models of naturalistic language comprehension

View ORCID ProfileSamuel A. Nastase, View ORCID ProfileYun-Fei Liu, View ORCID ProfileHanna Hillman, View ORCID ProfileAsieh Zadbood, View ORCID ProfileLiat Hasenfratz, View ORCID ProfileNeggin Keshavarzian, View ORCID ProfileJanice Chen, View ORCID ProfileChristopher J. Honey, View ORCID ProfileYaara Yeshurun, View ORCID ProfileMor Regev, Mai Nguyen, View ORCID ProfileClaire H. C. Chang, View ORCID ProfileChristopher Baldassano, View ORCID ProfileOlga Lositsky, Erez Simony, Michael A. Chow, View ORCID ProfileYuan Chang Leong, View ORCID ProfilePaula P. Brooks, Emily Micciche, Gina Choe, Ariel Goldstein, View ORCID ProfileTamara Vanderwal, View ORCID ProfileYaroslav O. Halchenko, View ORCID ProfileKenneth A. Norman, View ORCID ProfileUri Hasson
doi: https://doi.org/10.1101/2020.12.23.424091
Samuel A. Nastase
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Samuel A. Nastase
Yun-Fei Liu
2Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, MD, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Yun-Fei Liu
Hanna Hillman
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Hanna Hillman
Asieh Zadbood
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Asieh Zadbood
Liat Hasenfratz
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Liat Hasenfratz
Neggin Keshavarzian
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Neggin Keshavarzian
Janice Chen
2Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, MD, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Janice Chen
Christopher J. Honey
2Department of Psychological and Brain Sciences, Johns Hopkins University, Baltimore, MD, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Christopher J. Honey
Yaara Yeshurun
3School of Psychological Sciences, Tel Aviv University, Tel Aviv, Israel
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Yaara Yeshurun
Mor Regev
4Montreal Neurological Institute, McGill University, Montreal, QC, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Mor Regev
Mai Nguyen
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Claire H. C. Chang
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Claire H. C. Chang
Christopher Baldassano
5Department of Psychology, Columbia University, New York, NY, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Christopher Baldassano
Olga Lositsky
6Department of Cognitive, Linguistic and Psychological Sciences, Brown University, Providence, RI, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Olga Lositsky
Erez Simony
7Faculty of Electrical Engineering, Holon Institute of Technology, Holon, Israel
8Department of Neurobiology, Weizmann Institute of Science, Rehovot, Israel
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Michael A. Chow
9DataCamp, Inc., New York, NY, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Yuan Chang Leong
10Helen Wills Neuroscience Institute, University of California, Berkeley, CA, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Yuan Chang Leong
Paula P. Brooks
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Paula P. Brooks
Emily Micciche
11Peabody College, Vanderbilt University, Nashville, TN, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Gina Choe
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Ariel Goldstein
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Tamara Vanderwal
12Department of Psychiatry, University of British Columbia, and BC Children’s Hospital Research Institute, Vancouver, BC, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Tamara Vanderwal
Yaroslav O. Halchenko
13Department of Psychological and Brain Sciences and Department of Computer Science, Dartmouth College, Hanover, NH, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Yaroslav O. Halchenko
Kenneth A. Norman
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Kenneth A. Norman
Uri Hasson
1Princeton Neuroscience Institute and Department of Psychology, Princeton University, Princeton, NJ, USA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Uri Hasson
  • Abstract
  • Full Text
  • Info/History
  • Metrics
  • Data/Code
  • Preview PDF
Loading

Abstract

The “Narratives” collection aggregates a variety of functional MRI datasets collected while human subjects listened to naturalistic spoken stories. The current release includes 345 subjects, 891 functional scans, and 27 diverse stories of varying duration totaling ~4.6 hours of unique stimuli (~43,000 words). This data collection is well-suited for naturalistic neuroimaging analysis, and is intended to serve as a benchmark for models of language and narrative comprehension. We provide standardized MRI data accompanied by rich metadata, preprocessed versions of the data ready for immediate use, and the spoken story stimuli with time-stamped phoneme- and word-level transcripts. All code and data are publicly available with full provenance in keeping with current best practices in transparent and reproducible neuroimaging.

Background and summary

We use language to build a shared understanding of the world. In speaking, we package certain brain states into a sequence of linguistic elements that can be transmitted verbally, thorough vibrations in the air; in listening, we expand a verbal transmission into the intended brain states, bringing our own experiences to bear on the interpretation (Hasson et al., 2012). The neural machinery supporting this language faculty presents a particular challenge for neuroscience. Certain aspects of language are uniquely human (Berwick et al., 2013), limiting the efficacy of nonhuman animal models and restricting the use of invasive methods. Language is also very dynamic and contextualized; linguistic narratives evolve over time and the meaning of each element depends in part on the history of the elements preceding it, making it difficult to fully decompose natural language using traditional experimental manipulations (Hamilton & Huth, 2018; Hasson et al., 2018; Willems et al., 2020).

Noninvasive neuroimaging tools, such as functional MRI, have laid the groundwork for a neurobiology of language comprehension (Bookheimer, 2002; Vigneau et al., 2006; Hickok & Poeppel, 2007; Price, 2010, 2012; Friederici, 2011, 2012). Functional MRI measures local fluctuations in blood oxygenation—referred to as the blood-oxygen-level-dependent (BOLD) signal—associated with changes in neural activity over the course of an experimental paradigm (Kwong et al., 1992; Ogawa et al., 1992; Logothetis et al., 2001; Logothetis, 2008). Despite relatively poor temporal resolution (e.g. one sample every 1–2 seconds), fMRI allows us to map brain-wide responses to linguistic stimuli at a spatial resolution of millimeters. A large body of prior work has revealed a partially left-lateralized network of temporal and frontal cortices encoding acoustic-phonological (Démonet et al., 1992; Zatorre et al., 1992; Belin et al., 2000; Vouloumanos et al., 2001), syntactic (Dapretto & Bookheimer, 1999; Ben-Shachar et al., 2003; Noppeney & Price, 2004), and semantic features (Patterson et al., 2007; Binder et al., 2009) of linguistic stimuli (with considerable individual variability; Fedorenko et al., 2010; Mahowald & Fedorenko, 2016; Braga et al., 2020). This work has historically been grounded in highly-controlled experimental paradigms focusing on isolated phonemes (e.g. Jäncke et al., 2002; Obleser et al., 2007), single words (e.g. Petersen et al., 1988; Wise et al., 1991; Poldrack et al., 1999), or contrasting sentence manipulations (e.g. Just et al., 1996; Kuperberg et al., 2000; Ni et al., 2000; Scott et al., 2000; Vandenberghe et al., 2002; Humphries et al., 2006; cf. Brennan et al., 2012; Brennan et al., 2016). Recent work, however, has begun extending our understanding to more ecological contexts using naturalistic text (e.g. Wehbe et al., 2014) and speech (e.g. Huth et al., 2016).

In parallel, the machine learning community has made tremendous advances in natural language processing (Goldberg, 2017; Baroni, 2020). Neurally-inspired computational models are beginning to excel at complex linguistic tasks such as word-prediction, summarization, translation, and question-answering (e.g. Devlin et al., 2018; Radford et al., 2019). Rather than using symbolic lexical representations and syntactic trees, these models typically rely on vectorial representations of linguistic content (Turney et al., 2010): linguistic elements that are similar in some respect are encoded nearer to each other in a continuous embedding space, and seemingly complex linguistic relations can be recovered using relatively simple arithmetic operations (e.g. Mikolov et al., 2013; Manning et al., 2020). In contrast to experimental traditions in linguistics and neuroscience, machine learning has embraced complex, high-dimensional models trained on enormous corpora of real-world text, emphasizing predictive power over interpretability (Breiman, 2001; Yarkoni & Westfall, 2017; Varoquaux & Poldrack, 2019).

We expect that a reconvergence of these research trajectories supported by “big” neural data will be mutually beneficial (Hasson et al., 2020; Nastase et al., 2020a). Public, well-curated benchmark datasets can both accelerate research and serve as useful didactic tools (e.g. MNIST, CIFAR; Lecun et al., 1998; Krizhevsky et al., 2009; Rocca & Yarkoni, 2020). Furthermore, there is a societal benefit to sharing human brain data: fMRI data are expensive to acquire and the reuse of publicly shared fMRI data is estimated to have saved billions in public funding (Milham et al., 2018). Public data also receive much greater scrutiny, with the potential to reveal (and rectify) “bugs” in the data or metadata. Although public datasets released by large consortia have propelled exploratory research forward (e.g. Biswal et al., 2010; Van Essen et al., 2013), relatively few of these include naturalistic stimuli (cf. movie-viewing paradigms in Cam-CAN, HBN, HCP S1200; Shafto et al., 2014; Alexander et al., 2017; Taylor et al., 2017). On the other hand, maturing standards and infrastructure (e.g. Poldrack & Gorgolewski, 2014; Poldrack et al., 2017, 2019) have enabled increasingly widespread sharing of smaller datasets from the “long tail” of human neuroimaging research (Ferguson et al., 2014). We are beginning to see a proliferation of public neuroimaging datasets acquired using rich, naturalistic experimental paradigms (e.g. Hanke et al., 2014, 2016; Aliko et al., 2020; summarized in DuPre et al., 2020).

With this in mind, we introduce the “Narratives” collection of naturalistic story-listening fMRI data for evaluating models of language (Nastase et al., 2019b). The Narratives collection comprises fMRI data collected over the course of seven years by the Hasson and Norman Labs at the Princeton Neuroscience Institute while participants listened to 27 spoken story stimuli ranging from ~3 minutes to ~56 minutes for a total of ~4.6 hours of unique stimuli (~43,000 words; Table 1). The collection currently includes 345 unique subjects contributing 891 functional scans with accompanying anatomical data and metadata. In addition to the MRI data, we provide demographic data and comprehension scores where available. Finally, we provide the auditory story stimuli, as well as time-stamped phoneme- and word-level transcripts in hopes of accelerating analysis of the linguistic content of the data. The data are standardized according to the Brain Imaging Data Structure (BIDS 1.2.1; https://bids.neuroimaging.io/; RRID:SCR_016124; Gorgolewski et al., 2016), and are publicly available via OpenNeuro (https://openneuro.org/; RRID:SCR_005031; Poldrack & Gorgolewski, 2015): https://openneuro.org/datasets/ds002345. Derivatives of the data, including preprocessed versions of the data and stimulus annotations, are available with transparent provenance via DataLad (https://www.datalad.org/; RRID:SCR_003931; Hanke et al., 2020): http://datasets.datalad.org/?dir=/labs/hasson/narratives.

View this table:
  • View inline
  • View popup
  • Download powerpoint
Table 1.

The “Narratives” datasets summarized in terms of duration, word count, and sample size. Some subjects participated in multiple experiments resulting in overlapping samples. Note that for the “Milky Way” and “Shapes” datasets, we tabulate the duration and word count only once for closely related experimental manipulations, respectively (reflected in the total durations at bottom). For the “Schema” dataset, we indicate the sum of the duration and word counts across eight brief stories. We do not include the temporally scrambled versions of the “It’s Not the Fall That Gets You Dataset” in the duration and word totals.

We believe these data have considerable potential for reuse because naturalistic spoken narratives are rich enough to support testing of a wide range of meaningful hypotheses about language and the brain (Spiers & Maguire, 2007; Hasson & Honey, 2012; Hamilton & Huth, 2018; Matusz et al., 2019; Sonkusare et al., 2019; Nastase et al., 2020a; Willems et al., 2020). These hypotheses can be formalized as quantitative models and evaluated against the brain data (Kriegeskorte et al., 2008; Naselaris et al., 2011; Fig. 1a). For example, the Narratives data are particularly well-suited for evaluating models capturing linguistic content ranging from lower-level acoustic features (Santoro et al., 2014; de Heer et al., 2017; Kell et al., 2018) to higher-level semantic features (Mitchell et al., 2008; Huth et al., 2016; Pereira et al., 2018; Goldstein et al., 2020; Schrimpf et al., 2020). More broadly, naturalistic data of this sort can be useful for evaluating shared information across subjects (Hasson et al., 2004; Nastase et al., 2019a), individual differences (Vanderwal et al., 2017; Feilong et al., 2018; Finn et al., 2018, 2020), algorithms for functional alignment (Haxby et al., 2011, 2020; Chen et al., 2015; Guntupalli et al., 2016, 2018; Van Uden et al., 2018), models of event segmentation and narrative context (Milivojevic et al., 2016; Baldassano et al., 2017, 2018; Chang et al., 2018), and functional network organization (Simony et al., 2016; Kim et al., 2018; Betzel et al., 2020; Meer et al., 2020). In the following, we describe the Narratives data collection and provide several perspectives on data quality.

Fig. 1.
  • Download figure
  • Open in new tab
Fig. 1.

Schematic depiction of the naturalistic story-listening paradigm and data provenance. (a) At bottom, the full auditory story stimulus “Pie Man” by Jim O’Grady is plotted as a waveform of varying amplitude (y-axis) over the duration of 450 seconds (x-axis) corresponding to 300 fMRI volumes sampled at a TR of 1.5 seconds. An example clip (marked by vertical orange lines) is expanded and accompanied by the time-stamped word annotation (“I began my illustrious career in journalism in the Bronx, where I worked as a hard-boiled reporter for the Ram…”). The story stimulus can be described according to a variety of models; for example, acoustic, semantic, or narrative features can be extracted from or assigned to the stimulus. In a prediction or model-comparison framework, these models serve as formal hypotheses linking the stimulus to brain responses. At top, preprocessed fMRI response time-series from three example voxels for an example subject are plotted for the full duration of the story stimulus (x-axis: fMRI signal magnitude; y-axis: scan duration in TRs; red: early auditory cortex; orange: auditory association cortex; purple: temporoparietal junction). (b) At bottom, MRI data, metadata, and stimuli are formatted according to the BIDS standard and publicly available via OpenNeuro. All derivative data are version-controlled and publicly available via DataLad. The schematic preprocessing workflow includes the following steps: realignment, susceptibility distortion correction, and spatial normalization with fMRIPrep; unsmoothed and spatially smoothed workflows proceed in parallel; confound regression to mitigate artifacts from head motion and physiological noise; as well as intersubject correlation (ISC) analyses used for quality control in this manuscript. Each stage of the processing workflow is publicly available and indexed by a commit hash (left) providing a full, interactive history of the data collection.

Methods

Participants

Data were collected over the course of roughly seven years, from October, 2011 to September, 2018. Participants were recruited from the Princeton University student body as well as non-university-affiliated members of the broader community in Princeton, NJ. All participants provided informed, written consent prior to data collection in accordance with experimental procedures approved by Princeton University Institutional Review Board. Across all datasets, 345 adults participated in data collection (ages 18–53 years, mean age 22.2 ± 5.1 years, 204 reported female). Demographics for each dataset are reported in the “Narrative datasets” section. Both native and non-native English speakers are included in the dataset, but all subjects reported fluency in English. All subjects reported having normal hearing and no history of neurological disorders.

Experimental procedures

Upon arriving at the fMRI facility, participants first completed a simple demographics questionnaire, as well as a comprehensive MRI screening form. Participants were instructed to listen and pay attention to the story stimulus, remain still, and keep their eyes open. In some cases, subject wakefulness was monitored in real-time using an eye-tracker. Stimulus presentation was implemented using either Psychtoolbox (Brainard, 1997; Kleiner et al., 2007) or PsychoPy (Peirce, 2007, 2009; Peirce et al., 2019). In some cases a centrally-located fixation cross or dot was presented throughout the story stimulus; however, participants were not instructed to maintain fixation. Auditory story stimuli were delivered via MRI-compatible insert earphones (Sensimetrics, Model S14); headphones or foam padding were placed over the earphones to reduce scanner noise. In most cases, a volume check was performed to ensure that subjects could comfortably hear the auditory stimulus over the MRI acquisition noise prior to data collection.

For many datasets in this collection, story scans were accompanied by additional functional scans including different stories or other experimental paradigms, as well auxiliary anatomical scans. In all cases, an entire story stimulus was presented in a single scanning run; however, in some cases multiple independent stories were collected in a single scanning run (the “Slumlord” and “Reach for the Stars One Small Step at a Time” stories, and the “Schema” stories; see the “Narrative datasets” section below). In scanning sessions with multiple stories and scanning runs, participants were free to request volume adjustments between scans (although this was uncommon). Participants were debriefed as to the purpose of a given experiment at the end of a scanning session, and in some cases filled out follow-up questionnaires evaluating their comprehension of the stimuli. Specifical procedural details for each dataset are described in the “Narrative datasets” section below.

Stimuli

To add to the scientific value of the neuroimaging data, we provide waveform (WAV) audio files containing the spoken story stimulus for each dataset (e.g. pieman_audio.wav). Audio files were edited using the open source Audacity software (https://www.audacityteam.org). These audio files served as stimuli for the publicly-funded psychological and neuroscientific studies described herein. The stimuli are intended for non-profit, non-commercial scholarly use—principally feature extraction—under “fair use” or “fair dealing” provisions; (re)sharing and (re)use of these media files is intended to respect these guidelines (DuPre et al., 2020). Story stimuli span a variety of media, including commercially-produced radio and internet broadcasts, authors and actors reading written works, professional storytellers performing in front of live audiences, and subjects verbally recalling previous events (in the scanner). Manually written transcripts are provided with each story stimulus. The specific story stimuli for each dataset are described in more detail in the “Narrative datasets” section below. In total, the auditory stories sum to roughly 4.6 hours of unique stimuli corresponding to 11,149 TRs (excluding TRs acquired with no auditory story stimulus). Concatenating across all subjects, this amounts to roughly 6.4 days worth of story-listening fMRI data, or 369,496 TRs.

By including all stimuli in this data release, researchers can extract their choice of linguistic features for model evaluation (McNamara et al., 2017; Willems et al., 2020), opening the door to many novel scientific questions. To kick-start this process, we used Gentle 0.10.1 (Ochshorn & Hawkins, 2016) to create time-stamped phoneme- and word-level transcripts for each story stimulus in an automated fashion. Gentle is a robust, lenient forced-alignment algorithm that relies on the free and open source Kaldi automated speech recognition software (Povey et al., 2011) and the Fisher English Corpus (Cieri et al., 2004).

The written transcript and audio file were supplied to Gentle, which generates two principal outputs. First, a simple word-level transcript with onset and offset timing for each word is saved as a tabular CSV file. Each row in the CSV file corresponds to a word. The CSV file contains four columns (no header) where the first column indicates the word from the written transcript, the second column indicates the corresponding word in the vocabulary, the third column indicates the word onset timing from the beginning of the stimulus file, and the fourth column indicates the word offset timing. Second, Gentle generates a more detailed JSON file containing the full written transcript followed by a list of word entries where each entry contains fields indicating the word (“word”) and whether word was found in the vocabulary (“alignedWord”), the onset (“start”) and offset (“end”) timing of the word referenced to the beginning of the audio file, the duration of each phone comprising that word (“phones” containing “phone” and “duration”), and a field (“case”) indicating whether the word was correctly localized in the audio signal. Words that were not found in the vocabulary are indicated by “<unk>” for “unknown” in the second column of the CSV file or in the “alignedWord” field of the JSON file. Words that are not successfully localized in the audio file receive their own row but contain no timing information in the CSV file, and are marked as “not-found-in-audio” in the “case” field of the JSON file. All timing information generated by Gentle is indexed to the beginning of the stimulus audio file. This should be referenced against the BIDS-formatted events.tsv files accompanying each scan describing when the story stimulus began relative to the beginning of the scan (this varies across datasets as described in the “Narratives datasets” section below). Some word counts may be slightly inflated by the inclusion of disfluencies (e.g. “uh”) in the transcripts. Transcripts for the story stimuli contain 42,989 words in total across all stories; 789 words (1.8%) were not successfully localized by the forced-alignment algorithm and 651 words (1.5%) were not found in the vocabulary. Concatenating across all subjects yields 1,399,655 words occurring over the course of 369,496 TRs.

Gentle packages the audio file, written transcript, CSV and JSON files with an HTML file for viewing in a browser that allows for interactively listening to the audio file while each word is highlighted (with its corresponding phones) when it occurs in the audio. Note that our forced-alignment procedure using Gentle was fully automated and did not include any manual intervention. That said, the time-stamped transcripts are not perfect—speech disfluencies, multiple voices speaking simultaneously, sound effects, and scanner noise provide a challenge for automated transcription. We hope this annotation can serve as a starting point, and expect that better transcriptions will be created in the future. We invite researchers who derive additional features and annotations from the stimuli to contact the corresponding author and we will help incorporate these annotations into the publicly available dataset.

MRI data acquisition

All datasets were collected at the Princeton Neuroscience Institute Scully Center for Neuroimaging. Acquisition parameters are reproduced below in compliance with the guidelines put forward by the Organization for Human Brain Mapping (OHBM) Committee on Best Practices in Data Analysis and Sharing (COBIDAS; Nichols et al., 2017), and are also included in the BIDS-formatted JSON metadata files accompanying each scan. All studies used a repetition time (TR) of 1.5 seconds. The “Pie Man”, “Tunnel Under the World”, “Lucy”, “Pretty Mouth and Green My Eyes”, “Milky Way”, “Slumlord”, “Reach for the Stars One Small Step at a Time”, “It’s Not the Fall that Gets You”, “Merlin”, “Sherlock”, and “The 21st Year” datasets were collected on a 3T Siemens Magnetom Skyra (Erlangen, Germany) with a 20-channel phased-array head coil using the following acquisition parameters. Functional BOLD images were acquired in an interleaved fashion using gradient-echo echo-planar imaging (EPI) with an in-plane acceleration factor of 2 using GRAPPA: TR/TE = 1500/28 ms, flip angle = 64°, bandwidth = 1445 Hz/Px, in-plane resolution = 3 × 3 mm, slice thickness = 4 mm, matrix size = 64 × 64, FoV = 192 × 192 mm, 27 axial slices with roughly full brain coverage and no gap, anterior–posterior phase encoding, prescan normalization, fat suppression. At the beginning of each run, three dummy scans were acquired and discarded by the scanner to allow for signal stabilization. T1-weighted structural images were acquired using a high-resolution single-shot MPRAGE sequence with an in-plane acceleration factor of 2 using GRAPPA: TR/TE/TI = 2300/3.08/900 ms, flip angle = 9°, bandwidth = 240 Hz/Px, in-plane resolution 0.859 × 0.859 mm, slice thickness 0.9 mm, matrix size = 256 × 256, FoV = 172.8 × 220 × 220 mm, 192 sagittal slices, ascending acquisition, anterior–posterior phase encoding, prescan normalization, no fat suppression, 7 minutes 21 seconds total acquisition time.

The “Schema” and “Shapes” datasets were collected on a 3T Siemens Magnetom Prisma with a 64-channel head coil using the following acquisition parameters. Functional images were acquired in an interleaved fashion using gradient-echo EPI with a multiband (simultaneous multi-slice; SMS) acceleration factor of 4 using blipped CAIPIRINHA and no in-plane acceleration: TR/TE 1500/39 ms, flip angle = 50°, bandwidth = 1240 Hz/Px, in-plane resolution = 2.0 × 2.0 mm, slice thickness 2.0 mm, matrix size = 96 × 96, FoV = 192 × 192 mm, 60 axial slices with full brain coverage and no gap, anterior–posterior phase encoding, 6/8 partial Fourier, no prescan normalization, fat suppression, three dummy scans. T1-weighted structural images were acquired using a high-resolution single-shot MPRAGE sequence with an in-plane acceleration factor of 2 using GRAPPA: TR/TE/TI = 2530/2.67/1200 ms, flip angle = 7°, bandwidth = 200 Hz/Px, in-plane resolution 1.0 × 1.0 mm, slice thickness 1.0 mm, matrix size = 256 × 256, FoV = 176 × 256 × 256 mm, 176 sagittal slices, ascending acquisition, no fat suppression, 5 minutes 52 seconds total acquisition time.

The “Pie Man (PNI),” “Running from the Bronx,” “I Knew You Were Black,” and “The Man Who Forgot Ray Bradbury” datasets were collected on the same 3T Siemens Magnetom Prisma with a 64-channel head coil using different acquisition parameters. Functional images were acquired in an interleaved fashion using gradient-echo EPI with a multiband acceleration factor of 3 using blipped CAIPIRINHA and no in-plane acceleration: TR/TE 1500/31 ms, flip angle = 67°, bandwidth = 2480 Hz/Px, in-plane resolution = 2.5 × 2.5 mm, slice thickness 2.5 mm, matrix size = 96 × 96, FoV = 240 × 240 mm, 48 axial slices with full brain coverage and no gap, anterior–posterior phase encoding, prescan normalization, fat suppression, three dummy scans. T1-weighted structural images were acquired using a high-resolution single-shot MPRAGE sequence with an in-plane acceleration factor of 2 using GRAPPA: TR/TE/TI = 2530/3.3/1100 ms, flip angle = 7°, bandwidth = 200 Hz/Px, in-plane resolution 1.0 × 1.0 mm, slice thickness 1.0 mm, matrix size = 256 × 256, FoV = 176 × 256 × 256 mm, 176 sagittal slices, ascending acquisition, no fat suppression, prescan normalization, 5 minutes 53 seconds total acquisition time. T2-weighted structural images were acquired using a high-resolution single-shot MPRAGE sequence with an in-plane acceleration factor of 2 using GRAPPA: TR/TE = 3200/428 ms, flip angle = 120°, bandwidth = 200 Hz/Px, in-plane resolution 1.0 × 1.0 mm, slice thickness 1.0 mm, matrix size = 256 × 256, FoV = 176 × 256 × 256 mm, 176 sagittal slices, interleaved acquisition, no prescan normalization, no fat suppression, 4 minutes 40 seconds total acquisition time.

MRI preprocessing

MRI data were preprocessed using fMRIPrep (20.0.5; RRID:SCR_016216; Esteban et al., 2019b; Esteban et al., 2020b). FMRIPrep is a containerized, automated tool based on Nipype (1.4.2; RRID:SCR_002502; Gorgolewski et al., 2011; Esteban et al., 2020c) that adaptively adjusts to idiosyncrasies of the dataset (as captured by the metadata) to apply the best-in-breed preprocessing workflow. Many internal operations of fMRIPrep functional processing workflow use Nilearn (0.6.2; RRID:SCR_001362; Abraham et al., 2014). For more details of the pipeline, see the section corresponding to workflows in fMRIPrep’s documentation. The containerized fMRIPrep software was deployed using Singularity (3.5.2-1.1.sdl7; Kurtzer et al., 2017). The fMRIPrep outputs and visualization can be found in the fmriprep/ directory in derivatives/ available via the DataLad release. The fMRIPrep workflow produces two principal outputs: (a) the functional time series data in one more output space (e.g. MNI space), and (b) a collection of confound variables for each functional scan. In the following, we describe fMRIPrep’s anatomical and functional workflows, as well as subsequent spatial smoothing and confound regression implemented in AFNI (RRID:SCR_005927; Cox, 1996; Cox, 2012).

The anatomical MRI T1-weighted (T1w) images were corrected for intensity non-uniformity with N4BiasFieldCorrection (Tustison et al., 2010), distributed with ANTs (2.2.0; RRID:SCR_004757; Avants et al., 2008), and used as T1w-reference throughout the workflow. The T1w-reference was then skull-stripped with a Nipype implementation of the antsBrainExtraction.sh (from ANTs) using the OASIS30ANTs as the target template. Brain tissue segmentation of cerebrospinal fluid (CSF), white-matter (WM), and gray-matter (GM) was performed on the brain-extracted T1w using fast (FSL 5.0.9; RRID:SCR_002823; Zhang et al., 2001). Brain surfaces were reconstructed using recon-all (FreeSurfer 6.0.1; RRID:SCR_001847; Dale et al., 1999; Fischl, 2012), and the brain mask estimated previously was refined with a custom variation of the method to reconcile ANTs-derived and FreeSurfer-derived segmentations of the cortical gray-matter from Mindboggle (RRID:SCR_002438; Klein et al., 2017). Volume-based spatial normalization to two commonly-used standard spaces (MNI152NLin2009cAsym, MNI152NLin6Asym) was performed through nonlinear registration with antsRegistration (ANTs 2.2.0) using brain-extracted versions of both T1w reference and the T1w template. The following two volumetric templates were selected for spatial normalization and deployed using TemplateFlow (Esteban et al., 2020a): (a) ICBM 152 Nonlinear Asymmetrical template version 2009c (RRID:SCR_008796; TemplateFlow ID: MNI152NLin2009cAsym; Fonov et al., 2009), and (b) FSL’s MNI ICBM 152 non-linear 6th Generation Asymmetric Average Brain Stereotaxic Registration Model (RRID:SCR_002823; TemplateFlow ID: MNI152NLin6Asym; Evans et al., 2012). Surface-based normalization based on nonlinear registration of sulcal curvature was applied using the following three surface templates (FreeSurfer reconstruction nomenclature; Fischl et al., 1999): fsaverage, fsaverage6, fsaverage5.

The functional MRI data were preprocessed in the following way. First, a reference volume and its skull-stripped version were generated using a custom methodology of fMRIPrep. A deformation field to correct for susceptibility distortions was estimated using fMRIPrep’s fieldmap-less approach. The deformation field results from co-registering the BOLD reference to the same-subject T1w-reference with its intensity inverted (Huntenburg, 2014; Wang et al., 2017). Registration was performed with antsRegistration (ANTs 2.2.0), and the process was regularized by constraining deformation to be nonzero only along the phase-encoding direction, and modulated with an average fieldmap template (Treiber et al., 2016). Based on the estimated susceptibility distortion, a corrected EPI reference was calculated for more accurate co-registration with the anatomical reference. The BOLD reference was then co-registered to the T1w reference using bbregister (FreeSurfer 6.0.1), which implements boundary-based registration (Greve & Fischl, 2009). Co-registration was configured with six degrees of freedom. Head-motion parameters with respect to the BOLD reference (transformation matrices, and six corresponding rotation and translation parameters) are estimated before any spatiotemporal filtering using mcflirt (FSL 5.0.9; Jenkinson et al., 2002, 2012; Smith et al., 2004). BOLD runs were slice-time corrected using 3dTshift from AFNI (20160207; Cox & Hyde, 1997). The BOLD time-series were resampled onto the following surfaces: fsaverage, fsaverage6, fsaverage5. The BOLD time-series (including slice-timing correction when applied) were resampled onto their original, native space by applying a single, composite transform to correct for head-motion and susceptibility distortions. These resampled BOLD time-series are referred to as preprocessed BOLD in original space, or just preprocessed BOLD. The BOLD time-series were resampled into two volumetric standard spaces, correspondingly generating the following spatially-normalized, preprocessed BOLD runs: MNI152NLin2009cAsym, MNI152NLin6Asym. A reference volume and its skull-stripped version were first generated using a custom methodology of fMRIPrep. All resamplings were performed with a single interpolation step by composing all the pertinent transformations (i.e. head-motion transform matrices, susceptibility distortion correction, and co-registrations to anatomical and output spaces). Gridded (volumetric) resamplings were performed using antsApplyTransforms (ANTs 2.2.0), configured with Lanczos interpolation to minimize the smoothing effects of other kernels (Lanczos, 1964). Non-gridded (surface) resamplings were performed using mri_vol2surf (FreeSurfer 6.0.1).

Several confounding time-series were calculated based on the preprocessed BOLD: framewise displacement (FD), DVARS, and three region-wise global signals. FD and DVARS are calculated for each functional run, both using their implementations in Nipype (following the definitions by Power et al., 2014). The three global signals are extracted within the CSF, the WM, and the whole-brain masks. Additionally, a set of physiological regressors were extracted to allow for component-based noise correction (CompCor; Behzadi et al., 2007). Principal components are estimated after high-pass filtering the preprocessed BOLD time-series (using a discrete cosine filter with 128s cut-off) for the two CompCor variants: temporal (tCompCor) and anatomical (aCompCor). The tCompCor components are then calculated from the top 5% variable voxels within a mask covering the subcortical regions. This subcortical mask is obtained by heavily eroding the brain mask, which ensures it does not include cortical GM regions. For aCompCor, components are calculated within the intersection of the aforementioned mask and the union of CSF and WM masks calculated in T1w space, after their projection to the native space of each functional run (using the inverse BOLD-to-T1w transformation). Components are also calculated separately within the WM and CSF masks. For each CompCor decomposition, the k components with the largest singular values are retained, such that the retained components’ time series are sufficient to explain 50 percent of variance across the nuisance mask (CSF, WM, combined, or temporal). The remaining components are dropped from consideration. The head-motion estimates calculated in the correction step were also placed within the corresponding confounds file. The confound time series derived from head motion estimates and global signals were expanded with the inclusion of temporal derivatives and quadratic terms for each (Satterthwaite et al., 2013). Frames that exceeded a threshold of 0.5 mm FD or 1.5 standardised DVARS were annotated as motion outliers. All of these confound variables are provided with the dataset for researchers to use as they see fit. HTML files with quality control visualizations output by fMRIPrep are available via DataLad.

We provide spatially smoothed and non-smoothed versions of the preprocessed functional data returned by fMRIPrep. Analyses requiring voxelwise correspondence across subjects, such as ISC analysis (Pajula & Tohka, 2014), can benefit from spatial smoothing due to variability in functional–anatomical correspondence across individuals—at the expense of spatial specificity (finer-grained intersubject functional correspondence can be achieved using hyperalignment rather than spatial smoothing; Haxby et al., 2011, 2020; Nastase et al., 2020b). The smoothed and non-smoothed outputs (and subsequent analyses) can be found in the afni-smooth/ and afni-nosmooth/ directories in derivatives/ available via DataLad. To smooth the volumetric functional images, we used 3dBlurToFWHM in AFNI (19.3.0; Cox, 1996; Cox, 2012) which iteratively measures the global smoothness (ratio of variance of first differences across voxels to global variance) and local smoothness (estimated within 4 mm spheres), then applies local smoothing to less-smooth areas until the desired global smoothness is achieved. All smoothing operations were performed within a brain mask to ensure that non-brain values were not smoothed into the functional data. For surface-based functional data, we applied smoothing SurfSmooth in AFNI, which uses a similar iterative algorithm for smoothing surface data according to geodesic distance on the cortical mantle (Chung et al., 2005; Hagler et al., 2006). In both cases, data were smoothed until a target global smoothness of 6 mm FWHM was met (i.e. 2–3 times original voxel sizes; Triantafyllou et al., 2006). Equalizing smoothness is critical to harmonize data acquired across different scanners and protocols (Friedman et al., 2006).

We next temporally filtered the functional data to mitigate the effects of confounding variables. Unlike traditional task fMRI experiments with a well-defined event structure, the goal of regression was not to estimate regression coefficients for any given experimental conditions; rather, similar to resting-state functional connectivity analysis, the goal of regression was to model nuisance variables, resulting in a “clean” residual time series. However, unlike conventional resting-state paradigms, naturalistic stimuli enable intersubject analyses, which are less sensitive to idiosyncratic noises than within-subject functional connectivity analysis typically used with resting-state data (Simony et al., 2016; Simony & Chang, 2019). With this in mind, we used a modest confound regression model informed by the rich literature on confound regression for resting-state functional connectivity (e.g. Ciric et al., 2017; Parkes et al., 2018). AFNI’s 3dTproject was used to regress out the following nuisance variables: six head motion parameters (three translation, three rotation), the first five principal component time series from an eroded CSF and a white matter mask (Behzadi et al., 2007; Muschelli et al., 2014), cosine bases for high-pass filtering (using a discrete cosine filter with cutoff: 128 s, or ~.0078 Hz), and first- and second-order detrending polynomials. These variables were included in a single regression model to avoid reintroducing artifacts by sequential filtering (Lindquist et al., 2019). The scripts used to perform this regression and the residual time series are provided with this data release. This processing workflow ultimately yields smoothed and non-smoothed versions of the “clean” functional time series data in several volumetric and surface-based standard spaces.

Computing environment

In addition to software mentioned elsewhere in this manuscript, all data processing relied heavily on the free, open source GNU/Linux ecosystem and NeuroDebian distribution (https://neuro.debian.net/; RRID:SCR_004401; Hanke & Halchenko, 2011; Halchenko & Hanke, 2012), as well as the Anaconda distribution (https://www.anaconda.com/) and conda package manager (https://docs.conda.io/en/latest/; RRID:SCR_018317). Many analyses relied on scientific computing software in Python (https://www.python.org/; RRID:SCR_008394), including NumPy (http://www.numpy.org/; RRID:SCR_008633; van der Walt et al., 2011; Harris et al., 2020), SciPy (https://www.scipy.org/; RRID:SCR_008058; Jones et al., 2001; Virtanen et al., 2020), Pandas (https://pandas.pydata.org/; RRID:SCR_018214; McKinney, 2010), NiBabel (https://nipy.org/nibabel/; RRID:SCR_002498; Brett et al., 2020), IPython (http://ipython.org/; RRID:SCR_001658; Perez & Granger, 2007), and Jupyter (https://jupyter.org; RRID:SCR_018315; Kluyver et al., 2016), as well as Singularity containerization (https://sylabs.io/docs/; Kurtzer et al., 2017) and the Slurm workload manager (https://slurm.schedmd.com/; Yoo et al., 2003). All surface-based MRI data were visualized using SUMA (RRID:SCR_005927; Saad et al., 2004; Saad & Reynolds, 2012). All other figures were created using Matplotlib (http://matplotlib.org/; RRID:SCR_008624; Hunter, 2007), seaborn (http://seaborn.pydata.org/; RRID:SCR_018132), GIMP (http://www.gimp.org/; RRID:SCR_003182), and Inkscape (https://inkscape.org/; RRID:SCR_014479).

Narrative datasets

Each dataset in the “Narratives” collection is described below. The datasets are listed in roughly chronological order of acquisition. For each dataset, we provide the dataset-specific subject demographics, a summary of the stimulus and timing, as well as details of the experimental procedure or design.

“Pie Man”

The “Pie Man” dataset was collected between October, 2011 and March, 2013, and comprised 82 participants (ages 18–45 years, mean age 22.5 ± 4.3 years, 45 reported female). The “Pie Man” story was told by Jim O’Grady and recorded live at the The Moth, a non-profit storytelling event, in New York City in 2008 (freely available at https://themoth.org/stories/pie-man). The “Pie Man” audio stimulus was 450 seconds (7.5 minutes) long and began with 13 seconds of neutral introductory music followed by 2 seconds of silence, such that the story itself started at 0:15 and ended at 7:17, for a duration 422 s, with 13 seconds of silence at the end of the scan. The stimulus was started simultaneously with the acquisition of the first functional MRI volume, and the scans comprise 300 TRs, matching the duration of the stimulus. The transcript for the “Pie Man” story stimulus contains 957 words; 3 words (0.3%) were not successfully localized by the forced-alignment algorithm, and 19 words (2.0%) were not found in the vocabulary (e.g. proper names). The “Pie Man” data were in some cases collected in conjunction with temporally scrambled versions of the “Pie Man” stimulus, as well the “Tunnel Under the World” and “Lucy” datasets among others, and share subjects with these datasets (as specified in the participants.tsv file). The current release only includes data collected for the intact story stimulus (rather than the temporally scrambled versions of the stimulus). The following subjects received the “Pie Man” stimulus on two separate occasions (specified by run-1 or run-2 in the BIDS file naming convention): sub-001, sub-002, sub-003, sub-004, sub-005, sub-006, sub-008, sub-010, sub-011, sub-012, sub-013, sub-014, sub-015, sub-016. We recommend excluding subjects sub-001 (both run-1 and run-2), sub-013 (run-2), sub-014 (run-2), sub-021, sub-022, sub-038, sub-056, sub-068, and sub-069 (as specified in the scan_exclude.json file in the code/ directory) based on criteria explained in the “Intersubject correlation” section of “Data validation.” Subsets of the “Pie Man” data have been previously reported in numerous publications (Lerner et al., 2011, 2014; Ben-Yakov et al., 2012; Regev et al., 2013; Stephens et al., 2013; Chen et al., 2015; Simony et al., 2016; Baldassano et al., 2017; Liu et al., 2017; Vodrahalli et al., 2017; Yeshurun et al., 2017a; Zuo et al., 2020), and additional datasets not shared here have been collected using the “Pie Man” auditory story stimulus (e.g. Gross et al., 2013; Blank & Fedorenko, 2017; Iotzov et al., 2017). In the filename convention (and figures), “Pie Man” is labeled using the task alias pieman.

“Tunnel Under the World”

The “Tunnel Under the World” dataset was collected between May, 2012, and February, 2013, and comprised 23 participants (ages 18–31 years, mean age 22.5 ± 3.8 years, 14 reported female). The “Tunnel Under the World” science-fiction story was authored by Frederik Pohl in 1955 which was broadcast in 1956 as part of the X Minus One series, a collaboration between the National Broadcasting Company and Galaxy Science Fiction magazine (freely available at https://www.oldtimeradiodownloads.com). The “Tunnel Under the World” audio stimulus is 1534 seconds (~25.5 minutes) long. The stimulus was started after the first two functional MRI volumes (2 TRs, 3 seconds) were collected, with ~23 seconds of silence after the stimulus. The functional scans comprise 1040 TRs (1560 seconds), except for subjects sub-004 and sub-013 which have 1035 and 1045 TRs respectively. The transcript for “Tunnel Under the World” contains 3,435 words; 126 words (3.7%) were not successfully localized by the forced-alignment algorithm and 88 words (2.6%) were not found in the vocabulary. The “Tunnel Under the World” and “Lucy” datasets contain largely overlapping samples of subjects, though were collected in separate sessions, and were, in many cases, collected alongside “Pie Man” scans (as specified in the participants.tsv file). We recommend excluding the “Tunnel Under the World” scans for subjects sub-004 and sub-013 (as specified in the scan_exclude.json file in the code/ directory). The “Tunnel Under the World” data have been previously reported by Lositsky and colleagues (2016). In the filename convention, “Tunnel Under the World” is labeled using the task alias tunnel.

“Lucy”

The “Lucy” dataset was collected between October, 2012 and January, 2013 and comprised 16 participants (ages 18–31 years, mean age 22.6 ± 3.9 years, 10 reported female). The “Lucy” story was broadcast by the non-profit WNYC public radio in 2010 (freely available at https://www.wnycstudios.org/story/91705-lucy). The “Lucy” audio stimulus is 542 seconds (~9 minutes) long and was started after the first two functional MRI volumes (2 TRs, 3 seconds) were collected. The functional scans comprise 370 TRs (555 seconds). The transcript for “Lucy” contains 1,607 words; 22 words (1.4%) were not successfully localized by the forced-alignment algorithm and 27 words (1.7%) were not found in the vocabulary. The “Lucy” and “Tunnel Under the World” datasets contain largely overlapping samples of subjects, and were acquired contemporaneously with “Pie Man” data. We recommend excluding the “Lucy” scans for subjects sub-053 and sub-065 (as specified in the scan_exclude.json file in the code/ directory). The “Lucy” data have not previously been reported. In the filename convention, “Lucy” is labeled using the task alias lucy.

“Pretty Mouth and Green My Eyes”

The “Pretty Mouth and Green My Eyes” dataset was collected between March, 2013, and October, 2013, and comprised 40 participants (ages 18–34 years, mean age 21.4 ± 3.5 years, 19 reported female). The “Pretty Mouth and Green My Eyes” story was authored by J. D. Salinger for The New Yorker magazine (1951) and subsequently published in the Nine Stories collection (1953). The spoken story stimulus used for data collection was based on an adapted version of the original text that was shorter and included a few additional sentences, and was read by a professional actor. The “Pretty Mouth and Green My Eyes” audio stimulus is 712 seconds (~11.9 minutes) long and began with 18 seconds of neutral introductory music followed by 3 seconds of silence, such that the story started at 0:21 (after 14 TRs) and ended at 11:37, for a duration of 676 seconds (~451 TRs), with 15 seconds (10 TRs) of silence at the end of the scan. The functional scans comprised 475 TRs (712.5 seconds). The transcript for “Pretty Mouth and Green My Eyes” contains 1,970 words; 7 words (0.4%) were not successfully localized by the forced-alignment algorithm and 38 words (1.9%) were not found in the vocabulary.

The “Pretty Mouth and Green My Eyes” stimulus was presented to two groups of subjects in two different narrative contexts according between-subject experimental design: (a) in the “affair” group, subjects read a short passage implying that the main character was having an affair; (b) in the “paranoia” group, subjects read a short passage implying that the main character’s friend was unjustly paranoid (see Yeshurun et al., 2017b, for the full prompts). The two experimental groups were randomly assigned such that there were 20 subjects in each group and each subject only received the stimulus under a single contextual manipulation. The group assignments are indicated in the participants.tsv file, and the scans.tsv file for each subject. Both groups received the identical story stimulus despite receiving differing contextual prompts. Immediately following the scans, subjects completed a questionnaire assessing comprehension of the story. The questionnaire comprised 27 context-independent and 12 context-dependent questions (39 questions in total). The resulting comprehension scores are reported as the proportion of correct answers (ranging 0–1) in the participants.tsv file and scans.tsv file for each subject. We recommend excluding the “Pretty Mouth and Green My Eyes” scans for subjects sub-038 and sub-105 (as specified in the scan_exclude.json file in the code/ directory). The “Pretty Mouth and Green My Eyes” data have been reported in existing publications (Chen et al., 2015; Yeshurun, et al., 2017b). In the filename convention, “Pretty Mouth and Green My Eyes” is labeled using the task alias prettymouth.

“Milky Way”

The “Milky Way” dataset was collected between March, 2013, and April, 2017, and comprised 53 participants (ages 18–34 years, mean age 21.7 ± 4.1 years, 27 reported female). The “Milky Way” story stimuli were written by an affiliate of the experimenters and read by a member of the Princeton Neuroscience Institute not affiliated with the experimenters’ lab. There are three versions of the “Milky Way” story capturing three different experimental conditions: (a) the original story (labeled original), which describes a man who visits a hypnotist to overcome his obsession with an ex-girlfriend, but instead becomes fixated on Milky Way candy bars; (b) an alternative version (labeled vodka) where sparse word substitutions yield a narrative in which a woman obsessed with an American Idol judge visits a psychic and becomes fixated on vodka; (c) a control version (labeled synonyms) with a similar number of word substitutions to the vodka version, but instead using synonyms of words in the original version, yielding a very similar narrative to the original. Relative to the original version, both the vodka and synonyms versions substituted on average 2.5 ± 1.7 words per sentence (34 ± 21% of words per sentence). All three versions were read by the same actor and each sentence of the recording was temporally aligned across versions. All three versions of the “Milky Way” story were 438 seconds (7.3 min; 292 TRs) long and began with 18 seconds of neutral introductory music followed by a 3 seconds of silence (21 TRs total), such that the story started at 0:21 and ended at 7:05, for a duration 404 s, with 13 seconds of silence at the end of the scan. The functional runs comprised 297 volumes (444.5 seconds). The transcript for the original version of the stimulus contains 1,059 words; 7 words (0.7%) were not successfully localized by the forced-alignment algorithm and 16 words (1.5%) were not found in the vocabulary. The vodka version contains 1058 words; 2 words (0.2%) were not successfully localized and 21 words (2.0%) were not found in the vocabulary. The synonyms version contains 1,066 words; 10 words (0.9%) were not successfully localized and 13 words (1.2%) were not found in the vocabulary.

The three versions of the stimuli were assigned to subjects according to a between-subject design, such that there were 18 subjects in each of the three groups, and each subject received only one version of the story. The group assignments are indicated in the participants.tsv file, and the scans.tsv file for each subject. The stimulus filename includes the version (milkywayoriginal, milkywayvodka, milkywaysynonyms) and is specified in the events.tsv file accompanying each run. The data corresponding to the original and vodka versions were collected between March and October, 2013, while the synonyms data were collected in March and April, 2017. Subjects completed a 28-item questionnaire assessing story comprehension following the scanning session. The comprehension scores for the original and vodka groups are reported as the proportion of correct answers (ranging 0–1) in the participants.tsv file and scans.tsv file for each subject. We recommend excluding “Milky Way” scans for subjects sub-038, sub-105, and sub-123 (as specified in the scan_exclude.json file in the code/ directory). The “Milky Way” data have been previously reported by Yeshurun and colleagues (2017a). In the filename convention, “Milky Way” is labeled using the task alias milkyway.

“Slumlord” and “Reach for the Stars One Small Step at a Time”

The “Slumlord” and “Reach for the Stars One Small Step at a Time” dataset was collected between May, 2013, and October, 2013, and comprised 18 participants (ages 18–27 years, mean age 21.0 ± 2.3 years, 8 reported female). The “Slumlord” story was told by Jack Hitt and recorded live at The Moth, a non-profit storytelling event, in New York City in 2006 (freely available at https://themoth.org/stories/slumlord). The “Reach for the Stars One Small Step at a Time” story was told by Richard Garriott and also recorded live at The Moth in New York City in 2010 (freely available at https://themoth.org/stories/reach-for-the-stars). The combined audio stimulus is 1,802 seconds (~30 minutes) long in total and begins with 22.5 seconds of music followed by 3 seconds of silence. The “Slumlord” story begins at approximately 0:25 relative to the beginning of the stimulus file, ends at 15:28, for a duration of 903 seconds (602 TRs), and is followed by 12 seconds of silence. After another 22 seconds of music starting at 15:40, the “Reach for the Stars One Small Step at a Time” story starts at 16:06 (965 seconds; relative to the beginning of the stimulus file), ends at 29:50, for a duration of 825 seconds (~550 TRs), and is followed by 12 seconds of silence. The stimulus file was started after 3 TRs (4.5 seconds) as indicated in the events.tsv files accompanying each scan. The scans were acquired with a variable number of trailing volumes following the stimulus across subjects, but can be truncated as researchers see fit (e.g. to 1205 TRs). The transcript for the combined “Slumlord” and “Reach for the Stars One Small Step at a Time” stimulus contains 5,344 words; 116 words (2.2%) successfully localized by the forced-alignment algorithm, and 57 words (1.1%) were not found in the vocabulary. The transcript for “Slumlord” contains 2,715 words; 65 words (2.4%) were not successfully localized and 25 words (0.9%) were not found in the vocabulary. The transcript for “Reach for the Stars One Small Step at a Time” contains 2,629 words; 51 words (1.9%) were not successfully localized and 32 words (1.2%) were not found in the vocabulary. We recommend excluding sub-139 due to a truncated acquisition time (as specified in the scan_exclude.json file in the code/ directory).

After the scans, each subject completed an assessment of their comprehension of the “Slumlord” story. To evaluate comprehension, subjects were presented with a written transcript of the “Slumlord” story with 77 blank segments, and were asked to fill in the omitted word or phrase for each blank (free response) to the best of their ability. The free responses were evaluated using Amazon Mechanical Turk (MTurk) crowdsourcing platform. MTurk workers rated the response to each omitted segment on a scale from 0–4 of relatedness to the correct response, where 0 indicates no response provided, 1 indicates an incorrect response unrelated to the correct response, and 4 indicates an exact match to the correct response. The resulting comprehension scores are reported as the proportion out of a perfect score of 4 (ranging 0–1) in the participants.tsv file and scans.tsv file for each subject. Comprehension scores for “Reach for the Stars One Small Step at a Time” are not provided. The “Slumlord” and “Reach for the Stars One Small Step at a Time” data have not been previously reported; however, a separate dataset not included in this release has been collected using the “Slumlord” story stimulus (Regev et al., 2019). In the filename convention, the combined “Slumlord” and “Reach for the Stars One Small Step at a Time” data are labeled slumlordreach, and labeled slumlord and reach when analyzed separately.

“It’s Not the Fall that Gets You”

The “It’s Not the Fall that Gets You” dataset was collected between May, 2013, and October, 2013, and comprised 56 participants (ages 18–29 years, mean age 21.0 ± 2.4 years, 31 reported female). The “It’s Not the Fall that Gets You” story was told by Andy Christie and recorded live at The Moth, a non-profit storyteller event, in New York City in 2009 (freely available at https://themoth.org/stories/its-not-the-fall-that-gets-you). In addition to the original story (labeled intact), two experimental manipulations of the story stimulus were created: (a) in one version, the story stimulus was temporally scrambled at a coarse level (labeled longscram); (b) in the other version, the story stimulus was temporally scrambled at a finer level (labeled shortscram). In both cases, the story was spliced into segments at the level of sentences prior to scrambling, and segment boundaries used for the longscram condition are a subset of the boundaries used for the shortscram condition. The boundaries used for scrambling occurred on multiples of 1.5 seconds to align with the TR during scanning acquisition. We recommend using only the intact version for studying narrative processing because the scrambled versions do not have a coherent narrative structure (by design). All three versions of the “It’s Not the Fall that Gets You” stimulus were 582 seconds (9.7 minutes) long and began with 22 seconds of neutral introductory music followed by 3 seconds of silence such that the story started at 0:25 (relative to the beginning of the stimulus file) and ended at 9:32 for a duration of 547 seconds (365 TRs), with 10 seconds of silence at the end of the stimulus. The stimulus file was started after 3 TRs (4.5 seconds) as indicated in the events.tsv files accompanying each scan. The functional scans comprised 400 TRs (600 seconds). The transcript for the intact stimulus contains 1,601 words total; 40 words (2.5%) were not successfully localized by the forced-alignment algorithm and 20 words (1.2%) were not found in the vocabulary.

The scrambled stimuli were assigned to subjects according to a mixed within- and between-subject design, such that subjects received the intact stimulus and either the longscram stimulus (23 participants) or the shortscram stimulus (24 participants). The group assignments are indicated in the participants.tsv file, and the scans.tsv file for each subject. Due to the mixed within- and between-subject design, the files are named with the full notthefallintact, notthefalllongscram, notthefallshortscram task labels. We recommend excluding the intact scans for sub-317 and sub-335, the longscram scans for sub-066 and sub-335, and the shortscram scan for sub-333 (as specified in the scan_exclude.json file in the code/ directory). The “It’s Not the Fall that Gets You” data have been recently reported by Chien and Honey (2020).

“Merlin” and “Sherlock”

The “Merlin” and “Sherlock” datasets were collected between May, 2014, and March, 2015, and comprised 36 participants (ages 18–47 years, mean age 21.7 ± 4.7 years, 22 reported female). The “Merlin” and “Sherlock” stimuli were recorded while an experimental participant recalled events from previously viewed television episodes during fMRI scanning. Note that this spontaneous recollection task (during fMRI scanning) is cognitively distinct from reciting a well-rehearsed story to an audience, making this dataset different from others in this release. This release contains only the data for subjects listening to the auditory verbal recall; not data for the audiovisual stimuli or the speaker’s fMRI data (Zadbood et al., 2017). The “Merlin” stimulus file is 915 seconds (9.25 minutes) long and began with 25 seconds of music followed by 4 seconds of silence such that the story started at 0:29 and ended at 15:15 for a duration of 886 seconds (591 TRs). The transcript for the “Merlin” stimulus contains 2,245 words; 111 words (4.9%) were not successfully localized by the forced-alignment algorithm and 13 words (0.6%) were not found in the vocabulary. The “Sherlock” stimulus file is 1,081 seconds (~18 minutes) long and began with 25 seconds of music followed by 4 seconds of silence such that the story started at 0:29 and ended at 18:01 for a duration of 1,052 seconds (702 TRs). The stimulus files for both stories were started after 3 TRs (4.5 seconds) as indicated in the events.tsv files accompanying each scan. The “Merlin” and “Sherlock” scans were acquired with a variable number of trailing volumes following the stimulus across subjects, but can be truncated as researchers see fit (e.g. to 614 and 724 TRs, respectively). The transcript for “Sherlock” contains 2681 words; 171 words (6.4%) were not successfully localized and 17 words (0.6%) were not found in the vocabulary. The word counts for “Merlin” and “Sherlock” may be slightly inflated due to the inclusion of disfluencies (e.g. “uh”) in transcribing the spontaneous verbal recall.

All 36 subjects listened to both the “Merlin” and “Sherlock” verbal recall auditory stimuli. However, 18 subjects viewed the “Merlin” audiovisual clip prior to listening to both verbal recall stimuli, while the other 18 subjects viewed the “Sherlock” audiovisual clip prior to listening to both verbal recall stimuli. In the “Merlin” dataset, the 18 subjects that viewed the “Sherlock” clip (and not the “Merlin” audiovisual clip) are labeled with the naive condition as they heard the “Merlin” verbal recall auditory-only stimulus without previously having seen the “Merlin” audiovisual clip. The other 18 subjects in the “Merlin” dataset viewed the “Merlin” audiovisual clip (and not the “Sherlock” clip) prior to listening to the “Merlin” verbal recall and are labeled with the movie condition. Similarly, in the “Sherlock” dataset, the 18 subjects that viewed the “Merlin” audiovisual clip (and not the “Sherlock” audiovisual clip) are labeled with the naive condition, while the 18 subjects that viewed the “Sherlock” audiovisual clip (and not the “Merlin” audiovisual clip) are labeled with the movie condition. These condition labels are indicated in the both the participants.tsv file and the scans.tsv files for each subject. Following the scans, each subject performed a free recall task to assess memory of the story. Subjects were asked to write down events from the story in as much detail as possible with no time limit. The quality of comprehension for the free recall text was evaluated by three independent raters on a scale of 1–10. The resulting comprehension scores are reported as the sum across raters normalized by the perfect score (range 0–1) in the participants.tsv file and scans.tsv file for each subject. Comprehension scores are only provided for the naive condition. We recommend excluding the “Merlin” scan for subject sub-158 and the “Sherlock” scan for sub-139 (as specified in the scan_exclude.json file in the code/ directory. The “Merlin” and “Sherlock” data have been previously reported by Zadbood and colleagues (2017). In the filename convention, “Merlin” is labeled merlin and “Sherlock” is labeled sherlock.

“Schema”

The “Schema” dataset was collected between August, 2015, and March, 2016, and comprised 31 participants (ages 18–35 years, mean age 23.7 ± 4.8 years, 13 reported female). The “Schema” dataset comprised eight brief (~3-minute) auditory narratives based on popular television series and movies featuring restaurants and airport scenes: The Big Bang Theory, Friends, How I Met Your Mother, My Cousin Vinnny, The Santa Clause, Shame, Seinfeld, Up in the Air; labeled bigbang, friends, himym, vinny, santa, shame, seinfeld, upintheair, respectively. These labels are indicated in the events.tsv accompanying each scanner run, and the corresponding stimuli (e.g. bigbang_audio.wav). Two of the eight stories were presented in each of four runs. Each run had a fixed presentation order of a fixed set of stories, but the presentation order of the four runs was counterbalanced across subjects. The functional scans were acquired with a variable number of trailing volumes across subjects, and in each run two audiovisual film clips (not described in this data release) were also presented; the auditory stories must be spliced from the full runs according to the events.tsv file. Overall, the transcripts for the auditory story stimuli in the “Schema” dataset contain 3,788 words; 20 words (0.5%) were not successfully localized by the forced-alignment algorithm and 69 words (1.8%) were not found in the vocabulary. The “Schema” data were previously reported by Baldassano and colleagues (2018). In the filename convention, the “Schema” data are labeled using the task alias schema.

“Shapes”

The “Shapes” dataset was collected between May, 2015, and July, 2016, and comprised 59 participants (ages 18–35 years, mean age 23.0 ± 4.5 years, 42 reported female). The “Shapes” dataset includes two related auditory story stimuli describing a 7-minute original animation called “When Heider Met Simmel” (copyright Yale University). The movie uses two-dimensional geometric shapes to tell the story of a boy who dreams about a monster, and was inspired by Heider and Simmel (1944). The movie itself is available for download at https://www.headspacestudios.org. The two verbal descriptions based on the movie used as auditory stimuli in the “Shapes” dataset were: (a) a purely physical description of the animation (labeled shapesphysical), and (b) a social description of the animation intended to convey intentionality in the animated shapes (labeled shapessocial). Note that the physical description story (shapesphysical) is different from other stimuli in this release in that it describes the movements of geometrical shapes without any reference to the narrative embedded in the animation. Both stimulus files are 458 seconds (7.6 minutes) long and began with a 37-second introductory movie (only the audio channel is provided in the stimulus files) followed by 7 seconds of silence such that the story started at 0:45 and ended at roughly 7:32 for a duration of ~408 seconds (~272 TRs) and ended with ~5 seconds of silence. The stimulus files for both stories were started after 3 TRs (4.5 seconds) as indicated in the events.tsv files accompanying each scan. The functional scans were acquired with a variable number of trailing volumes following the stimulus across subjects, but can be truncated as researchers see fit (e.g. to 309 TRs). The transcript for the shapesphysical stimulus contains 951 words; 9 words (0.9%) were not successfully localized by the forced-alignment algorithm and 25 words (2.6%) were not found in the vocabulary. The transcript for the shapessocial stimulus contains 910 words; 6 words (0.7%) were not successfully localized and 14 words (1.5%) were not found in the vocabulary.

Each subject received both the auditory physical description of the animation (shapesphysical), the auditory social description of the animation (shapessocial), and the audiovisual animation itself (not included in the current data release) in three separate runs. Run order was counterbalanced across participants, meaning that some participants viewed the animation before hearing the auditory description. The run order for each subject (e.g. physical-social-movie) is specified in the participants.tsv file and the subject-specific scans.tsv files. Immediately following the story scan, subjects performed a free recall task in the scanner (the fMRI data during recall are not provided in this release). They were asked to describe the story in as much detail as possible using their own words. The quality of the recall was evaluated by an independent rater naive to the purpose of the experiment on a scale of 1–10. The resulting comprehension scores are reported normalized by 10 (range 0–1) and reported in the participants.tsv file and scans.tsv file for each subject. Comprehension scores are only provided for scans where the subject first listened to the story stimulus, prior to hearing the other version of the story or viewing the audiovisual stimulus. We recommend excluding the shapessocial scan for subject sub-238 (as specified in the scan_exclude.json file in the code/ directory. The “Shapes” data were previously reported by Nguyen and colleagues (2019).

“The 21st Year”

The “The 21st Year” dataset was collected between February, 2016, and January, 2018, and comprised 25 participants (ages 18–41 years, mean age 22.6 ± 4.7 years, 13 reported female). The story stimulus was written and read aloud by author Christina Lazaridi, and includes two seemingly unrelated storylines relayed in blocks of prose that eventually fuse into a single storyline. The stimulus file is 3,374 seconds (56.2 minutes) long and began with 18 seconds of introductory music followed by 3 seconds such that the story started at 0:21 and ended at 55:59 for a duration of 3,338 seconds (2,226 TRs) followed by 15 seconds of silence. The functional scans comprised 2,249 volumes (3,373.5 seconds). The story stimulus contains 8,267 words; 88 words (1.1%) were not successfully localized by the forced-alignment algorithm and 143 words (1.7%) were not found in the vocabulary. Data associated with “The 21st Year” have been previously reported by Chang and colleagues (2020). In the filename convention, data for “The 21st Year” are labeled using the task alias 21styear.

“Pie Man (PNI)” and “Running from the Bronx”

The “Pie Man (PNI)” and “Running from the Bronx” data were collected between May and September, 2018, and comprised 47 participants (ages 18–53 years, mean age 23.3 ± 7.4 years, 33 reported female). The “Pie Man (PNI)” and “Running from the Bronx” story stimuli were told by Jim O’Grady while undergoing fMRI scans at the Princeton Neuroscience Institute (PNI). The spoken story was recorded using a noise cancelling microphone mounted on the head coil and scanner noise was minimized using Audacity. The “Pie Man (PNI)” stimulus file is 415 seconds (~7 minutes) long and begins with a period of silence such that the story starts at :09 and ends at 6:49 for a duration of 400 seconds (267 TRs) followed by 7 seconds of silence. The “Pie Man (PNI)” stimulus conveys the same general story as the original “Pie Man” stimulus but with differences in timing and delivery (as well as poorer audio quality due to being recorded during scanning). The “Running from the Bronx” stimulus file is 561 seconds (9.4 minutes) long and begins with a period of silence such that the story starts at 0:15 and ends at 9:11 for a duration of 536 seconds (358 TRs) followed by 10 seconds of silence. The “Pie Man (PNI)” and “Running from the Bronx” functional scans comprised 294 TRs (441 seconds) and 390 TRs (585 seconds), respectively. Both stimulus files were started after 8 TRs (12 seconds) as indicated in the accompanying events.tsv files. The program card on the scanner computer was organized according to the ReproIn convention (Visconti di Oleggio Castello et al., 2020) to facilitate automated conversion to BIDS format using HeuDiConv (0.5.dev1; Halchenko et al., 2020). The transcript for the “Pie Man (PNI)” stimulus contains 992 words; 7 words (0.7%) were not successfully localized by the forced-alignment algorithm and 19 words (1.9%) were not found in the vocabulary. The transcript for “Running from the Bronx” contains 1,379 words; 23 words (1.7%) were not successfully localized and 19 words (1.4%) were not found in the vocabulary.

At the end of the scanning session, the subjects completed separate questionnaires for each story to evaluate comprehension. Each questionnaire comprised 30 multiple choice and fill-in-the-blank questions testing memory and understanding of the story. The resulting comprehension scores are reported as the proportion of correct answers (ranging 0–1) in the participants.tsv file and the scans.tsv file for each subject. Data for both stories have been previously reported (Lin et al., 2019; Nastase et al., 2020b). The “Pie Man (PNI)” and “Running from the Bronx” data are labeled with the piemanpni and bronx task aliases in the filename convention. The “Pie Man (PNI)” and “Running from the Bronx” data were collected in conjunction with the “I Knew You Were Black” and “The Man Who Forgot Ray Bradbury” data and share the same samples of subjects.

“I Knew You Were Black” and “The Man Who Forgot Ray Bradbury”

The “I Knew You Were Black” and “The Man Who Forgot Ray Bradbury” data were collected by between May and September, 2018, and comprised the sample sample of subjects from “Pie Man (PNI)” and “Running from the Bronx” data (47 participants, ages 18–53 years, mean age 23.3 ± 7.4 years, 33 reported female). Unlike the “Pie Man (PNI)” and “Running from the Bronx” story stimuli, the “I Knew You Were Black” and “The Man Who Forgot Ray Bradbury” stimuli are professionally recorded with high audio quality. The “I Knew You Were Black” story was told by Carol Daniel and recorded live at The Moth, a non-profit storytelling event, in New York City in 2018 (freely available at https://themoth.org/stories/i-knew-you-were-black). The “I Knew You Were Black” story is 800 seconds (13.3 minutes, 534 TRs) long and occupies the entire stimulus file. The “I Knew You Were Black” functional scans comprised 550 TRs (825 seconds). The transcript for the “I Knew You Were Black” stimulus contains 1,544 words; 5 words (0.3%) were not successfully localized by the forced-alignment algorithm and 4 words (0.3%) were not found in the vocabulary. “The Man Who Forgot Ray Bradbury” was written and read aloud by author Neil Gaiman at the Aladdin Theater in Portland, OR, in 2011 (freely available at https://soundcloud.com/neilgaiman/the-man-who-forgot-ray-bradbury). The “The Man Who Forgot Ray Bradbury” audio stimulus file is 837 seconds (~14 minutes, 558 TRs) long and occupies the entire stimulus file. The “The Man Who Forgot Ray Bradbury” functional scans comprised 574 TRs (861 seconds). The transcript for “The Man Who Forgot Ray Bradbury” contains 2,135 words; 16 words (0.7%) were not successfully localized and 29 words (1.4%) were not found in the vocabulary. For both datasets, the audio stimulus was prepended by 8 TRs (12 seconds) and followed by 8 TRs (12 seconds) of silence. The program card on the scanner computer was organized according to the ReproIn convention (Visconti di Oleggio Castello et al., 2020) to facilitate automated conversion to BIDS format using HeuDiConv (0.5.dev1; Halchenko et al., 2020).

Similarly to the “Pie Man (PNI)” and “Running from the Bronx” stories, the subjects completed separate questionnaires for each story to evaluate comprehension after scanning. Each questionnaire comprised 25 multiple choice and fill-in-the-blank questions testing memory and understanding of the story. The resulting comprehension scores are reported as the proportion of correct answers (ranging 0–1) in the participants.tsv file and the scans.tsv file for each subject. Data for both stories have been previously reported (Lin et al., 2019; Nastase et al., 2020b). The “I Knew You Were Black” and “The Man Who Forgot Ray Bradbury” data are labeled with the black and forgot task aliases in the filename convention.

Data records

Data and derivatives have been organized according to the machine-readable Brain Imaging Data Structure v1.2.1 (BIDS; Gorgolewski et al., 2016), which follows the FAIR principles for making data findable, accessible, interoperable, and reusable (Wilkinson et al., 2016). A detailed description of the BIDS specification can be found at http://bids.neuroimaging.io. Organizing the data according to the BIDS convention facilitates future research by enabling the use of automated BIDS-compliant software (i.e., BIDS Apps; Gorgolewski et al., 2017). Briefly, files and metadata are labeled using key-value pairs where key and value are separated by a hyphen, while key-value pairs are separated by underscores. Each subject is identified by an anonymized numerical identifier (e.g. sub-001). Each subject participated in one or more story-listening scans indicated by the task alias for the story (e.g. task-pieman). Subject identifiers are conserved across datasets; subjects contributing to multiple datasets are indexed by the same identifier (e.g. sub-001 contributes to both the “Pie Man” and “Tunnel Under the World” datasets).

The top-level BIDS-formatted narratives/ directory contains a tabular participants.tsv file where each row corresponds to a subject and includes demographic variables (age, sex), the stories that subject received (task), as well as experimental manipulations (condition) and comprehension scores where available (comprehension). Cases of omitted, missing, or inapplicable data are indicated by “n/a” according to the BIDS convention. The top-level BIDS directory also contains the following: (a) a dataset_description.json containing top-level metadata, including licensing, authors, acknowledgments, and funding sources; (b) a code/ directory containing scripts used for analyzing the data; (c) a stimuli/ directory containing the audio stimulus files (e.g. pieman_audio.wav), transcripts (e.g. pieman_transcript.txt), and a gentle/ directory containing time-stamped transcripts created using Gentle (i.e.; described in “Stimuli” in the “Methods” section); (d) a derivatives/ directory containing all MRI derivatives (described in more detail in the “MRI preprocessing” and “Technical validation” sections); and (e) descriptive README and CHANGES files.

BIDS-formatted MRI data for each subject are contained in separate directories named according to the subject identifiers. Within each subject’s directory, data are segregated into anatomical data (the anat/ directory) and functional data (the func/ directory). All MRI data are stored as gzipped (compressed) NIfTI-1 images (Cox et al., 2004). NIfTI images and the relevant metadata were reconstructed from the original Digital Imaging and Communications in Medicine (DICOM) images using the dcm2niix utility (1.0.20180518; Li et al., 2016). In instances where multiple scanner runs were collected for the same imaging modality or story in the same subject, the files are differentiated using a run label (e.g. run-1, run-2). All anatomical and functional MRI data files are accompanied by sidecar JSON files containing MRI acquisition parameters in compliance with the COBIDAS report (Nichols et al., 2017). Identifying metadata (e.g. name, birth date, acquisition date and time) have been omitted, and facial features have been removed from anatomical images using the pydeface automated de-facing software (2.0.0; Gulban et al., 2019). All MRI data and metadata are released under the Creative Commons CC0 license (https://creativecommons.org/), which allows for free reuse without restriction.

All data and derivatives are version-controlled using DataLad (0.13.3; https://www.datalad.org/; RRID:SCR_003931; Hanke et al., 2020). DataLad uses git (https://git-scm.com/; RRID:SCR_003932) and git-annex (https://git-annex.branchable.com/; RRID:SCR_019087) to track changes made to the data, providing a transparent history of the dataset. Each stage of the dataset’s evolution is indexed by a unique commit hash and can be interactively reinstantated. Major subdivisions of the data collection, such as the code/, stimuli/, and derivative directories, are encapsulated as “subdatasets” nested within the top-level BIDS dataset. These subdatasets can be independently manipulated and have their own standalone version history. For a primer on using DataLad, we recommend the DataLad Handbook (http://handbook.datalad.org/; Wagner et al., 2020).

Technical validation

Image quality metrics

To provide an initial assessment of the data, we used MRIQC (0.15.1; Esteban et al., 2017; Esteban et al., 2019a) to derive a variety of image quality metrics (IQMs). MRIQC was deployed using Singularity (Kurtzer et al., 2017). The MRIQC outputs can be found in the mriqc/ directory in derivatives/ available via the DataLad release, and includes dozens of IQMs per scan as well as a summary across the entire dataset. These IQMs can be visualized in a browser (e.g. group_bold.html) and are aggregated in tabular format across the entire data collection (e.g. group_bold.tsv). Here we do not exclude any subjects based on IQMs, but subsequent researchers can use the available IQMs to exclude scans as they see fit.

We briefly report three common metrics for functional images summarizing head motion, intrinsic spatial smoothness, and temporal signal quality (Fig. 2). To summarize head motion, we computed the mean framewise displacement (FD; Power et al., 2012, 2014) for each functional scan across all subjects and story stimuli (Fig. 2a). The median framewise displacement across scans for all subjects and story stimuli was 0.132 mm (SD = 0.058 mm). Although this indicates relatively low head motion, some subjects have notably high motion; researchers may opt to censor time points with high FD (e.g. greater 0.5 mm) or omit entire scans with high overall FD as they see fit. To quantify the intrinsic spatial smoothness of the raw functional data, we used 3dFWHMx in AFNI (19.3.0), which computes the ratio of variance of first differences across voxels to global variance across the image and expresses smoothness as the full width at half-maximum (FWHM) of a 2D Gaussian filter (Forman et al., 1995). Smoothness was computed in each subject’s native brain space within an anatomically-defined brain mask, and the functional time series were temporally detrended prior to smoothness estimation. Spatial smoothness varied considerably across datasets due to different acquisition parameters used with different scanner models (Fig. 2b). For example, scans acquired using older pulse sequences without multiband acceleration are smoothest in the slice acquisition (z-) axis, whereas newer scans acquired using multiband acceleration have higher overall spatial resolution (resulting in lower smoothness), and are smoothest in anterior–posterior phase-encoding (y-) axis. Finally, we computed the temporal signal-to-noise ratio (tSNR) as a measure of functional signal quality using MRIQC. A tSNR map was constructed for each scan by computing the voxelwise average of the signal magnitude over time divided by the standard deviation of the signal over time in each subject’s native space following realignment using 3dvolreg and brain extraction using 3dSkullStrip in AFNI (Krüger & Glover, 2001); the median of this tSNR map was used to summarize tSNR across voxels for a given scan (Fig. 2c). The median tSNR across all subjects and story stimuli was 50.772 (SD = 13.854). We also computed vertex-wise tSNR in fsaverage6 space following preprocessing with fMRIPrep to visualize the spatial distribution of tSNR across cortex (Fig. 2d). Low tSNR in orbitofrontal and anterior medio-temporal cortex reflects signal dropout due to the proximity of sinuses and ear canals (Ojemann et al., 1997).

Fig. 2.
  • Download figure
  • Open in new tab
Fig. 2.

Data quality summarized in terms of head motion, spatial smoothness, and temporal signal-to-noise ratio (tSNR). (a) Median framewise displacement (FD) summarizes head motion. Each black tick mark represents the median FD (x-axis) for a given scanning run and subject for each task (i.e. story; y-axis). Violin plots illustrate the distribution of median FDs across all runs and subjects for a given task. The red markers indicate the median FD across all runs for a given task, and the red error bars indicate the 95% bootstrap confidence interval for the median based on randomly sampling runs with replacement. At bottom, FD is summarized across all tasks. (b) Spatial smoothness is summarized using AFNI’s FWHM smoothness for each scanning run across each task. Spatial smoothness was computed in volumetric subject-specific functional (EPI) space using subject-specific whole-brain masks with minimal preprocessing (realignment and susceptibility distortion in fMRIPrep, detrending in AFNI). Each tick mark represents the spatial smoothness for a given run in a given acquisition axis (orange: x-axis, i.e. left–right; yellow: y-axis, i.e. anterior–posterior; red: z-axis, i.e. inferior–superior). Violin plots capture the distribution of smoothness estimates for x-, y-, and z-axes across all runs and subjects in a given task. The red markers indicate the median combined (i.e. geometric mean across x-, y-, and z-axes) smoothness across all runs and subjects for a given task (red error bars indicate 95% bootstrap confidence interval of median). At bottom, smoothness is summarized across all tasks. The multiple lobes of the distribution reflect differing acquisition parameters used with the Skyra and Prisma scanner models. (c) tSNR maps were constructed by computing the voxelwise mean signal over time divided by the standard deviation over time. Each black tick mark represents the median of the tSNR map (x-axis) for a given scanner run and subject for each task (y-axis). Violin plots reflect the distribution of tSNR values across all runs and subjects for a given task. The red markers indicate the mean tSNR across all runs for a given task (red error bars indicate 95% bootstrap confidence interval of mean). At bottom, the tSNR is summarized across all tasks. The two lobes of the distribution reflect older pulse sequences with larger voxels (and higher tSNR) and newer multiband pulse sequences with smaller voxels (and lower tSNR). (d) Distribution of tSNR across cortex. The color bar reflects the median vertex-wise tSNR. Note that unlike panel c, here tSNR was computed in fsaverage6 space for vertex-wise summarization and visualization, which may inflate the tSNR values.

Intersubject correlation

To measure the reliability of stimulus-evoked responses across subjects, we performed an intersubject correlation (ISC) analysis (Hasson et al., 2004, 2010; Nastase et al., 2019a). For each story stimulus, we computed ISC using the leave-one-out approach: for each subject, we averaged the response time series for the remaining subjects, then correlated this mean response time series for a given voxel or brain area with the response time series for the left-out subject in that voxel or brain area. Note that this approach yields higher correlation values than computing ISC between pairs of subjects. In large samples of subjects, leave-one-out ISC provides an estimate of the upper bound of variance that can be accounted for by between-subject models predicting voxel- or region-wise response time series (Nastase et al., 2019a; i.e. noise ceiling; Nili et al., 2014); that is, the mean data from other subjects serves as a surrogate for the optimal model that generalizes across subjects (within-subject models, however, may capture additional variance not shared across subjects). ISC was computed only on functional volumes corresponding to presentation of the story stimulus, and we trimmed the first 6 TRs (9 seconds) after story onset to reduce the impact of stimulus onset transients. All ISC values were Fisher z-transformed prior to averaging, and the mean was inverse Fisher z-transformed for visualization (Silver & Dunlap, 1987). ISC analysis was implemented using the Brain Imaging Analysis Kit (BrainIAK; RRID:SCR_014824; Cohen et al., 2017; Kumar et al., 2020a, 2020b).

ISCs were computed separately for different experimental groups or manipulations; for example, ISCs were computed separately for the “affair” and “paranoid” contextual manipulation conditions for “Pretty Mouth and Green My Eyes.” On the other hand, ISCs were computed as a single group for datasets where the only experimental manipulation was naively listening to the story stimulus versus listening to the story stimulus after viewing the audiovisual movie (“Merlin,” “Sherlock,” and “Shapes”). ISCs were computed separately for the “Slumlord” and “Reach for the Stars One Small Step at a Time” stories despite being presented in the same scanning run. The brief “Schema” stories and the scrambled versions of “It’s Not the Fall that Gets You” were excluded from ISC analysis for the sake of simplicity.

To assess the temporal synchronization of responses to the story stimuli, we computed ISC in an early auditory cortex (EAC) region of interest (ROI). The left- and right-hemisphere EAC ROIs were defined on the cortical surface in fsaverage6 space as combination of five areas (A1, MBelt, LBelt, PBelt, RI) from a multimodal cortical parcellation (Glasser et al., 2016; Mills, 2016). The non-smoothed response time series following preprocessing with fMRIPrep and confound regression with AFNI were averaged across vertices in the left and right EAC ROIs. We used two measures of temporal alignment in left EAC as quality-control criteria for excluding scans. First, we computed leave-one-out ISCs and set an exclusion cut-off at ISC < .10 in left EAC (Fig. 3a). Second, we computed leave-one-out ISC at temporal lags ranging from −45 to 45 seconds; for each left-out subject, we computed the correlation between the mean response time series for the remaining subjects and the left-out subject’s response time series at 61 lags ranging from −30 to 30 TRs (Fig 3b). We then set an exclusion cut-off at a peak ISC at a lag > ±3 TRs (4.5 seconds) to accommodate some inter-individual variability in the hemodynamic response (Aguirre et al., 1998; Handwerker et al., 2004). Both of these criteria are intended to exclude scans with poor entrainment to the stimulus (e.g. due to an error in acquisition), and yield a partly overlapping group of scans for exclusion. Overall, the combined criteria resulted in excluding 27 scans from 19 subjects, or 3.0% of the 891 total scans. We do not exclude these subjects from the dataset entirely because some analyses (e.g. comparing methods for mitigating artefacts) may in fact benefit from the inclusion of lower-quality scans. Rather, we provide a scan_exclude.json file listing scans to exclude (and a convenience function exclude_scans.py) in the code/ directory so that researchers can easily exclude scans flagged as poor quality according to the criteria above. Note that some studies (e.g. Chien & Honey, 2020) temporally shift response time series to ensure that peak lags match across subjects; here, we do not shift or edit the scans, but provide the lagged ISCs for all scans (group_roi-EAC_lags.json in derivatives/afni-nosmooth/) so that researchers can apply temporal shifting as they see fit. Ultimately, the mean ISC in left and right EAC across all scans was .549 (SD = .173) and .493 (SD = .179), respectively (Fig. 3).

Fig. 3.
  • Download figure
  • Open in new tab
Fig. 3.

Intersubject correlation (ISC) in early auditory cortex (EAC). (a) ISC in left and right EAC across all subjects and tasks (i.e. stories). Each tick mark represents the leave-one-out ISC (x-axis) for a left-out subject in either left-(purple) or right-(orange) hemisphere EAC across tasks (y-axis). Violin plots illustrate the distribution of leave-one-out ISC across all left-out subjects for each task. The circular markers indicate the mean ISC across left-out subjects for each task and hemisphere (error bars indicate the 95% bootstrap confidence interval of the mean). At bottom, ISC for left and right EAC is summarized across all tasks. (b) Lagged ISC captures temporal synchronization in left EAC. Leave-one-out ISC (y-axis) was computed for left EAC aat 61 lags ranging from −30 to +30 TRs (-45 to 45 seconds; x-axis). In the upper plot, each line represents the mean ISC across lags for each task. At bottom, all lagged ISCs are visualized in a single plot where each left-out subject corresponds to a semi-transparent gray line; the red line reflects the average lagged ISC across all subjects and tasks.

Finally, we visualized the cortical distribution of ISCs (Fig. 4). Leave-one-out ISCs were computed on vertex-wise response time series in fsaverage6 surface space following preprocessing using fMRIPrep, spatial smoothing to 6 mm, and confound regression using AFNI. ISCs were computed separately for each story stimulus (and experimental group or condition, as described above), then averaged across all stories to construct a summary ISC map (Fig. 4a). Here we are not interested in statistical testing and therefore arbitrarily threshold these ISC maps at a mean correlation of .10 for visualization; we refer researchers to work by Chen and colleagues (2016, 2017, 2020) for a detailed statistical treatment. Cortical areas with high ISC during story-listening were largely bilateral, spanning temporal and frontal areas reported in the fMRI literature on language, and extending into areas default-mode network (DMN) areas such as the angular gyrus and precuneus. Cortical ISC maps are strikingly similar across stories with considerable variability in sample size, duration, and content (Fig. 4b). We emphasize that a wide variety of stimulus content can drive cortical ISCs, and that decomposing the stimuli according to different models will ultimately provide a window onto the linguistic and narrative features encoded across cortex.

Fig. 4.
  • Download figure
  • Open in new tab
Fig. 4.

Intersubject correlations across cortex. (a) The mean vertex-wise leave-one-out ISC was computed across all subjects and stories. The color bar reflects the mean leave-one-out ISC and ranges from 0 to .60. ISC maps were arbitrarily thresholded at ISC = .10. (b) Mean ISC maps for 12 representative stories plotted according to the same color map and thresholded at ISC = .10. Where applicable (e.g. “Pretty Mouth and Green My Eyes”), ISCs were computed separately for different experimental groups or conditions, then averaged; ISCs were also computed separately for “Slumlord” and “Reach for the Stars One Small Step at a Time,” then averaged.

Usage notes

The core, unprocessed NIfTI-formatted MRI data with accompanying metadata and stimuli are publicly available on OpenNeuro: https://openneuro.org/datasets/ds002345. The data are organized according to the BIDS standard, and are therefore well-suited for automated processing using BIDS Apps (e.g. fMRIPrep, MRIQC; https://bids-apps.neuroimaging.io/) and the Brainlife cloud platform (https://brainlife.io/). All data and derivatives are hosted at the International Neuroimaging Data-sharing Initiative (INDI; http://fcon_1000.projects.nitrc.org/indi/retro/narratives.html; RRID:SCR_005361; Mennes et al., 2013) via the Neuroimaging Informatics Tools and Resources Clearinghouse (NITRC; https://www.nitrc.org/; RRID:SCR_003430; Kennedy et al., 2016). The full data collection is available via the DataLad data distribution (RRID:SCR_019089): http://datasets.datalad.org/?dir=/labs/hasson/narratives. The full data collection can be easily “installed” using DataLad (datalad install -r///labs/hasson/narratives). This clones the file hierarchy without downloading any large content files, which can be downloaded selectively as needed (datalad get).

Code availability

All code used for aggregating and analyzing the data is version-controlled and publicly available via the associated GitHub repository (https://github.com/snastase/narratives) and the code/ directory in the top-level BIDS directory distributed via DataLad (https://datasets.datalad.org/?dir=/labs/hasson/narratives).

Author contributions

S.A.N. converted the data to BIDS format, analyzed the data, and wrote the paper. S.A.N., Y.F.L., L.H., H.H., A.Z., and E.M. curated the data. J.C., C.J.H., E. S., M.A.C., M.R. and U.H. designed and collected the “Pie Man” dataset. O.L. and U.H. designed and collected the “Tunnel Under the World” and “Lucy” datasets. Y.Y. and U.H. designed and collected the “Pretty Mouth and Green My Eyes” and “Milky Way” datasets. J.C., M.R., M.C. and U.H. designed and collected the “Slumlord”, “Reach for the Stars One Small Step at a Time”, and “It’s Not the Fall that Gets You” datasets. A.Z., J.C., Y.C.L., K.A.N., and U.H. designed and collected the “Merlin” and “Sherlock” datasets. C.B., U.H., and K.A.N. designed and collected the “Schema” dataset. M.N., T.V., and U.H. designed and collected the “Shapes” dataset. C.H.C.C., Y.Y., K.A.N., and U.H. designed and collected the “The 21st Year” dataset. S.N., P.P.B., and U.H. designed and collected the “Pie Man”, “Running from the Bronx”, “I Knew You Were Black”, and “The Man Who Forgot Ray Bradbury” datasets. G.C. and A.G. provided technical assistance in stimulus transcription. Y.O.H. provided technical assistance with DataLad. U.H., K.A.N., and Y.O.H. supervised the project.

Competing interests

The authors declare no competing interests.

Acknowledgments

We thank Leigh Nystrom, Mark Pinsk, Garrett McGrath, and the administrative staff at the Scully Center for the Neuroscience of Mind and Behavior and the Princeton Neuroscience Institute, as well as Elizabeth McDevitt, Anne Mennen, and members of Pygers support group. We thank Franklin Feingold and Alexandre Franco for assistance in data sharing, as well as Chris Gorgolewski, Tal Yarkoni, Satrajit S. Ghosh, Cameron Craddock, Avital Hahamy, Mohamed Amer, Indranil Sur, Xiao Lin, and Ajay Divarakian for helpful feedback on the data and analysis. We also thank the International Neuroimaging Data-Sharing Initiative (INDI) for hosting the full derivatives dataset. This work was supported by the National Institutes of Health (NIH) grants R01-MH094480 (U.H.), DP1-HD091948 (U.H.), R01-MH112566 (U.H.), R01-MH112357 (K.A.N., U.H), T32-MH065214 (K.A.N), by the Defense Advanced Research Projects Agency (DARPA) Brain-to-Brain Seedling contract number FA8750-18-C-0213 (U.H.), and by the Intel Corporation. The views, opinions, and/or conclusions contained in this paper are those of the authors and should not be interpreted as representing the official views or policies, either expressed or implied of the NIH, DARPA, or Intel.

Footnotes

  • https://openneuro.org/datasets/ds002345

  • http://datasets.datalad.org/?dir=/labs/hasson/narratives

  • https://github.com/snastase/narratives

  • https://snastase.github.io/datasets/ds002345

References

  1. ↵
    Abraham, A., Pedregosa, F., Eickenberg, M., Gervais, P., Mueller, A., Kossaifi, J., Gramfort, A., Thirion, B., & Varoquaux, G. (2014). Machine learning for neuroimaging with scikit-learn. Frontiers in Neuroinformatics, 8, 14. https://doi.org/10.3389/fninf.2014.00014
    OpenUrl
  2. ↵
    Aguirre, G. K., Zarahn, E., & D’esposito, M. (1998). The variability of human, BOLD hemodynamic responses. NeuroImage, 8(4), 360–369. https://doi.org/10.1006/nimg.1998.0369
    OpenUrlCrossRefPubMedWeb of Science
  3. ↵
    Alexander, L. M., Escalera, J., Ai, L., Andreotti, C., Febre, K., Mangone, A., Vega-Potler, N., Langer, N., Alexander, A., Kovacs, M., Litke, S., O’Hagan, B., Andersen, J., Bronstein, B., Bui, A., Bushey, M., Butler, H., Castagna, V., Camacho, N., Chan, E., Citera, D., Clucas, J., Cohen, S., Dufek, S., Eaves, M., Fradera, B., Gardner, J., Grant-Villegas, N., Green, G., Gregory, C., Hart, E., Harris, S., Horton, M., Khan, D., Kabotyanski, K., Karmel, B., Kelly, S. P., Kleinman, K., Koo, B., Kramer, E., Lennon, E., Lord, C., Mantello, G., Margolis, A., Merikangas, K. R., Milham, J., Minniti, G., Neuhaus, R., Levine, A., Osman, Y., Parra, L. C., Pugh, K. R., Racanello, A., Restrepo, A., Saltzman, T., Septimus, B., Tobe, R., Waltz, R., Williams, A., Yeo, A., Castellanos, F. X., Klein, A., Paus, T., Leventhal, B. L., Craddock, R. C., Koplewicz, H. S., & Milham, M. P. (2017). An open resource for transdiagnostic research in pediatric mental health and learning disorders. Scientific Data, 4, 170181. https://doi.org/10.1038/sdata.2017.181
    OpenUrl
  4. ↵
    Aliko, S., Huang, J., Gheorghiu, F., Meliss, S., & Skipper, J. I. (2020). A naturalistic neuroimaging database for understanding the brain using ecological stimuli. Scientific Data, 7, 347. https://doi.org/10.1038/s41597-020-00680-2
    OpenUrl
  5. ↵
    Avants, B. B., Epstein, C. L., Grossman, M., & Gee, J. C. (2008). Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain. Medical Image Analysis, 12(1), 26–41. https://doi.org/10.1016/j.media.2007.06.004
    OpenUrlCrossRefPubMedWeb of Science
  6. ↵
    Baldassano, C., Chen, J., Zadbood, A., Pillow, J. W., Hasson, U., & Norman, K. A. (2017). Discovering event structure in continuous narrative perception and memory. Neuron, 95(3), 709–721.e5. https://doi.org/10.1016/j.neuron.2017.06.041
    OpenUrl
  7. ↵
    Baldassano, C., Hasson, U., & Norman, K. A. (2018). Representation of real-world event schemas during narrative perception. Journal of Neuroscience, 38(45), 9689–9699. https://doi.org/10.1523/jneurosci.0251-18.2018
    OpenUrlAbstract/FREE Full Text
  8. ↵
    Baroni, M. (2020). Linguistic generalization and compositionality in modern artificial neural networks. Philosophical Transactions of the Royal Society of London B: Biological Sciences, 375(1791), 20190307. https://doi.org/10.1098/rstb.2019.0307
    OpenUrl
  9. ↵
    Behzadi, Y., Restom, K., Liau, J., & Liu, T. T. (2007). A component based noise correction method (CompCor) for BOLD and perfusion based fMRI. NeuroImage, 37(1), 90–101. https://doi.org/10.1016/j.neuroimage.2007.04.042
    OpenUrlCrossRefPubMedWeb of Science
  10. ↵
    Belin, P., Zatorre, R. J., Lafaille, P., Ahad, P., & Pike, B. (2000). Voice-selective areas in human auditory cortex. Nature, 403(6767), 309–312. https://doi.org/10.1038/35002078
    OpenUrlCrossRefPubMedWeb of Science
  11. ↵
    Ben-Shachar, M., Hendler, T., Kahn, I., Ben-Bashat, D., & Grodzinsky, Y. (2003). The neural reality of syntactic transformations: evidence from functional magnetic resonance imaging. Psychological Science, 14(5), 433–440. https://doi.org/10.1111/1467-9280.01459
    OpenUrlCrossRefPubMedWeb of Science
  12. ↵
    Ben-Yakov, A., Honey, C. J., Lerner, Y., & Hasson, U. (2012). Loss of reliable temporal structure in event-related averaging of naturalistic stimuli. NeuroImage, 63(1), 501–506. https://doi.org/10.1016/j.neuroimage.2012.07.008
    OpenUrlCrossRefPubMedWeb of Science
  13. ↵
    Berwick, R. C., Friederici, A. D., Chomsky, N., & Bolhuis, J. J. (2013). Evolution, brain, and the nature of language. Trends in Cognitive Sciences, 17(2), 89–98. https://doi.org/10.1016/j.tics.2012.12.002
    OpenUrlCrossRefPubMed
  14. ↵
    Betzel, R. F., Byrge, L., Esfahlani, F. Z., & Kennedy, D. P. (2020). Temporal fluctuations in the brain’s modular architecture during movie-watching. NeuroImage, 213, 116687. https://doi.org/10.1016/j.neuroimage.2020.116687
    OpenUrl
  15. ↵
    Binder, J. R., Desai, R. H., Graves, W. W., & Conant, L. L. (2009). Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cerebral Cortex, 19(12), 2767–2796. https://doi.org/10.1093/cercor/bhp055
    OpenUrlCrossRefPubMedWeb of Science
  16. ↵
    Biswal, B. B., Mennes, M., Zuo, X.-N., Gohel, S., Kelly, C., Smith, S. M., Beckmann, C. F., Adelstein, J. S., Buckner, R. L., Colcombe, S., Dogonowski, A.-M., Ernst, M., Fair, D., Hampson, M., Hoptman, M. J., Hyde, J. S., Kiviniemi, V. J., Kötter, R., Li, S.-J., Lin, C.-P., Lowe, M. J., Mackay, C. M., Madden, D. J., Madsen, K. H., Margulies, D. S., Mayberg, H. S., McMahon, K., Monk, C. S., Mostofsky, S. H., Nagel, B. J., Pekar, J. J., Peltier, S. J., Pettersen, S. E., Riedl, V., Rombouts, S. A. R. B., Rypma, B., Schlaggar, B. L., Schmidt, S., Seidler, R. D., Siegle, G. J., Sorg, C., Teng, G.-J., Veijola, J., Villringer, A., Walter, M., Wang, L., Weng, X.-C., Whitfield-Gabrieli, S., Williamson, P., Windischberger, C., Zang, Y.-F., Zhang, H.-Y., Castellanos, F. X., & Milham, M. P. (2010). Toward discovery science of human brain function. Proceedings of the National Academy of Sciences of the United States of America, 107(10), 4734–4739. https://doi.org/10.1073/pnas.0911855107
    OpenUrlAbstract/FREE Full Text
  17. ↵
    Blank, I. A., & Fedorenko, E. (2017). Domain-general brain regions do not track linguistic input as closely as language-selective regions. Journal of Neuroscience, 37(41), 9999–10011. https://doi.org/10.1523/jneurosci.3642-16.2017
    OpenUrlAbstract/FREE Full Text
  18. ↵
    Bookheimer, S. (2002). Functional MRI of language: new approaches to understanding the cortical organization of semantic processing. Annual Review of Neuroscience, 25, 151–188. https://doi.org/10.1146/annurev.neuro.25.112701.142946
    OpenUrlCrossRefPubMedWeb of Science
  19. ↵
    Braga, R. M., DiNicola, L. M., Becker, H. C., & Buckner, R. L. (2020). Situating the left-lateralized language network in the broader organization of multiple specialized large-scale distributed networks. Journal of Neurophysiology. https://doi.org/10.1152/jn.00753.2019
  20. ↵
    Brainard, D. H. (1997). The Psychophysics Toolbox. Spatial Vision, 10(4), 433–436. https://doi.org/10.1163/156856897X00357
    OpenUrlCrossRefPubMedWeb of Science
  21. ↵
    Breiman, L. (2001). Statistical modeling: the two cultures. Statistical Science, 16(3), 199–231. https://doi.org/10.1214/ss/1009213726
    OpenUrlCrossRefPubMedWeb of Science
  22. ↵
    Brennan, J., Nir, Y., Hasson, U., Malach, R., Heeger, D. J., & Pylkkänen, L. (2012). Syntactic structure building in the anterior temporal lobe during natural story listening. Brain and Language, 120(2), 163–173. https://doi.org/10.1016/j.bandl.2010.04.002
    OpenUrlCrossRefPubMedWeb of Science
  23. ↵
    Brennan, J. R., Stabler, E. P., Van Wagenen, S. E., Luh, W.-M., & Hale, J. T. (2016). Abstract linguistic structure correlates with temporal activity during naturalistic comprehension. Brain and Language, 157–158, 81–94. https://doi.org/10.1016/j.bandl.2016.04.008
    OpenUrlCrossRefPubMed
  24. ↵
    Brett, M., Markiewicz, C. J., Hanke, M., Côté, M.-A., Cipollini, B., McCarthy, P., Jarecka, D., Cheng, C. P., Halchenko, Y. O., Cottaar, M., Ghosh, S., Larson, E., Wassermann, D., Gerhard, S., Lee, G. R., Wang, H.-T., Kastman, E., Kaczmarzyk, J., Guidotti, R., Duek, O., Rokem, A., Madison, C., Morency, F. C., Moloney, B., Goncalves, M., Markello, R., Riddell, C., Burns, C., Millman, J., Gramfort, A., Leppäkangas, J., Sólon, A., van den Bosch, J. J. F., Vincent, R. D., Braun, H., Subramaniam, K., Gorgolewski, K. J., Raamana, P. R., Nolan, N. B., Baker, E. M., Hayashi, S., Pinsard, B., Haselgrove, C., Hymers, M., Esteban, O., Koudoro, S., Oosterhof, N. N., Amirbekian, B., Nimmo-Smith, I., Nguyen, L., Reddigari, S., St-Jean, S., Panfilov, E., Garyfallidis, E., Varoquaux, G., Legarreta, J. H., Hahn, K. S., Hinds, O. P., Fauber, B., Poline, J.-B., Stutters, J., Jordan, K., Cieslak, M., Moreno, M. E., Haenel, V., Schwartz, Y., Baratz, Z., Darwin, B. C., Thirion, B., Papadopoulos Orfanoos, D., Pérez-García, F., Solovey, I., Gonzalez, I., Palasubramaniam, J., Lecher, J., Leinweber, K., Raktivan, K., Fischer, P., Gervais, P., Gadde, S., Ballinger, T., Roos, T., Reddam, V. R., & freec84. (2020). nipy/nibabel: 3.1.1. https://doi.org/10.5281/zenodo.3924343
  25. ↵
    Chang, C. H. C., Lazaridi, C., Yeshurun, Y., Norman, K. A., & Hasson, U. (2020). Relating the past with the present: Information integration and segregation during ongoing narrative processing. bioRxiv. https://doi.org/10.1101/2020.01.16.908731
  26. ↵
    Chang, L. J., Jolly, E., Cheong, J. H., Rapuano, K., Greenstein, N., Chen, P.-H. A., & Manning, J. R. (2018). Endogenous variation in ventromedial prefrontal cortex state dynamics during naturalistic viewing reflects affective experience. bioRxiv. https://doi.org/10.1101/487892
  27. ↵
    Chen, G., Shin, Y.-W., Taylor, P. A., Glen, D. R., Reynolds, R. C., Israel, R. B., & Cox, R. W. (2016). Untangling the relatedness among correlations, part I: nonparametric approaches to inter-subject correlation analysis at the group level. NeuroImage, 142, 248–259. https://doi.org/10.1016/j.neuroimage.2016.05.023
    OpenUrl
  28. ↵
    Chen, G., Taylor, P. A., Qu, X., Molfese, P. J., Bandettini, P. A., Cox, R. W., & Finn, E. S. (2020). Untangling the relatedness among correlations, part III: inter-subject correlation analysis through Bayesian multilevel modeling for naturalistic scanning. NeuroImage, 216, 116474. https://doi.org/10.1016/j.neuroimage.2019.116474
    OpenUrl
  29. ↵
    Chen, G., Taylor, P. A., Shin, Y.-W., Reynolds, R. C., & Cox, R. W. (2017). Untangling the relatedness among correlations, part II: inter-subject correlation group analysis through linear mixed-effects modeling. NeuroImage, 147, 825–840. https://doi.org/10.1016/j.neuroimage.2016.08.029
    OpenUrl
  30. ↵
    1. C. Cortes,
    2. N. D. Lawrence,
    3. D. D. Lee,
    4. M. Sugiyama, &
    5. R. Garnett
    Chen, P.-H., Chen, J., Yeshurun, Y., Hasson, U., Haxby, J., & Ramadge, P. J. (2015). A reduced-dimension fMRI shared response model. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, & R. Garnett (Eds.), Advances in Neural Information Processing Systems (pp. 460–468). Curran Associates, Inc. https://proceedings.neurips.cc/paper/2015/file/b3967a0e938dc2a6340e258630febd5a-Paper.pdf
  31. ↵
    Chien, H.-Y. S., & Honey, C. J. (2020). Constructing and forgetting temporal context in the human cerebral cortex. Neuron, 106(4), 675–686.e11. https://doi.org/10.1016/j.neuron.2020.02.013
    OpenUrl
  32. ↵
    Chung, M. K., Robbins, S. M., Dalton, K. M., Davidson, R. J., Alexander, A. L., & Evans, A. C. (2005). Cortical thickness analysis in autism with heat kernel smoothing. NeuroImage, 25(4), 1256–1265. https://doi.org/10.1016/j.neuroimage.2004.12.052
    OpenUrlCrossRefPubMedWeb of Science
  33. ↵
    Cieri, C., Miller, D., & Walker, K. (2004). The Fisher Corpus: a resource for the next generations of speech-to-text. In Proceedings of the Fourth International Conference on Language Resources and Evaluation (pp. 69–71). http://www.lrec-conf.org/proceedings/lrec2004/pdf/767.pdf
  34. ↵
    Ciric, R., Wolf, D. H., Power, J. D., Roalf, D. R., Baum, G. L., Ruparel, K., Shinohara, R. T., Elliott, M. A., Eickhoff, S. B., Davatzikos, C., Gur, R. C., Gur, R. E., Bassett, D. S., & Satterthwaite, T. D. (2017). Benchmarking of participant-level confound regression strategies for the control of motion artifact in studies of functional connectivity. NeuroImage, 154, 174–187. https://doi.org/10.1016/j.neuroimage.2017.03.020
    OpenUrlCrossRef
  35. ↵
    Cohen, J. D., Daw, N., Engelhardt, B., Hasson, U., Li, K., Niv, Y., Norman, K. A., Pillow, J., Ramadge, P. J., Turk-Browne, N. B., & Willke, T. L. (2017). Computational approaches to fMRI analysis. Nature Neuroscience, 20(3), 304–313. https://doi.org/10.1038/nn.4499
    OpenUrlCrossRefPubMed
  36. ↵
    Cox, R. W. (1996). AFNI: software for analysis and visualization of functional magnetic resonance neuroimages. Computers and Biomedical Research, 29(3), 162–173. https://doi.org/10.1006/cbmr.1996.0014
    OpenUrlCrossRefPubMedWeb of Science
  37. ↵
    Cox, R. W. (2012). AFNI: what a long strange trip it’s been. NeuroImage, 62(2), 743–747. https://doi.org/10.1016/j.neuroimage.2011.08.056
    OpenUrlCrossRefPubMedWeb of Science
  38. ↵
    Cox, R. W., Ashburner, J., Breman, H., Fissell, K., Haselgrove, C., Holmes, C. J., Lancaster, J. L., Rex, D. E., Smith, S. M., Woodward, J. B., & Strother, S. C. (2004). A (sort of) new image data format standard: NIfTI-1. 10th Annual Meeting of the Organization for Human Brain Mapping, Budapest, Hungary. http://nifti.nimh.nih.gov/nifti-1/documentation/hbm_nifti_2004.pdf
  39. ↵
    Cox, R. W., & Hyde, J. S. (1997). Software tools for analysis and visualization of fMRI data. NMR in Biomedicine, 10(4-5), 171–178. https://doi.org/10.1002/(SICI)1099-1492(199706/08)10:4/5%3C171::AID-NBM453%3E3.0.CO;2-L
    OpenUrlCrossRefPubMedWeb of Science
  40. ↵
    Dale, A. M., Fischl, B., & Sereno, M. I. (1999). Cortical surface-based analysis: I. segmentation and surface reconstruction. NeuroImage, 9(2), 179–194. https://doi.org/10.1006/nimg.1998.0395
    OpenUrlCrossRefPubMedWeb of Science
  41. ↵
    Dapretto, M., & Bookheimer, S. Y. (1999). Form and content: dissociating syntax and semantics in sentence comprehension. Neuron, 24(2), 427–432. https://doi.org/10.1016/s0896-6273(00)80855-7
    OpenUrlCrossRefPubMedWeb of Science
  42. ↵
    de Heer, W. A., Huth, A. G., Griffiths, T. L., Gallant, J. L., & Theunissen, F. E. (2017). The hierarchical cortical organization of human speech processing. Journal of Neuroscience, 37(27), 6539–6557. https://doi.org/10.1523/jneurosci.3267-16.2017
    OpenUrlAbstract/FREE Full Text
  43. ↵
    Démonet, J. F., Chollet, F., Ramsay, S., Cardebat, D., Nespoulous, J. L., Wise, R., Rascol, A., & Frackowiak, R. (1992). The anatomy of phonological and semantic processing in normal subjects. Brain, 115(6), 1753–1768. https://doi.org/10.1093/brain/115.6.1753
    OpenUrlCrossRefPubMedWeb of Science
  44. ↵
    Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2018). BERT: pre-training of deep bidirectional transformers for language understanding. arXiv. http://arxiv.org/abs/1810.04805
  45. ↵
    DuPre, E., Hanke, M., & Poline, J.-B. (2020). Nature abhors a paywall: how open science can realize the potential of naturalistic stimuli. NeuroImage, 216, 116330. https://doi.org/10.1016/j.neuroimage.2019.116330
    OpenUrl
  46. ↵
    Esteban, O., Birman, D., Schaer, M., Koyejo, O. O., Poldrack, R. A., & Gorgolewski, K. J. (2017). MRIQC: advancing the automatic prediction of image quality in MRI from unseen sites. PLOS One, 12(9), e0184661. https://doi.org/10.1371/journal.pone.0184661
    OpenUrlCrossRef
  47. ↵
    Esteban, O., Blair, R., Markiewicz, C. J., Varada, J., Nielson, D., Poldrack, R. A., & Gorgolewski, K. J. (2019a). MRIQC: Advancing the automatic prediction of image quality in MRI from unseen sites (Version 0.15.1). Zenodo. http://doi.org/10.5281/zenodo.3352432
  48. ↵
    Esteban, O., Ciric, R., Markiewicz, C. J., Poldrack, R. A., & Gorgolewski, K. J. (2020a). TemplateFlow Client: accessing the library of standardized neuroimaging standard spaces. Zenodo. https://doi.org/10.5281/zenodo.2583289
  49. ↵
    Esteban, O., Markiewicz, C. J., Blair, R. W., Moodie, C. A., Isik, A. I., Erramuzpe, A., Kent, J. D., Goncalves, M., DuPre, E., Snyder, M., Oya, H., Ghosh, S. S., Wright, J., Durnez, J., Poldrack, R. A., & Gorgolewski, K. J. (2019b). fMRIPrep: a robust preprocessing pipeline for functional MRI. Nature Methods, 16(1), 111–116. https://doi.org/10.1038/s41592-018-0235-4
    OpenUrl
  50. ↵
    Esteban, O., Markiewicz, C. J., DuPre, E., Goncalves, M., Kent, J. D., Ciric, R., Blair, R. W., Poldrack, R. A., & Gorgolewski, K. J. (2020b). fMRIPrep: a robust preprocessing pipeline for functional MRI (Version 20.0.5). Zenodo. http://doi.org/10.5281/zenodo.3724468
  51. ↵
    Esteban, O., Markiewicz, C. J., Johnson, H., Ziegler, E., Manhães-Savio, A., Jarecka, D., Burns, C., Ellis, D. G., Hamalainen, C., Notter, M. P., Yvernault, B., Salo, T., Waskom, M., Goncalves, M., Jordan, K., Wong, J., Dewey, B. E., Madison, C., Benderoff, E., Clark, D., Loney, F., Clark, D., Keshavan, A., Joseph, M., Nielsen, D. M., Dayan, M., Modat, M., Gramfort, A., Bougacha, S., Pinsard, B., Berleant, S., Christian, H., Rokem, A., Visconti di Oleggio Castello, M., Halchenko, Y. O., Kaczmarzyk, J., Varoquaux, G., Ciric, R., Moloney, B., DuPre, E., Koudoro, S., Clark, M. G., Cipollini B., Wassermann, D., Guillon, J., Markello, R., Hanke, M., Buchanan, C., Tungaraza, R., Gillman, A., Pauli, W. M., de Hollander, G., Sikka, S., Forbes, J., Mordom, D., Iqbal, S., Mancini, M., Malone, I. B., Dubois, M., Schwartz, Y., Frohlich, C., Tabas, A., Welch, D., Richie-Halford, A., Tilley II, S., Watanabe, A., Nichols, N. B., Huntenberg, J. M., Eshaghi, A., Ginsburg, D., Schaefer, A., Bottenhorn, K., Cumba, C., Acland, B., Heinsfeld, A. S., Kastman, E., Kent, J. D., Kleesiek, J., Ghayoor, A., Erickson, D., Giavasis, S., De La Vega, A., Liem, F., Küttner, R., Perez-Guevara, M. F., Lee, J. A., Millman, J., Lai, J., Zhou, D., Haselgrove, C., Glen, D., Doll, A., Renfro, M., Correa, C., Liu, S., Lampe, L., Kong, X.-Z., Hallquist, M., Kim, S., Kahn, A. E., Glatard, T., Triplett, W., Chawla, K., Salvatore, J., Pérez-García, F., Feilong, M., Park, A., Craddock, R. C., Hinds, O., Poldrack, R., Perkins, L. N., Stojic, H., Chetverikov, A., Inati, S., Grignard, M., Snoek, L., Sisk, L. M., Leinweber, K., Wen, J., Urchs, S., Blair, R., Matsubara, K., Floren, A., Mattfeld, A., Gerhard, S., Stadler, J., Cooper, G., Haehn, D., Broderick, W., Andberg, S. K., Noel, M., Cieslak, M., Durnez, J., Condamine, E., Papadopoulos Orfanos, D., Geisler, D., Meyers, B., Tambini, A., Weinstein, A., González Orozco, A. A., Harms, R., Khanuja, R., Sharp, P., Stanley, O., Lee, N., Crusoe, M. R., Brett, M., Falkiewicz, M., Weninger, L., Podranski, K., Linkersdörfer, J., Flandin, G., Lerma-Usabiaga, G., Tarbert, C., Cheung, B., Van, A., Davison, A., Shachnev, D., Molina-Romero, M., Rothmei, S., Bilgel, M., Schlamp, K., Ort, E., McNamee, D., Arias, J., Bielievtsov, D., Steele, C. J., Huang, L., Gonzalez, I., Warner, J., Margulies, D., Contier, O., Marina, A., Saase, V., Nickson, T., Varada, J., Schwabacher, I., Pellman, J., Pannetier, N., McDermottroe, C., Mihai, P. G., Lai, J., Gorgolewski, K. J., & Ghosh, S. (2020c). nipy/nipype (Version 1.4.2). Zenodo. https://doi.org/10.5281/zenodo.3668316
  52. ↵
    Evans, A. C., Janke, A. L., Collins, D. L., & Baillet, S. (2012). Brain templates and atlases. NeuroImage, 62(2), 911–922. https://doi.org/10.1016/j.neuroimage.2012.01.024
    OpenUrlCrossRefPubMedWeb of Science
  53. ↵
    Fedorenko, E., Hsieh, P.-J., Nieto-Castañón, A., Whitfield-Gabrieli, S., & Kanwisher, N. (2010). New method for fMRI investigations of language: defining ROIs functionally in individual subjects. Journal of Neurophysiology, 104(2), 1177–1194. https://doi.org/10.1152/jn.00032.2010
    OpenUrlCrossRefPubMedWeb of Science
  54. ↵
    Feilong, M., Nastase, S. A., Guntupalli, J. S., & Haxby, J. V. (2018). Reliable individual differences in fine-grained cortical functional architecture. NeuroImage, 183, 375–386. https://doi.org/10.1016/j.neuroimage.2018.08.029
    OpenUrl
  55. ↵
    Ferguson, A. R., Nielson, J. L., Cragin, M. H., Bandrowski, A. E., & Martone, M. E. (2014). Big data from small data: data-sharing in the “long tail” of neuroscience. Nature Neuroscience, 17(11), 1442–1447. https://doi.org/10.1038/nn.3838
    OpenUrlCrossRefPubMed
  56. ↵
    Finn, E. S., Corlett, P. R., Chen, G., Bandettini, P. A., & Constable, R. T. (2018). Trait paranoia shapes inter-subject synchrony in brain activity during an ambiguous social narrative. Nature Communications, 9, 2043. https://doi.org/10.1038/s41467-018-04387-2
    OpenUrl
  57. ↵
    Finn, E. S., Glerean, E., Khojandi, A. Y., Nielson, D., Molfese, P. J., Handwerker, D. A., & Bandettini, P. A. (2020). Idiosynchrony: from shared responses to individual differences during naturalistic neuroimaging. NeuroImage, 215, 116828. https://doi.org/10.1016/j.neuroimage.2020.116828
    OpenUrl
  58. ↵
    Fischl, B. (2012). FreeSurfer. NeuroImage, 62(2), 774–781. https://doi.org/10.1016/j.neuroimage.2012.01.021
    OpenUrlCrossRefPubMedWeb of Science
  59. ↵
    Fischl, B., Sereno, M. I., Tootell, R. B., & Dale, A. M. (1999). High-resolution intersubject averaging and a coordinate system for the cortical surface. Human Brain Mapping, 8(4), 272–284. https://doi.org/10.1002/(SICI)1097-0193(1999)8:4%3C272::AID-HBM10%3E3.0.CO;2-4
    OpenUrlCrossRefPubMedWeb of Science
  60. ↵
    Fonov, V. S., Evans, A. C., McKinstry, R. C., Almli, C. R., & Collins, D. L. (2009). Unbiased nonlinear average age-appropriate brain templates from birth to adulthood. NeuroImage, 47(Supplement 1), S102. https://doi.org/10.1016/S1053-8119(09)70884-5
    OpenUrlCrossRef
  61. ↵
    Forman, S. D., Cohen, J. D., & Fitzgerald, M. (1995). Improved assessment of significant activation in functional magnetic resonance imaging (fMRI): use of a cluster‐size threshold. Magnetic Resonance in Medicine, 33, 636–647. https://doi.org/10.1002/mrm.1910330508
    OpenUrlCrossRefPubMedWeb of Science
  62. ↵
    Friederici, A. D. (2011). The brain basis of language processing: from structure to function. Physiological Reviews, 91(4), 1357–1392. https://doi.org/10.1152/physrev.00006.2011
    OpenUrlCrossRefPubMed
  63. ↵
    Friederici, A. D. (2012). The cortical language circuit: from auditory perception to sentence comprehension. Trends in Cognitive Sciences, 16(5), 262–268. https://doi.org/10.1016/j.tics.2012.04.001
    OpenUrlCrossRefPubMedWeb of Science
  64. ↵
    Friedman, L., Glover, G. H., Krenz, D., Magnotta, V., & The FIRST BIRN. (2006). Reducing inter-scanner variability of activation in a multicenter fMRI study: role of smoothness equalization. NeuroImage, 32(4), 1656–1668. https://doi.org/10.1016/j.neuroimage.2006.03.062
    OpenUrlCrossRefPubMedWeb of Science
  65. ↵
    Glasser, M. F., Coalson, T. S., Robinson, E. C., Hacker, C. D., Harwell, J., Yacoub, E., Ugurbil, K., Andersson, J., Beckmann, C. F., Jenkinson, M., Smith, S. M., & Van Essen, D. C. (2016). A multi-modal parcellation of human cerebral cortex. Nature, 536(7615), 171–178. https://doi.org/10.1038/nature18933
    OpenUrlCrossRefPubMed
  66. ↵
    Goldberg, Y. (2017). Neural network methods for natural language processing. Synthesis Lectures on Human Language Technologies, 10(1), 1–309. https://doi.org/10.2200/S00762ED1V01Y201703HLT037
    OpenUrl
  67. ↵
    Goldstein, A., Zada, Z., Buchnik, E., Schain, M., Price, A., Aubrey, B., Nastase, S. A., Feder, A., Emanual D., Cohen, A., Jensen, A., Gazula, H., Choe, G., Roa, A., Kim, C., Casto, C., Lora, F., Flinker, A., Devore, S., Doyle, W., Dugan, P., Friedman, D., Hassidim, A., Brenner, M., Matias, Y., Norman, K. A., Devinsky, O., & Hasson, U. (2020). Thinking ahead: prediction in context as a keystone of language in humans and machines. bioRxiv. https://doi.org/10.1101/2020.12.02.403477
  68. ↵
    Gorgolewski, K., Burns, C. D., Madison, C., Clark, D., Halchenko, Y. O., Waskom, M. L., & Ghosh, S. S. (2011). Nipype: a flexible, lightweight and extensible neuroimaging data processing framework in python. Frontiers in Neuroinformatics, 5, 13. https://doi.org/10.3389/fninf.2011.00013
    OpenUrl
  69. ↵
    Gorgolewski, K. J., Alfaro-Almagro, F., Auer, T., Bellec, P., Capotă, M., Chakravarty, M. M., Churchill, N. W., Cohen, A. L., Craddock, R. C., Devenyi, G. A., Eklund, A., Esteban, O., Flandin, G., Ghosh, S. S., Guntupalli, J. S., Jenkinson, M., Keshavan, A., Kiar, G., Liem, F., Raamana, P. R., Raffelt, D., Steele, C. J., Quirion, P.-O., Smith, R. E., Strother, S. C., Varoquaux, G., Wang, Y., Yarkoni, T., & Poldrack, R. A. (2017). BIDS apps: Improving ease of use, accessibility, and reproducibility of neuroimaging data analysis methods. PLOS Computational Biology, 13(3), e1005209. https://doi.org/10.1371/journal.pcbi.1005209
    OpenUrl
  70. ↵
    Gorgolewski, K. J., Auer, T., Calhoun, V. D., Craddock, R. C., Das, S., Duff, E. P., Flandin, G., Ghosh, S. S., Glatard, T., Halchenko, Y. O., Handwerker, D. A., Hanke, M., Keator, D., Li, X., Michael, Z., Maumet, C., Nichols, B. N., Nichols, T. E., Pellman, J., Poline, J.-B., Rokem, A., Schaefer, G., Sochat, V., Triplett, W., Turner, J. A., Varoquaux, G., & Poldrack, R. A. (2016). The brain imaging data structure, a format for organizing and describing outputs of neuroimaging experiments. Scientific Data, 3, 160044. https://doi.org/10.1038/sdata.2016.44
    OpenUrl
  71. ↵
    Greve, D. N., & Fischl, B. (2009). Accurate and robust brain image alignment using boundary-based registration. NeuroImage, 48(1), 63–72. https://doi.org/10.1016/j.neuroimage.2009.06.060
    OpenUrlCrossRefPubMedWeb of Science
  72. ↵
    Gross, J., Hoogenboom, N., Thut, G., Schyns, P., Panzeri, S., Belin, P., & Garrod, S. (2013). Speech rhythms and multiplexed oscillatory sensory coding in the human brain. PLOS Biology, 11(12), e1001752. https://doi.org/10.1371/journal.pbio.1001752
    OpenUrlCrossRefPubMed
  73. ↵
    Gulban, O. F., Nielson, D., Poldrack, R., Lee, J., Gorgolewski, C., Vanessasaurus, & Ghosh, S. (2019). poldracklab/pydeface (Version 2.0.0). Zenodo. https://doi.org/10.5281/zenodo.3524401
  74. Guntupalli, J. S., Feilong, M., & Haxby, J. V. (2018). A computational model of shared fine-scale structure in the human connectome. PLOS Computational Biology, 14(4), e1006120. https://doi.org/10.1371/journal.pcbi.1006120
    OpenUrl
  75. Guntupalli, J. S., Hanke, M., Halchenko, Y. O., Connolly, A. C., Ramadge, P. J., & Haxby, J. V. (2016). A model of representational spaces in human cortex. Cerebral Cortex, 26(6), 2919–2934. https://doi.org/10.1093/cercor/bhw068
    OpenUrlCrossRefPubMed
  76. ↵
    Hagler, D. J., Jr., Saygin, A. P., & Sereno, M. I. (2006). Smoothing and cluster thresholding for cortical surface-based group analysis of fMRI data. NeuroImage, 33(4), 1093–1103. https://doi.org/10.1016/j.neuroimage.2006.07.036
    OpenUrlCrossRefPubMedWeb of Science
  77. ↵
    Halchenko, Y., Goncalves, M., di Oleggio Castello, M. V., Ghosh, S., Hanke, M., Salo, T., Dae, Kent, J., pvelasco, Amlien, I., Brett, M., Tilley, S., Markiewicz, C., Lukas, D. C., Gorgolewski, C., Kim, S., Stadler, J., Kaczmarzyk, J., Lee, J., Lurie, D., Pellman, J., Braun, H., Melo, B., Poldrack, B., Schiffler, B., Szczepanik, M., Samuels, K., Carlin, J., Feingold, F., & Kahn, A. (2020). nipy/heudiconv (Version 0.8.0). Zenodo. https://doi.org/10.5281/zenodo.3760062
  78. ↵
    Halchenko, Y. O., & Hanke, M. (2012). Open is not enough. Let’s take the next step: an integrated, community-driven computing platform for neuroscience. Frontiers in Neuroinformatics, 6, 22. https://doi.org/10.3389/fninf.2012.00022
    OpenUrl
  79. Hamilton, L. S., & Huth, A. G. (2020). The revolution will not be controlled: natural stimuli in speech neuroscience. Language, Cognition and Neuroscience, 35(5), 573–582. https://doi.org/10.1080/23273798.2018.1499946
    OpenUrl
  80. ↵
    Handwerker, D. A., Ollinger, J. M., & D’Esposito, M. (2004). Variation of BOLD hemodynamic responses across subjects and brain regions and their effects on statistical analyses. NeuroImage, 21(4), 1639–1651. https://doi.org/10.1016/j.neuroimage.2003.11.029
    OpenUrlCrossRefPubMedWeb of Science
  81. ↵
    Hanke, M., Adelhöfer, N., Kottke, D., Iacovella, V., Sengupta, A., Kaule, F. R., Nigbur, R., Waite, A. Q., Baumgartner, F., & Stadler, J. (2016). A studyforrest extension, simultaneous fMRI and eye gaze recordings during prolonged natural stimulation. Scientific Data, 3, 160092. https://doi.org/10.1038/sdata.2016.92
    OpenUrl
  82. ↵
    Hanke, M., Baumgartner, F. J., Ibe, P., Kaule, F. R., Pollmann, S., Speck, O., Zinke, W., & Stadler, J. (2014). A high-resolution 7-Tesla fMRI dataset from complex natural stimulation with an audio movie. Scientific Data, 1, 140003. https://doi.org/10.1038/sdata.2014.3
    OpenUrl
  83. ↵
    Hanke, M., & Halchenko, Y. O. (2011). Neuroscience runs on GNU/Linux. Frontiers in Neuroinformatics, 5, 8. https://doi.org/10.3389/fninf.2011.00008
    OpenUrl
  84. ↵
    Hanke, M., Halchenko, Y. O., Poldrack, B., Meyer, K., Solanky, D. S., Alteva, G., Gors, J., MacFarlane, D., Olaf Häusler, C., Olson, T., Waite, A., De La Vega, A., Sochat, V., Keshavan, A., Ma, F., Christian, H., Poelen, J., Skytén, K., Visconti di Oleggio Castello, M., Hardcastle, N., Stoeter, T., Lau, V. C., Markiewicz, C. J., & Wagner, A. S. (2020). datalad/datalad (Version 0.13.3). Zenodo. https://doi.org/10.5281/zenodo.4006562
  85. ↵
    Harris, C. R., Millman, K. J., van der Walt, S. J., Gommers, R., Virtanen, P., Cournapeau, D., Wieser, E., Taylor, J., Berg, S., Smith, N. J., Kern, R., Picus, M., Hoyer, S., van Kerkwijk, M. H., Brett, M., Haldane, A., Del Río, J. F., Wiebe, M., Peterson, P., Gérard-Marchant, P., Sheppard, K., Reddy, T., Weckesser, W., Abbasi, H., Gohlke, C., & Oliphant, T. E. (2020). Array programming with NumPy. Nature, 585(7825), 357–362. https://doi.org/10.1038/s41586-020-2649-2
    OpenUrlCrossRef
  86. ↵
    Hasson, U., Egidi, G., Marelli, M., & Willems, R. M. (2018). Grounding the neurobiology of language in first principles: the necessity of non-language-centric explanations for language comprehension. Cognition, 180, 135–157. https://doi.org/10.1016/j.cognition.2018.06.018
    OpenUrlCrossRefPubMed
  87. ↵
    Hasson, U., Ghazanfar, A. A., Galantucci, B., Garrod, S., & Keysers, C. (2012). Brain-to-brain coupling: a mechanism for creating and sharing a social world. Trends in Cognitive Sciences, 16(2), 114–121. https://doi.org/10.1016/j.tics.2011.12.007
    OpenUrlCrossRefPubMedWeb of Science
  88. ↵
    Hasson, U., & Honey, C. J. (2012). Future trends in neuroimaging: neural processes as expressed within real-life contexts. NeuroImage, 62(2), 1272–1278. https://doi.org/10.1016/j.neuroimage.2012.02.004
    OpenUrlCrossRefPubMedWeb of Science
  89. ↵
    Hasson, U., Malach, R., & Heeger, D. J. (2010). Reliability of cortical activity during natural stimulation. Trends in Cognitive Sciences, 14(1), 40–48. https://doi.org/10.1016/j.tics.2009.10.011
    OpenUrlCrossRefPubMedWeb of Science
  90. ↵
    Hasson, U., Nastase, S. A., & Goldstein, A. (2020). Direct fit to nature: an evolutionary perspective on biological and artificial neural networks. Neuron, 105(3), 416–434. https://doi.org/10.1016/j.neuron.2019.12.002
    OpenUrl
  91. ↵
    Hasson, U., Nir, Y., Levy, I., Fuhrmann, G., & Malach, R. (2004). Intersubject synchronization of cortical activity during natural vision. Science, 303(5664), 1634–1640. https://doi.org/10.1126/science.1089506
    OpenUrlAbstract/FREE Full Text
  92. ↵
    Haxby, J. V., Guntupalli, J. S., Connolly, A. C., Halchenko, Y. O., Conroy, B. R., Gobbini, M. I., Hanke, M., & Ramadge, P. J. (2011). A common, high-dimensional model of the representational space in human ventral temporal cortex. Neuron, 72(2), 404–416. https://doi.org/10.1016/j.neuron.2011.08.026
    OpenUrlCrossRefPubMedWeb of Science
  93. ↵
    Haxby, J. V., Guntupalli, J. S., Nastase, S. A., & Feilong, M. (2020). Hyperalignment: Modeling shared information encoded in idiosyncratic cortical topographies. eLife, 9, e56601. https://doi.org/10.7554/eLife.56601
    OpenUrlCrossRef
  94. ↵
    Heider, F., & Simmel, M. (1944). An experimental study of apparent behavior. American Journal of Psychology, 57(2), 243–259. https://doi.org/10.2307/1416950
    OpenUrlCrossRefWeb of Science
  95. ↵
    Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews Neuroscience, 8(5), 393–402. https://doi.org/10.1038/nrn2113
    OpenUrlCrossRefPubMedWeb of Science
  96. ↵
    Humphries, C., Binder, J. R., Medler, D. A., & Liebenthal, E. (2006). Syntactic and semantic modulation of neural activity during auditory sentence comprehension. Journal of Cognitive Neuroscience, 18(4), 665–679. https://doi.org/10.1162/jocn.2006.18.4.665
    OpenUrlCrossRefPubMedWeb of Science
  97. ↵
    Huntenburg, J. M. (2014). Evaluating nonlinear coregistration of BOLD EPI and T1w images [Master’s thesis, Freie Universität Berlin]. http://hdl.handle.net/11858/00-001M-0000-002B-1CB5-A
  98. ↵
    Hunter, J. D. (2007). Matplotlib: A 2D graphics environment. Computing in Science & Engineering, 9(3), 90–95. https://doi.org/10.1109/MCSE.2007.55
    OpenUrlCrossRef
  99. ↵
    Huth, A. G., de Heer, W. A., Griffiths, T. L., Theunissen, F. E., & Gallant, J. L. (2016). Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532(7600), 453–458. https://doi.org/10.1038/nature17637
    OpenUrlCrossRefPubMed
  100. ↵
    Iotzov, I., Fidali, B. C., Petroni, A., Conte, M. M., Schiff, N. D., & Parra, L. C. (2017). Divergent neural responses to narrative speech in disorders of consciousness. Annals of Clinical and Translational Neurology, 4(11), 784–792. https://doi.org/10.1002/acn3.470
    OpenUrlCrossRef
  101. ↵
    Jäncke, L., Wüstenberg, T., Scheich, H., & Heinze, H.-J. (2002). Phonetic perception and the temporal cortex. NeuroImage, 15(4), 733–746. https://doi.org/10.1006/nimg.2001.1027
    OpenUrlCrossRefPubMedWeb of Science
  102. ↵
    Jenkinson, M., Bannister, P., Brady, M., & Smith, S. (2002). Improved optimization for the robust and accurate linear registration and motion correction of brain images. NeuroImage, 17(2), 825–841. https://doi.org/10.1016/s1053-8119(02)91132-8
    OpenUrlCrossRefPubMedWeb of Science
  103. ↵
    Jenkinson, M., Beckmann, C. F., Behrens, T. E. J., Woolrich, M. W., & Smith, S. M. (2012). FSL. NeuroImage, 62(2), 782–790. https://doi.org/10.1016/j.neuroimage.2011.09.015
    OpenUrlCrossRefPubMedWeb of Science
  104. ↵
    Jones, E., Oliphant, T., & Peterson, P. (2001). SciPy: open source scientific tools for Python.
  105. ↵
    Just, M. A., Carpenter, P. A., Keller, T. A., Eddy, W. F., & Thulborn, K. R. (1996). Brain activation modulated by sentence comprehension. Science, 274(5284), 114–116. https://doi.org/10.1126/science.274.5284.114
    OpenUrlAbstract/FREE Full Text
  106. ↵
    Kell, A. J. E., Yamins, D. L. K., Shook, E. N., Norman-Haignere, S. V., & McDermott, J. H. (2018). A task-optimized neural network replicates human auditory behavior, predicts brain responses, and reveals a cortical processing hierarchy. Neuron, 98(3), 630–644.e16. https://doi.org/10.1016/j.neuron.2018.03.044
    OpenUrl
  107. ↵
    Kennedy, D. N., Haselgrove, C., Riehl, J., Preuss, N., & Buccigrossi, R. (2016). The NITRC image repository. NeuroImage, 124(Part B), 1069–1073. https://doi.org/10.1016/j.neuroimage.2015.05.074
    OpenUrl
  108. ↵
    Kim, D., Kay, K., Shulman, G. L., & Corbetta, M. (2018). A new modular brain organization of the BOLD signal during natural vision. Cerebral Cortex, 28(9), 3065–3081. https://doi.org/10.1093/cercor/bhx175
    OpenUrl
  109. ↵
    Klein, A., Ghosh, S. S., Bao, F. S., Giard, J., Häme, Y., Stavsky, E., Lee, N., Rossa, B., Reuter, M., Chaibub Neto, E., & Keshavan, A. (2017). Mindboggling morphometry of human brains. PLOS Computational Biology, 13(2), e1005350. https://doi.org/10.1371/journal.pcbi.1005350
    OpenUrl
  110. ↵
    Kleiner, M., Brainard, D., & Pelli, D. (2007). What’s new in Psychtoolbox-3? In Perception, 36(1), ECVP ‘07 Abstracts (p. 14). https://doi.org/10.1177/03010066070360S101
    OpenUrlCrossRefPubMed
  111. ↵
    1. F. Loizides &
    2. B. Schmidt
    Kluyver, T., Ragan-Kelley, B., Pérez, F., Granger, B., Bussonnier, M., Frederic, J., Kelley, K., Hamrick, J., Grout, J., Corlay, S., & Others. (2016). Jupyter Notebooks—a publishing format for reproducible computational workflows. In F. Loizides & B. Schmidt (Eds.), Positioning and Power in Academic Publishing: Players, Agents and Agendas (pp. 87–90). IOS Press. https://doi.org/10.3233/978-1-61499-649-1-87
  112. ↵
    Kriegeskorte, N., Mur, M., & Bandettini, P. A. (2008). Representational similarity analysis—connecting the branches of systems neuroscience. Frontiers in Systems Neuroscience, 2, 4. https://doi.org/10.3389/neuro.06.004.2008
    OpenUrl
  113. ↵
    Krizhevsky, A., Nair, V., & Hinton, G. (2009). Learning multiple layers of features from tiny images. Technical report, Department of Computer Science, University of Toronto, Canada. https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf
  114. ↵
    Krüger, G., & Glover, G. H. (2001). Physiological noise in oxygenation-sensitive magnetic resonance imaging. Magnetic Resonance in Medicine, 46(4), 631–637. https://doi.org/10.1002/mrm.1240
    OpenUrlCrossRefPubMedWeb of Science
  115. ↵
    Kumar, M., Anderson, M. J., Antony, J. W., Baldassano, C., Brooks, P. P., Cai, M. B., Chen, P.-H. C., Ellis, C. T., Henselman-Petrusek, G., Huberdeau, D., Hutchinson, J. B., Li, P. Y., Lu, Q., Manning, J. R., Mennen, A. C., Nastase, S. A., Richard, H., Schapiro, A. C., Schuck, N. W., Shvartsman, M., Sundaraman, N., Suo, D., Turek, J. S., Vo, V. A., Wallace, G., Wang, Y., Zhang, H., Zhu, X., Capota, M., Cohen, J. D., Hasson, U., Li, K., Ramadge, P. J., Turk-Browne, N. B., Willke, T. L., & Norman, K. A. (2020a). BrainIAK: The Brain Imaging Analysis Kit. OSF Preprints. https://doi.org/10.31219/osf.io/db2ev
  116. ↵
    Kumar, M., Ellis, C. T., Lu, Q., Zhang, H., Capota, M., Willke, T. L., Ramadge, P. J., Turk-Browne, N., & Norman, K. (2020b). BrainIAK tutorials: user-friendly learning materials for advanced fMRI analysis. PLOS Computational Biology, 16(1), e1007549. https://doi.org/10.1371/journal.pcbi.1007549
    OpenUrl
  117. ↵
    Kuperberg, G. R., McGuire, P. K., Bullmore, E. T., Brammer, M. J., Rabe-Hesketh, S., Wright, I. C., Lythgoe, D. J., Williams, S. C., & David, A. S. (2000). Common and distinct neural substrates for pragmatic, semantic, and syntactic processing of spoken sentences: an fMRI study. Journal of Cognitive Neuroscience, 12(2), 321–341. https://doi.org/10.1162/089892900562138
    OpenUrlCrossRefPubMedWeb of Science
  118. ↵
    Kurtzer, G. M., Sochat, V., & Bauer, M. W. (2017). Singularity: scientific containers for mobility of compute. PLOS One, 12(5), e0177459. https://doi.org/10.1371/journal.pone.0177459
    OpenUrlCrossRefPubMed
  119. ↵
    Kwong, K. K., Belliveau, J. W., Chesler, D. A., Goldberg, I. E., Weisskoff, R. M., Poncelet, B. P., Kennedy, D. N., Hoppel, B. E., Cohen, M. S., & Turner, R. (1992). Dynamic magnetic resonance imaging of human brain activity during primary sensory stimulation. Proceedings of the National Academy of Sciences of the United States of America, 89(12), 5675–5679. https://doi.org/10.1073/pnas.89.12.5675
    OpenUrlAbstract/FREE Full Text
  120. ↵
    Lanczos, C. (1964). Evaluation of noisy data. Journal of the Society for Industrial and Applied Mathematics Series B Numerical Analysis, 1(1), 76–85. https://doi.org/10.1137/0701007
    OpenUrlCrossRef
  121. ↵
    Lecun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278–2324. https://doi.org/10.1109/5.726791
    OpenUrlCrossRefWeb of Science
  122. ↵
    Lerner, Y., Honey, C. J., Katkov, M., & Hasson, U. (2014). Temporal scaling of neural responses to compressed and dilated natural speech. Journal of Neurophysiology, 111(12), 2433–2444. https://doi.org/10.1152/jn.00497.2013
    OpenUrlCrossRefPubMed
  123. ↵
    Lerner, Y., Honey, C. J., Silbert, L. J., & Hasson, U. (2011). Topographic mapping of a hierarchy of temporal receptive windows using a narrated story. Journal of Neuroscience, 31(8), 2906–2915. https://doi.org/10.1523/jneurosci.3684-10.2011
    OpenUrlAbstract/FREE Full Text
  124. ↵
    Lindquist, M. A., Geuter, S., Wager, T. D., & Caffo, B. S. (2019). Modular preprocessing pipelines can reintroduce artifacts into fMRI data. Human Brain Mapping, 40, 2358–2376. https://doi.org/10.1002/hbm.24528
    OpenUrlCrossRefPubMed
  125. ↵
    Lin, X., Sur, I., Nastase, S. A., Divakaran, A., Hasson, U., & Amer, M. R. (2019). Data-efficient mutual information neural estimator. arXiv. http://arxiv.org/abs/1905.03319
  126. ↵
    Liu, Y., Piazza, E. A., Simony, E., Shewokis, P. A., Onaral, B., Hasson, U., & Ayaz, H. (2017). Measuring speaker–listener neural coupling with functional near infrared spectroscopy. Scientific Reports, 7, 43293. https://doi.org/10.1038/srep43293
    OpenUrl
  127. ↵
    Li, X., Morgan, P. S., Ashburner, J., Smith, J., & Rorden, C. (2016). The first step for neuroimaging data analysis: DICOM to NIfTI conversion. Journal of Neuroscience Methods, 264, 47–56. https://doi.org/10.1016/j.jneumeth.2016.03.001
    OpenUrlCrossRefPubMed
  128. ↵
    Logothetis, N. K. (2008). What we can do and what we cannot do with fMRI. Nature, 453(7197), 869–878. https://www.nature.com/articles/nature06976
    OpenUrlCrossRefPubMedWeb of Science
  129. ↵
    Logothetis, N. K., Pauls, J., Augath, M., Trinath, T., & Oeltermann, A. (2001). Neurophysiological investigation of the basis of the fMRI signal. Nature, 412(6843), 150–157. https://doi.org/10.1038/35084005
    OpenUrlCrossRefPubMedWeb of Science
  130. ↵
    Lositsky, O., Chen, J., Toker, D., Honey, C. J., Shvartsman, M., Poppenk, J. L., Hasson, U., & Norman, K. A. (2016). Neural pattern change during encoding of a narrative predicts retrospective duration estimates. eLife, 5, e16070. https://doi.org/10.7554/eLife.16070
    OpenUrlCrossRefPubMed
  131. ↵
    Mahowald, K., & Fedorenko, E. (2016). Reliable individual-level neural markers of high-level language processing: A necessary precursor for relating neural variability to behavioral and genetic variability. NeuroImage, 139, 74–93. https://doi.org/10.1016/j.neuroimage.2016.05.073
    OpenUrlCrossRefPubMed
  132. ↵
    Manning, C. D., Clark, K., Hewitt, J., Khandelwal, U., & Levy, O. (2020). Emergent linguistic structure in artificial neural networks trained by self-supervision. Proceedings of the National Academy of Sciences of the United States of America, 117(48) 30046–30054. https://doi.org/10.1073/pnas.1907367117
    OpenUrlAbstract/FREE Full Text
  133. ↵
    Matusz, P. J., Dikker, S., Huth, A. G., & Perrodin, C. (2019). Are we ready for real-world neuroscience? Journal of Cognitive Neuroscience, 31(3), 327–338. https://doi.org/10.1162/jocn_e_01276
    OpenUrlCrossRefPubMed
  134. ↵
    McKinney, W. (2010). Data structures for statistical computing in Python. In Proceedings of the 9th Python in Science Conference (pp. 51–56). http://conference.scipy.org/proceedings/scipy2010/pdfs/mckinney.pdf
  135. ↵
    McNamara, Q., De La Vega, A., & Yarkoni, T. (2017). Developing a comprehensive framework for multimodal feature extraction. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 1567–1574). https://doi.org/10.1145/3097983.3098075
  136. ↵
    Meer, J. N. v. d., Breakspear, M., Chang, L. J., Sonkusare, S., & Cocchi, L. (2020). Movie viewing elicits rich and reliable brain state dynamics. Nature Communications, 11(1), 5004. https://doi.org/10.1038/s41467-020-18717-w
    OpenUrl
  137. ↵
    Mennes, M., Biswal, B. B., Castellanos, F. X., & Milham, M. P. (2013). Making data sharing work: the FCP/INDI experience. NeuroImage, 82, 683–691. https://doi.org/10.1016/j.neuroimage.2012.10.064
    OpenUrl
  138. ↵
    1. C. J. C. Burges,
    2. L. Bottou,
    3. M. Welling,
    4. Z. Ghahramani, &
    5. K. Q. Weinberger
    Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013). Distributed representations of words and phrases and their compositionality. In C. J. C. Burges, L. Bottou, M. Welling, Z. Ghahramani, & K. Q. Weinberger (Eds.), Advances in Neural Information Processing Systems (pp. 3111–3119). Curran Associates, Inc. https://proceedings.neurips.cc/paper/2013/file/9aa42b31882ec039965f3c4923ce901b-Paper.pdf
  139. ↵
    Milham, M. P., Craddock, R. C., Son, J. J., Fleischmann, M., Clucas, J., Xu, H., Koo, B., Krishnakumar, A., Biswal, B. B., Castellanos, F. X., Colcombe, S., Di Martino, A., Zuo, X.-N., & Klein, A. (2018). Assessment of the impact of shared brain imaging data on the scientific literature. Nature Communications, 9(1), 2818. https://doi.org/10.1038/s41467-018-04976-1
    OpenUrl
  140. ↵
    Milivojevic, B., Varadinov, M., Vicente Grabovetsky, A., Collin, S. H. P., & Doeller, C. F. (2016). Coding of event nodes and narrative context in the hippocampus. Journal of Neuroscience, 36(49), 12412–12424. https://doi.org/10.1523/jneurosci.2889-15.2016
    OpenUrlAbstract/FREE Full Text
  141. ↵
    Mills, K. (2016). HCP-MMP1.0 projected on fsaverage. Figshare. https://doi.org/10.6084/m9.figshare.3498446.v2
  142. ↵
    Mitchell, T. M., Shinkareva, S. V., Carlson, A., Chang, K.-M., Malave, V. L., Mason, R. A., & Just, M. A. (2008). Predicting human brain activity associated with the meanings of nouns. Science, 320(5880), 1191–1195. https://doi.org/10.1126/science.1152876
    OpenUrlAbstract/FREE Full Text
  143. ↵
    Muschelli, J., Nebel, M. B., Caffo, B. S., Barber, A. D., Pekar, J. J., & Mostofsky, S. H. (2014). Reduction of motion-related artifacts in resting state fMRI using aCompCor. NeuroImage, 96, 22–35. https://doi.org/10.1016/j.neuroimage.2014.03.028
    OpenUrlCrossRefPubMed
  144. ↵
    Naselaris, T., Kay, K. N., Nishimoto, S., & Gallant, J. L. (2011). Encoding and decoding in fMRI. NeuroImage, 56(2), 400–410. https://doi.org/10.1016/j.neuroimage.2010.07.073
    OpenUrlCrossRefPubMedWeb of Science
  145. ↵
    Nastase, S. A., Gazzola, V., Hasson, U., & Keysers, C. (2019a). Measuring shared responses across subjects using intersubject correlation. Social Cognitive and Affective Neuroscience, 14(6), 667–685. https://doi.org/10.1093/scan/nsz037
    OpenUrlCrossRef
  146. ↵
    Nastase, S. A., Goldstein, A., & Hasson, U. (2020a). Keep it real: rethinking the primacy of experimental control in cognitive neuroscience. NeuroImage, 222, 117254. https://doi.org/10.1016/j.neuroimage.2020.117254
    OpenUrl
  147. ↵
    Nastase, S. A., Liu, Y.-F., Hillman, H., Norman, K. A., & Hasson, U. (2020b). Leveraging shared connectivity to aggregate heterogeneous datasets into a common response space. NeuroImage, 217, 116865. https://doi.org/10.1016/j.neuroimage.2020.116865
    OpenUrl
  148. ↵
    Nastase, S. A., Liu, Y.-F., Hillman, H., Zadbood, A., Hasenfratz, L., Keshavarzian, N., Chen, J., Honey, C. J., Yeshurun, Y., Regev, M., Nguyen, M., Chang, C. H. C., Baldassano, C. B., Lositsky, O., Simony, E., Chow, M. A., Leong, Y. C., Brooks, P. P., Micciche, E., Choe, G., Goldstein, A., Vanderwal, T., Halchenko, Y. O., Norman, K. A., & Hasson, U. (2019b). Narratives: fMRI data for evaluating models of naturalistic language comprehension. OpenNeuro, ds002345. https://doi.org/10.18112/openneuro.ds002345.v1.1.4
  149. ↵
    Nguyen, M., Vanderwal, T., & Hasson, U. (2019). Shared understanding of narratives is correlated with shared neural responses. NeuroImage, 184, 161–170. https://doi.org/10.1016/j.neuroimage.2018.09.010
    OpenUrlCrossRef
  150. ↵
    Nichols, T. E., Das, S., Eickhoff, S. B., Evans, A. C., Glatard, T., Hanke, M., Kriegeskorte, N., Milham, M. P., Poldrack, R. A., Poline, J.-B., Proal, E., Thirion, B., Van Essen, D. C., White, T., & Yeo, B. T. T. (2017). Best practices in data analysis and sharing in neuroimaging using MRI. Nature Neuroscience, 20(3), 299–303. https://doi.org/10.1038/nn.4500
    OpenUrlCrossRefPubMed
  151. ↵
    Nili, H., Wingfield, C., Walther, A., Su, L., Marslen-Wilson, W., & Kriegeskorte, N. (2014). A toolbox for representational similarity analysis. PLOS Computational Biology, 10(4), e1003553. https://doi.org/10.1371/journal.pcbi.1003553
    OpenUrl
  152. ↵
    Ni, W., Constable, R. T., Mencl, W. E., Pugh, K. R., Fulbright, R. K., Shaywitz, S. E., Shaywitz, B. A., Gore, J. C., & Shankweiler, D. (2000). An event-related neuroimaging study distinguishing form and content in sentence processing. Journal of Cognitive Neuroscience, 12(1), 120–133. https://doi.org/10.1162/08989290051137648
    OpenUrlCrossRefPubMedWeb of Science
  153. ↵
    Noppeney, U., & Price, C. J. (2004). An FMRI study of syntactic adaptation. Journal of Cognitive Neuroscience, 16(4), 702–713. https://doi.org/10.1162/089892904323057399
    OpenUrlCrossRefPubMedWeb of Science
  154. ↵
    Obleser, J., Zimmermann, J., Van Meter, J., & Rauschecker, J. P. (2007). Multiple stages of auditory speech perception reflected in event-related FMRI. Cerebral Cortex, 17(10), 2251–2257. https://doi.org/10.1093/cercor/bhl133
    OpenUrlCrossRefPubMedWeb of Science
  155. ↵
    Ochshorn, R. M., & Hawkins, M. (2016). Gentle: a robust yet lenient forced aligner built on Kaldi. https://lowerquality.com/gentle/
  156. ↵
    Ogawa, S., Tank, D. W., Menon, R., Ellermann, J. M., Kim, S. G., Merkle, H., & Ugurbil, K. (1992). Intrinsic signal changes accompanying sensory stimulation: functional brain mapping with magnetic resonance imaging. Proceedings of the National Academy of Sciences of the United States of America, 89(13), 5951–5955. https://doi.org/10.1073/pnas.89.13.5951
    OpenUrlAbstract/FREE Full Text
  157. ↵
    Ojemann, J. G., Akbudak, E., Snyder, A. Z., McKinstry, R. C., Raichle, M. E., & Conturo, T. E. (1997). Anatomic localization and quantitative analysis of gradient refocused echo-planar fMRI susceptibility artifacts. NeuroImage, 6(3), 156–167. https://doi.org/10.1006/nimg.1997.0289
    OpenUrlCrossRefPubMedWeb of Science
  158. ↵
    Pajula, J., & Tohka, J. (2014). Effects of spatial smoothing on inter-subject correlation based analysis of FMRI. Magnetic Resonance Imaging, 32(9), 1114–1124. https://doi.org/10.1016/j.mri.2014.06.001
    OpenUrlCrossRefPubMed
  159. ↵
    Parkes, L., Fulcher, B., Yücel, M., & Fornito, A. (2018). An evaluation of the efficacy, reliability, and sensitivity of motion correction strategies for resting-state functional MRI. NeuroImage, 171, 415–436. https://doi.org/10.1016/j.neuroimage.2017.12.073
    OpenUrlCrossRef
  160. ↵
    Patterson, K., Nestor, P. J., & Rogers, T. T. (2007). Where do you know what you know? The representation of semantic knowledge in the human brain. Nature Reviews Neuroscience, 8(12), 976–987. https://doi.org/10.1038/nrn2277
    OpenUrlCrossRefPubMedWeb of Science
  161. ↵
    Peirce, J., Gray, J. R., Simpson, S., MacAskill, M., Höchenberger, R., Sogo, H., Kastman, E., & Lindeløv, J. K. (2019). PsychoPy2: Experiments in behavior made easy. Behavior Research Methods, 51(1), 195–203. https://doi.org/10.3758/s13428-018-01193-y
    OpenUrlCrossRefPubMed
  162. ↵
    Peirce, J. W. (2007). PsychoPy—Psychophysics software in Python. Journal of Neuroscience Methods, 162(1–2), 8–13. https://doi.org/10.1016/j.jneumeth.2006.11.017
    OpenUrlCrossRefPubMedWeb of Science
  163. ↵
    Peirce, J. W. (2009). Generating stimuli for neuroscience using PsychoPy. Frontiers in Neuroinformatics, 2, 10. https://doi.org/10.3389/neuro.11.010.2008
    OpenUrl
  164. ↵
    Pereira, F., Lou, B., Pritchett, B., Ritter, S., Gershman, S. J., Kanwisher, N., Botvinick, M., & Fedorenko, E. (2018). Toward a universal decoder of linguistic meaning from brain activation. Nature Communications, 9(1), 963. https://doi.org/10.1038/s41467-018-03068-4
    OpenUrl
  165. ↵
    Perez, F., & Granger, B. E. (2007). IPython: a system for interactive scientific computing. Computing in Science Engineering, 9(3), 21–29. https://doi.org/10.1109/MCSE.2007.53
    OpenUrlCrossRef
  166. ↵
    Petersen, S. E., Fox, P. T., Posner, M. I., Mintun, M., & Raichle, M. E. (1988). Positron emission tomographic studies of the cortical anatomy of single-word processing. Nature, 331(6157), 585–589. https://doi.org/10.1038/331585a0
    OpenUrlCrossRefPubMedWeb of Science
  167. ↵
    Poldrack, R. A., Baker, C. I., Durnez, J., Gorgolewski, K. J., Matthews, P. M., Munafò, M. R., Nichols, T. E., Poline, J.-B., Vul, E., & Yarkoni, T. (2017). Scanning the horizon: towards transparent and reproducible neuroimaging research. Nature Reviews Neuroscience, 18(2), 115–126. https://doi.org/10.1038/nrn.2016.167
    OpenUrlCrossRefPubMed
  168. ↵
    Poldrack, R. A., & Gorgolewski, K. J. (2014). Making big data open: data sharing in neuroimaging. Nature Neuroscience, 17(11), 1510–1517. https://doi.org/10.1038/nn.3818
    OpenUrlCrossRefPubMed
  169. ↵
    Poldrack, R. A., & Gorgolewski, K. J. (2015). OpenfMRI: open sharing of task fMRI data. NeuroImage, 144(Part B), 259–261. https://doi.org/10.1016/j.neuroimage.2015.05.073
    OpenUrl
  170. ↵
    Poldrack, R. A., Gorgolewski, K. J., & Varoquaux, G. (2019). Computational and informatic advances for reproducible data analysis in neuroimaging. Annual Review of Biomedical Data Science, 2, 119–138. https://doi.org/10.1146/annurev-biodatasci-072018-021237
    OpenUrl
  171. ↵
    Poldrack, R. A., Wagner, A. D., Prull, M. W., Desmond, J. E., Glover, G. H., & Gabrieli, J. D. (1999). Functional specialization for semantic and phonological processing in the left inferior prefrontal cortex. NeuroImage, 10(1), 15–35. https://doi.org/10.1006/nimg.1999.0441
    OpenUrlCrossRefPubMedWeb of Science
  172. ↵
    Povey, D., Ghoshal, A., Boulianne, G., Burget, L., Glembek, O., Goel, N., Hannemann, M., Motlicek, P., Qian, Y., Schwarz, P., Silovsky, J., Stemmer, G., & Vesely, K. (2011). The Kaldi speech recognition toolkit. In IEEE 2011 Workshop on Automatic Speech Recognition and Understanding. http://publications.idiap.ch/index.php/publications/show/2265
  173. ↵
    Power, J. D., Barnes, K. A., Snyder, A. Z., Schlaggar, B. L., & Petersen, S. E. (2012). Spurious but systematic correlations in functional connectivity MRI networks arise from subject motion. NeuroImage, 59(3), 2142–2154. https://doi.org/10.1016/j.neuroimage.2011.10.018
    OpenUrlCrossRefPubMedWeb of Science
  174. ↵
    Power, J. D., Mitra, A., Laumann, T. O., Snyder, A. Z., Schlaggar, B. L., & Petersen, S. E. (2014). Methods to detect, characterize, and remove motion artifact in resting state fMRI. NeuroImage, 84, 320–341. https://doi.org/10.1016/j.neuroimage.2013.08.048
    OpenUrlCrossRefPubMedWeb of Science
  175. ↵
    Price, C. J. (2010). The anatomy of language: a review of 100 fMRI studies published in 2009. Annals of the New York Academy of Sciences, 1191, 62–88. https://doi.org/10.1111/j.1749-6632.2010.05444.x
    OpenUrlCrossRefPubMedWeb of Science
  176. ↵
    Price, C. J. (2012). A review and synthesis of the first 20 years of PET and fMRI studies of heard speech, spoken language and reading. NeuroImage, 62(2), 816–847. https://doi.org/10.1016/j.neuroimage.2012.04.062
    OpenUrlCrossRefPubMedWeb of Science
  177. ↵
    Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. Technical report, OpenAI. https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
  178. ↵
    Regev, M., Honey, C. J., Simony, E., & Hasson, U. (2013). Selective and invariant neural responses to spoken and written narratives. Journal of Neuroscience, 33(40), 15978–15988. https://doi.org/10.1523/jneurosci.1580-13.2013
    OpenUrlAbstract/FREE Full Text
  179. ↵
    Regev, M., Simony, E., Lee, K., Tan, K. M., Chen, J., & Hasson, U. (2019). Propagation of information along the cortical hierarchy as a function of attention while reading and listening to stories. Cerebral Cortex, 29(10), 4017–4034. https://doi.org/10.1093/cercor/bhy282
    OpenUrl
  180. ↵
    Rocca, R., & Yarkoni, T. (2020). Putting psychology to the test: rethinking model evaluation through benchmarking and prediction. PsyArXiv. https://doi.org/10.31234/osf.io/e437b
  181. ↵
    Saad, Z. S., & Reynolds, R. C. (2012). SUMA. NeuroImage, 62(2), 768–773. https://doi.org/10.1016/j.neuroimage.2011.09.016
    OpenUrlCrossRefPubMed
  182. ↵
    Saad, Z. S., Reynolds, R. C., & Argall, B. (2004). SUMA: an interface for surface-based intra-and inter-subject analysis with AFNI. In IEEE International Symposium on Biomedical Imaging: From Nano to Macro (pp. 1510–1513). https://doi.org/10.1109/ISBI.2004.1398837
  183. ↵
    Santoro, R., Moerel, M., De Martino, F., Goebel, R., Ugurbil, K., Yacoub, E., & Formisano, E. (2014). Encoding of natural sounds at multiple spectral and temporal resolutions in the human auditory cortex. PLOS Computational Biology, 10(1), e1003412. https://doi.org/10.1371/journal.pcbi.1003412
    OpenUrl
  184. ↵
    Satterthwaite, T. D., Elliott, M. A., Gerraty, R. T., Ruparel, K., Loughead, J., Calkins, M. E., Eickhoff, S. B., Hakonarson, H., Gur, R. C., Gur, R. E., & Wolf, D. H. (2013). An improved framework for confound regression and filtering for control of motion artifact in the preprocessing of resting-state functional connectivity data. NeuroImage, 64, 240–256. https://doi.org/10.1016/j.neuroimage.2012.08.052
    OpenUrlCrossRefPubMedWeb of Science
  185. ↵
    Schrimpf, M., Blank, I., Tuckute, G., Kauf, C., Hosseini, E. A., Kanwisher, N., Tenenbaum, J., & Fedorenko, E. (2020). Artificial neural networks accurately predict language processing in the brain. bioRxiv. https://doi.org/10.1101/2020.06.26.174482
  186. ↵
    Scott, S. K., Blank, C. C., Rosen, S., & Wise, R. J. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain, 123(12), 2400–2406. https://doi.org/10.1093/brain/123.12.2400
    OpenUrlCrossRefPubMedWeb of Science
  187. ↵
    Shafto, M. A., Tyler, L. K., Dixon, M., Taylor, J. R., Rowe, J. B., Cusack, R., Calder, A. J., Marslen-Wilson, W. D., Duncan, J., Dalgleish, T., Henson, R. N., Brayne, C., Matthews, F. E., & Cam-CAN. (2014). The Cambridge Centre for Ageing and Neuroscience (Cam-CAN) study protocol: a cross-sectional, lifespan, multidisciplinary examination of healthy cognitive ageing. BMC Neurology, 14, 204. https://doi.org/10.1186/s12883-014-0204-1
    OpenUrl
  188. ↵
    Silver, N. C., & Dunlap, W. P. (1987). Averaging correlation coefficients: should Fisher’s z transformation be used? Journal of Applied Psychology, 72(1), 146–148. https://doi.org/10.1037/0021-9010.72.1.146
    OpenUrlCrossRefWeb of Science
  189. ↵
    Simony, E., & Chang, C. (2019). Analysis of stimulus-induced brain dynamics during naturalistic paradigms. NeuroImage, 216, 116461. https://doi.org/10.1016/j.neuroimage.2019.116461
    OpenUrl
  190. ↵
    Simony, E., Honey, C. J., Chen, J., Lositsky, O., Yeshurun, Y., Wiesel, A., & Hasson, U. (2016). Dynamic reconfiguration of the default mode network during narrative comprehension. Nature Communications, 7, 12141. https://doi.org/10.1038/ncomms12141
    OpenUrl
  191. ↵
    Smith, S. M., Jenkinson, M., Woolrich, M. W., Beckmann, C. F., Behrens, T. E. J., Johansen-Berg, H., Bannister, P. R., De Luca, M., Drobnjak, I., Flitney, D. E., Niazy, R. K., Saunders, J., Vickers, J., Zhang, Y., De Stefano, N., Brady, J. M., & Matthews, P. M. (2004). Advances in functional and structural MR image analysis and implementation as FSL. NeuroImage, 23(Supplement 1), S208–S219. https://doi.org/10.1016/j.neuroimage.2004.07.051
    OpenUrlCrossRefPubMedWeb of Science
  192. ↵
    Sonkusare, S., Breakspear, M., & Guo, C. (2019). Naturalistic stimuli in neuroscience: critically acclaimed. Trends in Cognitive Sciences, 23(8), 699–714. https://doi.org/10.1016/j.tics.2019.05.004
    OpenUrlCrossRef
  193. ↵
    Spiers, H. J., & Maguire, E. A. (2007). Decoding human brain activity during real-world experiences. Trends in Cognitive Sciences, 11(8), 356–365. https://doi.org/10.1016/j.tics.2007.06.002
    OpenUrlCrossRefPubMedWeb of Science
  194. ↵
    Stephens, G. J., Honey, C. J., & Hasson, U. (2013). A place for time: the spatiotemporal structure of neural dynamics during natural audition. Journal of Neurophysiology, 110(9), 2019–2026. https://doi.org/10.1152/jn.00268.2013
    OpenUrlCrossRefPubMed
  195. ↵
    Taylor, J. R., Williams, N., Cusack, R., Auer, T., Shafto, M. A., Dixon, M., Tyler, L. K., Cam-Can, & Henson, R. N. (2017). The Cambridge Centre for Ageing and Neuroscience (Cam-CAN) data repository: structural and functional MRI, MEG, and cognitive data from a cross-sectional adult lifespan sample. NeuroImage, 144(Part B), 262–269. https://doi.org/10.1016/j.neuroimage.2015.09.018
    OpenUrlCrossRef
  196. ↵
    Treiber, J. M., White, N. S., Steed, T. C., Bartsch, H., Holland, D., Farid, N., McDonald, C. R., Carter, B. S., Dale, A. M., & Chen, C. C. (2016). Characterization and correction of geometric distortions in 814 diffusion weighted images. PLOS One, 11(3), e0152472. https://doi.org/10.1371/journal.pone.0152472
    OpenUrlCrossRef
  197. ↵
    Triantafyllou, C., Hoge, R. D., & Wald, L. L. (2006). Effect of spatial smoothing on physiological noise in high-resolution fMRI. NeuroImage, 32(2), 551–557. https://doi.org/10.1016/j.neuroimage.2006.04.182
    OpenUrlCrossRefPubMedWeb of Science
  198. ↵
    Turney, P. D., Pantel, P., & Others. (2010). From frequency to meaning: vector space models of semantics. Journal of Artificial Intelligence Research, 37, 141–188. https://doi.org/10.1613/jair.2934
    OpenUrl
  199. ↵
    Tustison, N. J., Avants, B. B., Cook, P. A., Zheng, Y., Egan, A., Yushkevich, P. A., & Gee, J. C. (2010). N4ITK: improved N3 bias correction. IEEE Transactions on Medical Imaging, 29(6), 1310–1320. https://doi.org/10.1109/TMI.2010.2046908
    OpenUrlCrossRefPubMedWeb of Science
  200. ↵
    Vandenberghe, R., Nobre, A. C., & Price, C. J. (2002). The response of left temporal cortex to sentences. Journal of Cognitive Neuroscience, 14(4), 550–560. https://doi.org/10.1162/08989290260045800
    OpenUrlCrossRefPubMedWeb of Science
  201. ↵
    Vanderwal, T., Eilbott, J., Finn, E. S., Craddock, R. C., Turnbull, A., & Castellanos, F. X. (2017). Individual differences in functional connectivity during naturalistic viewing conditions. NeuroImage, 157, 521–530. https://doi.org/10.1016/j.neuroimage.2017.06.027
    OpenUrlCrossRef
  202. ↵
    van der Walt, S., Colbert, S. C., & Varoquaux, G. (2011). The NumPy array: a structure for efficient numerical computation. Computing in Science & Engineering, 13(2), 22–30. https://doi.org/10.1109/MCSE.2011.37
    OpenUrl
  203. ↵
    Van Essen, D. C., Smith, S. M., Barch, D. M., Behrens, T. E. J., Yacoub, E., Ugurbil, K., & WU-Minn HCP Consortium. (2013). The WU-Minn Human Connectome Project: an overview. NeuroImage, 80, 62–79. https://doi.org/10.1016/j.neuroimage.2013.05.041
    OpenUrlCrossRefPubMedWeb of Science
  204. ↵
    Van Uden, C. E., Nastase, S. A., Connolly, A. C., Feilong, M., Hansen, I., Gobbini, M. I., & Haxby, J. V. (2018). Modeling semantic encoding in a common neural representational space. Frontiers in Neuroscience, 12, 437. https://doi.org/10.3389/fnins.2018.00437
    OpenUrl
  205. ↵
    Varoquaux, G., & Poldrack, R. A. (2019). Predictive models avoid excessive reductionism in cognitive neuroimaging. Current Opinion in Neurobiology, 55, 1–6. https://doi.org/10.1016/j.conb.2018.11.002
    OpenUrl
  206. ↵
    Vigneau, M., Beaucousin, V., Hervé, P. Y., Duffau, H., Crivello, F., Houdé, O., Mazoyer, B., & Tzourio-Mazoyer, N. (2006). Meta-analyzing left hemisphere language areas: phonology, semantics, and sentence processing. NeuroImage, 30(4), 1414–1432. https://doi.org/10.1016/j.neuroimage.2005.11.002
    OpenUrlCrossRefPubMedWeb of Science
  207. ↵
    Virtanen, P., Gommers, R., Oliphant, T. E., Haberland, M., Reddy, T., Cournapeau, D., Burovski, E., Peterson, P., Weckesser, W., Bright, J., van der Walt, S. J., Brett, M., Wilson, J., Millman, K. J., Mayorov, N., Nelson, A. R. J., Jones, E., Kern, R., Larson, E., Carey, C. J., Polat, I., Feng, Y., Moore, E. W., VanderPlas, J., Laxalde, D., Perktold, J., Cimrman, R., Henriksen, I., Quintero, E. A., Harris, C. R., Archibald, A. M., Ribeiro, A. H., Pedregosa, F., van Mulbregt, P., and SciPy 1.0 Contributors. (2020). SciPy 1.0: fundamental algorithms for scientific computing in Python. Nature Methods, 17(3), 261–272. https://doi.org/10.1038/s41592-019-0686-2
    OpenUrl
  208. ↵
    Visconti di Oleggio Castello, M., Dobson, J. E., Sackett, T., Kodiweera, C., Haxby, J. V., Goncalves, M., Ghosh, S., & Halchenko, Y. O. (2020). ReproNim/reproin (Version 0.6.0). Zenodo. https://doi.org/10.5281/zenodo.3625000
  209. ↵
    Vodrahalli, K., Chen, P.-H., Liang, Y., Baldassano, C., Chen, J., Yong, E., Honey, C., Hasson, U., Ramadge, P., Norman, K. A., & Arora, S. (2017). Mapping between fMRI responses to movies and their natural language annotations. NeuroImage, 180(Part A), 223–231. https://doi.org/10.1016/j.neuroimage.2017.06.042
    OpenUrl
  210. ↵
    Vouloumanos, A., Kiehl, K. A., Werker, J. F., & Liddle, P. F. (2001). Detection of sounds in the auditory stream: event-related fMRI evidence for differential activation to speech and nonspeech. Journal of Cognitive Neuroscience, 13(7), 994–1005. https://doi.org/10.1162/089892901753165890
    OpenUrlCrossRefPubMedWeb of Science
  211. ↵
    Wagner, A. S., Waite, L. K., Meyer, K., Heckner, M. K., Kadelka, T., Reuter, N., Waite, A. Q., Poldrack, B., Markiewicz, C. J., Halchenko, Y. O., Vavra, P., Chormai, P., Poline, J.-B., Paas, L. K., Herholz, P., Mochalski, L. N., Kraljevic, N., Wiersch, L., Hutton, A., Pustina, D., Baagil, H. H., Glatard, T., Oliveira, S., Ippoliti, G., Mönch, C., Huijser, D., & Hanke, M. (2020). The DataLad Handbook. Zenodo. https://doi.org/10.5281/zenodo.3608611
  212. ↵
    Wang, S., Peterson, D. J., Gatenby, J. C., Li, W., Grabowski, T. J., & Madhyastha, T. M. (2017). Evaluation of field map and nonlinear registration methods for correction of susceptibility artifacts in diffusion MRI. Frontiers in Neuroinformatics, 11, 17. https://doi.org/10.3389/fninf.2017.00017
    OpenUrl
  213. ↵
    Wehbe, L., Murphy, B., Talukdar, P., Fyshe, A., Ramdas, A., & Mitchell, T. (2014). Simultaneously uncovering the patterns of brain regions involved in different story reading subprocesses. PLOS One, 9(11), e112575. https://doi.org/10.1371/journal.pone.0112575
    OpenUrlCrossRefPubMed
  214. ↵
    Wilkinson, M. D., Dumontier, M., Aalbersberg, I. J. J., Appleton, G., Axton, M., Baak, A., Blomberg, N., Boiten, J.-W., da Silva Santos, L. B., Bourne, P. E., Bouwman, J., Brookes, A. J., Clark, T., Crosas, M., Dillo, I., Dumon, O., Edmunds, S., Evelo, C. T., Finkers, R., Gonzalez-Beltran, A., Gray, A. J., Groth, P., Goble, C., Grethe, J., Heringa, J., ‘t Hoen, P. A., Hooft, R., Kuhn, T., Kok, R., Kok, J., Lusher, S. J., Martone, M. E., Mons, A., Packer, A. L., Persson, B., Rocca-Serra, P., Roos, M., van Schaik, R., Sansone, S. A., Schultes, E., Sengstag, T., Slater, T., Strawn, G., Swertz, M. A., Thompson, M., van der Lei, J., van Mulligan, E., Velterop, J., Waagmeester, A., Wittenburg, P., Wolstencroft, K., Zhao, J., & Mons, B. (2016). The FAIR guiding principles for scientific data management and stewardship. Scientific Data, 3, 160018. https://doi.org/10.1038/sdata.2016.18
    OpenUrl
  215. ↵
    Willems, R. M., Nastase, S. A., & Milivojevic, B. (2020). Narratives for neuroscience. Trends in Neurosciences, 43(5), 271–273. https://doi.org/10.1016/j.tins.2020.03.003
    OpenUrl
  216. ↵
    Wise, R., Chollet, F., Hadar, U., Friston, K., Hoffner, E., & Frackowiak, R. (1991). Distribution of cortical neural networks involved in word comprehension and word retrieval. Brain, 114(4), 1803–1817. https://doi.org/10.1093/brain/114.4.1803
    OpenUrlCrossRefPubMedWeb of Science
  217. ↵
    Yarkoni, T., & Westfall, J. (2017). Choosing prediction over explanation in psychology: lessons from machine learning. Perspectives on Psychological Science, 12(6), 1100–1122. https://doi.org/10.1177/1745691617693393
    OpenUrlCrossRef
  218. ↵
    Yeshurun, Y., Nguyen, M., & Hasson, U. (2017a). Amplification of local changes along the timescale processing hierarchy. Proceedings of the National Academy of Sciences of the United States of America, 114(35), 9475–9480. https://doi.org/10.1073/pnas.1701652114
    OpenUrlAbstract/FREE Full Text
  219. ↵
    Yeshurun, Y., Swanson, S., Simony, E., Chen, J., Lazaridi, C., Honey, C. J., & Hasson, U. (2017b). Same story, different story: the neural representation of interpretive frameworks. Psychological Science, 28(3), 307–319. https://doi.org/10.1177/0956797616682029
    OpenUrlCrossRefPubMed
  220. ↵
    1. Feitelson D.,
    2. Rudolph L., &
    3. Schwiegelshohn U.
    Yoo A.B., Jette M.A., Grondona M. (2003) SLURM: simple Linux utility for resource management. In Feitelson D., Rudolph L., & Schwiegelshohn U. (Eds.), Lecture Notes in Computer Science: Job Scheduling Strategies for Parallel Processing (pp. 44–60). Springer, Berlin, Heidelberg. https://doi.org/10.1007/10968987_3
  221. ↵
    Zadbood, A., Chen, J., Leong, Y. C., Norman, K. A., & Hasson, U. (2017). How we transmit memories to other brains: constructing shared neural representations via communication. Cerebral Cortex, 27(10), 4988–5000. https://doi.org/10.1093/cercor/bhx202
    OpenUrlCrossRefPubMed
  222. ↵
    Zatorre, R. J., Evans, A. C., Meyer, E., & Gjedde, A. (1992). Lateralization of phonetic and pitch discrimination in speech processing. Science, 256(5058), 846–849. https://doi.org/10.1126/science.1589767
    OpenUrlAbstract/FREE Full Text
  223. ↵
    Zhang, Y., Brady, M., & Smith, S. (2001). Segmentation of brain MR images through a hidden Markov random field model and the expectation-maximization algorithm. IEEE Transactions on Medical Imaging, 20(1), 45–57. https://doi.org/10.1109/42.906424
    OpenUrlCrossRefPubMedWeb of Science
  224. ↵
    Zuo, X., Honey, C. J., Barense, M. D., Crombie, D., Norman, K. A., Hasson, U., & Chen, J. (2020). Temporal integration of narrative information in a hippocampal amnesic patient. NeuroImage, 213, 116658. https://doi.org/10.1016/j.neuroimage.2020.116658
    OpenUrl
Back to top
PreviousNext
Posted December 23, 2020.
Download PDF
Data/Code
Email

Thank you for your interest in spreading the word about bioRxiv.

NOTE: Your email address is requested solely to identify you as the sender of this article.

Enter multiple addresses on separate lines or separate them with commas.
Narratives: fMRI data for evaluating models of naturalistic language comprehension
(Your Name) has forwarded a page to you from bioRxiv
(Your Name) thought you would like to see this page from the bioRxiv website.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Share
Narratives: fMRI data for evaluating models of naturalistic language comprehension
Samuel A. Nastase, Yun-Fei Liu, Hanna Hillman, Asieh Zadbood, Liat Hasenfratz, Neggin Keshavarzian, Janice Chen, Christopher J. Honey, Yaara Yeshurun, Mor Regev, Mai Nguyen, Claire H. C. Chang, Christopher Baldassano, Olga Lositsky, Erez Simony, Michael A. Chow, Yuan Chang Leong, Paula P. Brooks, Emily Micciche, Gina Choe, Ariel Goldstein, Tamara Vanderwal, Yaroslav O. Halchenko, Kenneth A. Norman, Uri Hasson
bioRxiv 2020.12.23.424091; doi: https://doi.org/10.1101/2020.12.23.424091
Digg logo Reddit logo Twitter logo CiteULike logo Facebook logo Google logo Mendeley logo
Citation Tools
Narratives: fMRI data for evaluating models of naturalistic language comprehension
Samuel A. Nastase, Yun-Fei Liu, Hanna Hillman, Asieh Zadbood, Liat Hasenfratz, Neggin Keshavarzian, Janice Chen, Christopher J. Honey, Yaara Yeshurun, Mor Regev, Mai Nguyen, Claire H. C. Chang, Christopher Baldassano, Olga Lositsky, Erez Simony, Michael A. Chow, Yuan Chang Leong, Paula P. Brooks, Emily Micciche, Gina Choe, Ariel Goldstein, Tamara Vanderwal, Yaroslav O. Halchenko, Kenneth A. Norman, Uri Hasson
bioRxiv 2020.12.23.424091; doi: https://doi.org/10.1101/2020.12.23.424091

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Subject Area

  • Neuroscience
Subject Areas
All Articles
  • Animal Behavior and Cognition (2516)
  • Biochemistry (4961)
  • Bioengineering (3458)
  • Bioinformatics (15159)
  • Biophysics (6876)
  • Cancer Biology (5372)
  • Cell Biology (7699)
  • Clinical Trials (138)
  • Developmental Biology (4514)
  • Ecology (7120)
  • Epidemiology (2059)
  • Evolutionary Biology (10202)
  • Genetics (7494)
  • Genomics (9761)
  • Immunology (4811)
  • Microbiology (13166)
  • Molecular Biology (5124)
  • Neuroscience (29332)
  • Paleontology (203)
  • Pathology (833)
  • Pharmacology and Toxicology (1459)
  • Physiology (2125)
  • Plant Biology (4728)
  • Scientific Communication and Education (1007)
  • Synthetic Biology (1336)
  • Systems Biology (4000)
  • Zoology (768)