Skip to main content
bioRxiv
  • Home
  • About
  • Submit
  • ALERTS / RSS
Advanced Search
New Results

High-level cognition during story listening is reflected in high-order dynamic correlations in neural activity patterns

Lucy L. W. Owen, Thomas H. Chang, View ORCID ProfileJeremy R. Manning
doi: https://doi.org/10.1101/763821
Lucy L. W. Owen
1Department of Psychological and Brain Sciences, Dartmouth College, Hanover, NH
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Thomas H. Chang
1Department of Psychological and Brain Sciences, Dartmouth College, Hanover, NH
2Amazon.com, Seattle, WA
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Jeremy R. Manning
1Department of Psychological and Brain Sciences, Dartmouth College, Hanover, NH
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Jeremy R. Manning
  • For correspondence: jeremy.r.manning@dartmouth.edu
  • Abstract
  • Full Text
  • Info/History
  • Metrics
  • Supplementary material
  • Data/Code
  • Preview PDF
Loading

Abstract

Our thoughts arise from coordinated patterns of interactions between brain structures that change with our ongoing experiences. High-order dynamic correlations in neural activity patterns reflect different subgraphs of the brain’s connectome that display homologous lower-level dynamic correlations. We tested the hypothesis that high-level cognition is supported by high-order dynamic correlations in brain activity patterns. We developed an approach to estimating high-order dynamic correlations in timeseries data, and we applied the approach to neuroimaging data collected as human participants either listened to a ten-minute story, listened to a temporally scrambled version of the story, or underwent a resting state scan. We trained across-participant pattern classifiers to decode (in held-out data) when in the session each neural activity snapshot was collected. We found that classifiers trained to decode from high-order dynamic correlations yielded the best performance on data collected as participants listened to the (unscrambled) story. By contrast, classifiers trained to decode data from scrambled versions of the story or during the resting state scan yielded the best performance when they were trained using first-order dynamic correlations or non-correlational activity patterns. We suggest that as our thoughts become more complex, they are supported by higher-order patterns of dynamic network interactions throughout the brain.

Introduction

A central goal in cognitive neuroscience is to elucidate the neural code: the mapping between (a) mental states or cognitive representations and (b) neural activity patterns. One means of testing models of the neural code is to ask how accurately that model is able to “translate” neural activity patterns into known (or hypothesized) mental states or cognitive representations (e.g., Haxby et al., 2001; Huth et al., 2016, 2012; Kamitani & Tong, 2005; Mitchell et al., 2008; Nishimoto et al., 2011; Norman et al., 2006; Pereira et al., 2018; Tong & Pratte, 2012). Training decoding models on different types of neural features (Fig. 1a) can also help to elucidate which specific aspects of neural activity patterns are informative about cognition– and, by extension, which types of neural activity patterns might comprise the neural code. For example, prior work has used region of interest analyses to estimate the anatomical locations of specific neural representations (e.g., Etzel et al., 2009), or to compare the relative contributions to the neural code of multivariate activity patterns versus dynamic correlations between neural activity patterns (e.g., Fong et al., 2019; Manning et al., 2018). An emerging theme in this literature is that cognition is mediated by dynamic interactions between brain structures (Bassett et al., 2006; Demertzi et al., 2019; Lurie et al., 2018; Mack et al., 2017; Preti et al., 2017; Solomon et al., 2019; Sporns & Honey, 2006; Turk-Browne, 2013; Zou et al., 2019).

Figure 1:
  • Download figure
  • Open in new tab
Figure 1: Neural patterns.

.a. A space of neural features. Within-brain analyses are carried out within a single brain, whereas across-brain analyses compare neural patterns across two or more individuals’ brains. Univariate analyses characterize the activities of individual units (e.g., nodes, small networks, hierarchies of networks, etc.), whereas multivariate analyses characterize the patterns of activities across units. Order 0 patterns involve individual nodes; order 1 patterns involve node-node interactions; order 2 (and higher) patterns relate to interactions between homologous networks. Each of these patterns may be static (e.g., averaging over time) or dynamic. b. Summarizing neural patterns. To efficiently compute with complex neural patterns, it can be useful to characterize the patterns using summary measures. Dimensionality reduction algorithms project the patterns onto lower-dimensional spaces whose dimensions reflect weighted combinations or non-linear transformations of the dimensions in the original space. Graph measures characterize each unit’s participation in its associated network.

Studies of the neural code to date have primarily focused on univariate or multivariate neural patterns (for review see Norman et al., 2006), or (more recently) on patterns of dynamic first-order correlations (i.e., interactions between pairs of brain structures; Demertzi et al., 2019; Fong et al., 2019; Lurie et al., 2018; Manning et al., 2018; Preti et al., 2017; Zou et al., 2019). We wondered what the future of this line of work might hold. For example, is the neural code mediated by higher-order interactions between brain structures (e.g., see Reimann et al., 2017)? Second-order correlations reflect homologous patterns of correlation. In other words, if the dynamic patterns of correlations between two regions, A and B, are similar to those between two other regions, C and D, this would be reflected in the second-order correlations between (A–B) and (C–D). In this way, second-order correlations identify similarities and differences between subgraphs of the brain’s connectome. Analogously, third-order correlations reflect homologies between second-order correlations– i.e., homologous patterns of homologous interactions between brain regions. More generally, higher-order correlations reflect homologies between patterns of lower-order correlations. We can then ask: which “orders” of interaction are most reflective of high-level cognitive processes?

Another central question pertains to the extent to which the neural code is carried by activity patterns that directly reflect ongoing cognition (e.g., following Haxby et al., 2001; Norman et al., 2006), versus the dynamic properties of the network structure itself, independent of specific activity patterns in any given set of regions (e.g., following Bassett et al., 2006). For example, graph measures such as centrality and degree (Bullmore & Sporns, 2009) may be used to estimate how a given brain structure is “communicating” with other structures, independently of the specific neural representations carried by those structures. If one considers a brain region’s position in the network (e.g., its eigenvector centrality) as a dynamic property, one can compare how the positions of different regions are correlated, and/or how those patterns of correlations change over time. We can also compute higher-order patterns in these correlations to characterize homologous subgraphs in the connectome that display similar changes in their constituent brain structures’ interactions with the rest of the brain.

To gain insights into the above aspects of the neural code, we developed a computational framework for estimating dynamic high-order correlations in timeseries data. This framework provides an important advance, in that it enables us to examine patterns of higher-order correlations that are computationally intractable to estimate via conventional methods. Given a multivariate timeseries, our framework provides timepoint-by-timepoint estimates of the first-order correlations, second-order correlations, and so on. Our approach combines a kernel-based method for computing dynamic correlations in timeseries data with a dimensionality reduction step (Fig. 1b) that projects the resulting dynamic correlations into a low-dimensional space. We explored two dimensionality reduction approaches: principle components analysis (PCA; Pearson, 1901), which preserves an approximately invertible transformation back to the original data (e.g., this follows related approaches taken by Gonzalez-Castillo et al., 2019; McIntosh & Jirsa, 2019; Toker & Sommer, 2019); and a second non-invertible algorithm that explored patterns in eigenvector centrality (Landau, 1895). This latter approach characterizes correlations between each feature dimension’s relative position in the network in favor of the specific activity histories of different features (also see Betzel et al., 2019; Reimann et al., 2017; Sizemore et al., 2018).

We validated our approach using synthetic data where the underlying correlations were known. We then applied our framework to a neuroimaging dataset collected as participants listened to either an audio recording of a ten-minute story, listened to a temporally scrambled version of the story, or underwent a resting state scan (Simony et al., 2016). We used a subset of the data to train across-participant classifiers to decode listening times using a blend of neural features (comprising neural activity patterns, as well as different orders of dynamic correlations between those patterns that were inferred using our computational framework). We found that both the PCA-based and eigenvector centrality-based approaches yielded neural patterns that could be used to decode accurately (i.e., well above chance). Both approaches also yielded the best decoding accuracy for data collected during (intact) story listening when high-order (PCA: secondorder; eigenvector centrality: fourth-order) dynamic correlation patterns were included as features. When we trained classifiers on the scrambled stories or resting state data, only lower-order dynamic patterns were informative to the decoders. Taken together, our results indicate that high-level cognition is supported by high-order dynamic patterns of communication between brain structures.

Methods

Our general approach to efficiently estimating high-order dynamic correlations comprises four general steps (Fig. 2). First, we derive a kernel-based approach to computing dynamic pairwise correlations in a T (timepoints) by K (features) multivariate timeseries, X0. This yields a T by O(K2) matrix of dynamic correlations, Y1, where each row comprises the upper triangle and diagonal of the correlation matrix at a single timepoint, reshaped into a row vector (this reshaped vector is Embedded Image -dimensional). Second, we apply a dimensionality reduction step to project the matrix of dynamic correlations back onto a K-dimensional space. This yields a T by K matrix, X1, that reflects an approximation of the dynamic correlations reflected in the original data. Third, we use repeated applications of the kernel-based dynamic correlation step to Xn and the dimensionality reduction step to the resulting Yn+1 to estimate high-order dynamic correlations. Each application of these steps to a T by K time series Xn yields a T by K matrix, Xn+1, that reflects the dynamic correlations between the columns of Xn. In this way, we refer to n as the order of the timeseries, where X0 (order 0) denotes the original data and Xn denotes (approximated) nth-order dynamic correlations between the columns of X0. Finally, we use a cross-validation–based decoding approach to evaluate how well information contained in a given order (or weighted mixture of orders) may be used to decode relevant cognitive states. If including a given Xn in the feature set yields higher classification accuracy on held-out data, we interpret this as evidence that the given cognitive states are reflected in patterns of nth-order correlations. All of the code used to produce the figures and results in this manuscript, along with links to the corresponding datasets, may be found at github.com/ContextLab/timecorr-paper. In addition, we have released a Python toolbox for computing dynamic high-order correlations in timeseries data; our toolbox may be found at timecorr.readthedocs.io.

Figure 2:
  • Download figure
  • Open in new tab
Figure 2: Estimating dynamic high-order correlations.

Given a T by K matrix of multivariate timeseries data, Xn (where n ∈ ℕ, n ≥ 0), we use Equation 4 to compute a timeseries of K by K correlation matrices, Yn+1. We then approximate Yn+1 with the T by K matrix Xn+1. This process may be repeated to scalably estimate iteratively higher-order correlations in the data. Note that the transposes of Xn and Xn+1 are displayed in the figure for compactness.

Kernel-based approach for computing dynamic correlations

Given a T by K matrix of observations, X, we can compute the (static) Pearson’s correlation between any pair of columns, X(·, i) and X(·, j) using (Pearson, 1901): Embedded Image Embedded Image Embedded Image

We can generalize this formula to compute time-varying correlations by incorporating a kernel function that takes a time t as input, and returns how much the observed data at each timepoint τ ∈ [−∞, ∞] contributes to the estimated instantaneous correlation at time t (Fig. 3; also see Allen et al., 2012, for a similar approach).

Figure 3:
  • Download figure
  • Open in new tab
Figure 3: Examples of kernel functions.

Each panel displays per-timepoint weights for a kernel centered at t = 50, evaluated at 100 timepoints (τ ∈ [1, …, 100]). a. Uniform kernel. The weights are timepoint-invariant; observations at all timepoints are weighted equally, and do not change as a function of τ. This is a special case kernel function that reduces dynamic correlations to static correlations. b. Dirac d kernel. Only the observation at timepoint t is given a non-zero weight (of 1). c. Gaussian kernels. Each kernel’s weights fall off in time according to a Gaussian probability density function centered on time t. Weights derived using several different example width parameters (σ2) are displayed. d. Laplace kernels. Each kernel’s weights fall off in time according to a Laplace probability density function centered on time t. Weights derived using several different example width parameters (b) are displayed. e. Mexican hat (Ricker wavelet) kernels. Each kernel’s weights fall off in time according to a Ricker wavelet centered on time t. This function highlights the contrasts between local versus surrounding activity patterns in estimating dynamic correlations. Weights derived using several different example width parameters (σ) are displayed.

Given a kernel function κt(·) for timepoint t, evaluated at timepoints τ ∈ [1, …, T], we can update the static correlation formula in Equation 1 to estimate the instantaneous correlation at timepoint t: Embedded Image Embedded Image Embedded Image

Here Embedded Image reflects the correlation at time t between columns i and j of X, estimated using the kernel κt. We evaluate Equation 4 in turn for each pair of columns in X and for kernels centered on each timepoint in the timeseries, respectively, to obtain a T by K by K timeseries of dynamic correlations, Y. For convenience, we then reshape the upper triangles and diagonals of each timepoint’s correlation matrix into a row vector to obtain an equivalent T by Embedded Image matrix.

Dynamic inter-subject functional connectivity (DISFC)

Equation 4 provides a means of taking a single observation matrix, Xn and estimating the dynamic correlations from moment to moment, Yn+1. Suppose that one has access to a set of multiple observation matrices that reflect the same phenomenon. For example, one might collect neuroimaging data from several experimental participants, as each participant performs the same task (or sequence of tasks). Let Embedded Image reflect the T by K observation matrices (n = 0) or reduced correlation matrices (n > 0) for each of P participants in an experiment. We can use inter-subject functional connectivity (ISFC; Simony et al., 2016) to compute the stimulus-driven correlations reflected in the multi-participant dataset at a given timepoint t using: Embedded Image where M extracts and vectorizes the upper triangle and diagonal of a symmetric matrix, Z is the Fisher z-transformation (Zar, 2010): Embedded Image

R is the inverse of Z: Embedded Image and Embedded Image denotes the correlation matrix at timepoint t (Eqn. 4) between each column of Embedded Image and each column of the average Xn from all other participants, Embedded Image: Embedded Image where \p denotes the set of all participants other than participant p. In this way, the T by Embedded Image DISFC matrix Embedded Image provides a time-varying extension of the ISFC approach developed by Simony et al. (2016).

Low-dimensional representations of dynamic correlations

Given a T by Embedded Image matrix of nth-order dynamic correlations, Yn, we propose two general approaches to computing a T by K low-dimensional representation of those correlations, Xn. The first approach uses dimensionality reduction algorithms to project Yn onto a K-dimensional space. The second approach uses graph measures to characterize the relative positions of each feature (k ∈ [1, …, K]) in the network defined by the correlation matrix at each timepoint.

Dimensionality reduction-based approaches to computing Xn

The modern toolkit of dimensionality reduction algorithms include Principal Components Analysis (PCA; Pearson, 1901), Probabilistic PCA (PPCA; Tipping & Bishop, 1999), Exploratory Factor Analysis (EFA; Spearman, 1904), Independent Components Analysis (ICA; Comon et al., 1991; Jutten & Herault, 1991), t-Stochastic Neighbor Embedding (t-SNE; van der Maaten & Hinton, 2008), Uniform Manifold Approximation and Projection (UMAP; McInnes & Healy, 2018), non-negative matrix factorization (NMF; Lee & Seung, 1999), Topographic Factor Analysis (TFA; Manning et al., 2014), Hierarchical Topographic Factor analysis (HTFA; Manning et al., 2018), Topographic Latent Source Analysis (TLSA; Gershman et al., 2011), dictionary learning (J. Mairal et al., 2009; J. B. Mairal et al., 2009), and deep auto-encoders (Hinton & Salakhutdinov, 2006), among others. While complete characterizations of each of these algorithms is beyond the scope of the present manuscript, the general intuition driving these approaches is to compute the T by K matrix, X, that is closest to the original T by J matrix, Y, where (typically) K " J. The different approaches place different constraints on what properties X must satisfy and which aspects of the data are compared (and how) in order to optimize how well X approximates Y.

Applying dimensionality reduction algorithms to Y yields an X whose columns reflect weighted combinations (or nonlinear transformations) of the original columns of Y. This has two main consequences. First, with each repeated dimensionality reduction, the resulting Xn has lower and lower fidelity (with respect to what the “true” Yn might have looked like without using dimensionality reduction to maintain scalability). In other words, computing Xn is a lossy operation. Second, whereas each column of Yn may be mapped directly onto specific pairs of columns of Xn−1, the columns of Xn reflect weighted combinations and/or nonlinear transformations of the columns of Yn. Many dimensionality reduction algorithms are invertible (or approximately invertible). However, attempting to map a given Xn back onto the original feature space of X0 will usually require 𝒪(TK2n) space and therefore becomes intractable as n or K grow large.

Graph measure approaches to computing Xn

The above dimensionality reduction approaches to approximating a given Yn with a lower-dimensional Xn preserve a (potentially recombined and transformed) mapping back to the original data in X0. We also explore graph measures that instead characterize each feature’s relative position in the broader network of interactions and connections. To illustrate the distinction between the two general approaches we explore, suppose a network comprises nodes A, B, and C. If A and B exhibit uncorrelated activity patterns, the functional connection (correlation) between them will be (by definition) close to 0. However, if A and B each interact with C in similar ways, we might attempt to capture those similarities using a measure that reflects how A and B interact with other members of the network.

In general, graph measures take as input a matrix of interactions (e.g., using the above notation, a K by K correlation matrix or binarized correlation matrix reconstituted from a single timepoint’s row of Y), and return as output a set of K measures describing how each node (feature) sits within that correlation matrix with respect to the rest of the population. Widely used measures include betweenness centrality (the proportion of shortest paths between each pair of nodes in the population that involves the given node in question; e.g., Barthélemy, 2004; Freeman, 1977; Geisberger et al., 2008; Newman, 2005; Opsahl et al., 2010); diversity and dissimilarity (characterizations of how differently connected a given node is from others in the population; e.g., Lin, 2009; Rao, 1982; Ricotta & Szeidl, 2006); eigenvector centrality and pagerank centrality (measures of how influential a given node is within the broader network; e.g., Bonacich, 2007; Halu et al., 2013; Lohmann et al., 2010; Newman, 2008); transfer entropy and flow coefficients (a measure of how much information is flowing from a given node to other nodes in the network; e.g., Honey et al., 2007; Schreiber, 2000); k-coreness centrality (a measure of the connectivity of a node within its local subgraph; e.g., Alvarez-Hamelin et al., 2005; Christakis & Fowler, 2010); within-module degree (a measure of how many connections a node has to its close neighbors in the network; e.g., Rubinov & Sporns, 2010); participation coefficient (a measure of the diversity of a node’s connections to different subgraphs in the network; e.g., Rubinov & Sporns, 2010); and subgraph centrality (a measure of a node’s participation in all of the network’s subgraphs; e.g., Estrada & Rodríguez-Velázquez, 2005); among others.

For a given graph measure, η: ℝK×K → ℝK, we can use η to tranform each row of Yn in a way that characterizes the corresponding graph properties of each column. This results in a new T by K matrix, Xn, that reflects how the features reflected in the columns of Xn−1 participate in the network during each timepoint (row).

Dynamic higher-order correlations

Because Xn has the same shape as the original data X0, approximating Yn with a lower-dimensional Xn enables us to estimate high-order dynamic correlations in a scalable way. Given a T by K input matrix, the output of Equation 4 requires 𝒪(TK2) space to store. Repeated applications of Equation 4 (i.e., computing dynamic correlations between the columns of the outputted dynamic correlation matrix) each require exponentially more space; in general the nth-order dynamic correlations of a T by K timeseries occupies 𝒪(TK2n) space. However, when we approximate or summarize the output of Equation 4 with a T by K matrix (as described above), it becomes feasible to compute even very high-order correlations in high-dimensional data. Specifically, approximating the nth-order dynamic correlations of a T by K timeseries requires only 𝒪(TK2) additional space– the same as would be required to compute first-order dynamic correlations. In other words, the space required to store n + 1 multivariate timeseries reflecting up to nth order correlations in the original data scales linearly with n using our approach (Fig. 2).

Data

We examined two types of data: synthetic data and human functional neuroimaging data. We constructed and leveraged the synthetic data to evaluate our general approach (for a related validation approach see Thompson et al., 2018). Specifically, we tested how well Equation 4 could be used to recover known dynamic correlations using different choices of kernel (κ; Fig. 3), for each of several synthetic datasets that exhibited different temporal properties. We applied our approach to a functional neuroimaging dataset to test the hypothesis that ongoing cognitive processing is reflected in high-order dynamic correlations. We used an across-participant classification test to estimate whether dynamic correlations of different orders contain information about which timepoint in a story participants were listening to.

Synthetic data

We constructed a total of 40 different multivariate timeseries, collectively reflecting a total of 4 qualitatively different patterns of dynamic correlations (i.e., 10 datasets reflecting each type of dynamic pattern). Each timeseries comprised 50 features (dimensions) that varied over 300 timepoints. The observations at each timepoint were drawn from a zero-mean multivariate Gaussian distribution with a covariance matrix defined for each timepoint as described below. We drew the observations at each timepoint independently from the draws at all other timepoints; in other words, for each observation st ∼ 𝒩(0, Σt) at timepoint t, p(st) = p(st|s\t).

Constant

We generated data with stable underlying correlations to evaluate how Equation 4 characterized correlation “dynamics” when the ground truth correlations were static. We constructed 10 multivariate timeseries whose observations were each drawn from a single (stable) Gaussian distribution. For each dataset (indexed by m), we constructed a random covariance matrix, Σm: Embedded Image Embedded Image i, j ∈ [1, 2, …, 50]. In other words, all of the observations (for each of the 300 timepoints) within each dataset were drawn from a multivariate Gaussian distribution with the same covariance matrix, and the 10 datasets each used a different covariance matrix.

Random

We generated a second set of 10 synthetic datasets whose observations at each timepoint were drawn from a Gaussian distribution with a new randomly constructed (using Eqn. 12) covariance matrix. Because each timepoint’s covariance matrix was drawn independently from the covariance matrices for all other timepoints, these datasets provided a test of reconstruction accuracy in the absence of any meaningful underlying temporal structure in the dynamic correlations underlying the data.

Ramping

We generated a third set of 10 synthetic datasets whose underlying correlations changed gradually over time. For each dataset, we constructed two anchor covariance matrices using Equation 12, Σstart and Σend. For each of the 300 timepoints in each dataset, we drew the observations from a multivariate Gaussian distribution whose covariance matrix at each timepoint t ∈ [0, …, 299] was given by Embedded Image

The gradually changing correlations underlying these datasets allow us to evaluate the recovery of dynamic correlations when each timepoint’s correlation matrix is unique (as in the random datasets), but where the correlation dynamics are structured.

Event

We generated a fourth set of 10 synthetic datasets whose underlying correlation matrices exhibited prolonged intervals of stability, interspersed with abrupt changes. For each dataset, we used Equation 12 to generate 5 random covariance matrices. We constructed a timeseries where each set of 60 consecutive samples was drawn from a Gaussian with the same covariance matrix. These datasets were intended to simulate a system that undergoes occasional abrupt state changes.

Functional neuroimaging data collected during story listening

We examined an fMRI dataset collected by Simony et al. (2016) that the authors have made publicly available at arks.princeton.edu/ark:/88435/dsp015d86p269k. The dataset comprises neuroimaging data collected as participants listened to an audio recording of a story (intact condition; 36 participants), listened to temporally scrambled recordings of the same story (17 participants in the paragraph-scrambled condition listened to the paragraphs in a randomized order and 36 in the word-scrambled condition listened to the words in a randomized order), or lay resting with their eyes open in the scanner (rest condition; 36 participants). Full neuroimaging details may be found in the original paper for which the data were collected (Simony et al., 2016).

Hierarchical topographic factor analysis (HTFA)

Following our prior related work, we used HTFA (Manning et al., 2018) to derive a compact representation of the neuroimaging data. In brief, this approach approximates the timeseries of voxel activations (44,415 voxels) using a much smaller number of radial basis function (RBF) nodes (in this case, 700 nodes, as determined by an optimization procedure described by Manning et al., 2018). This provides a convenient representation for examining full-brain network dynamics. All of the analyses we carried out on the neuroimaging dataset were performed in this lower-dimensional space. In other words, each participant’s data matrix, X0, was a number-of-timepoints by 700 matrix of HTFA-derived factor weights (where the row and column labels were matched across participants). Code for carrying out HTFA on fMRI data may be found as part of the BrainIAK toolbox (Capota et al., 2017), which may be downloaded at brainiak.org.

Temporal decoding

We sought to identify neural patterns that reflected participants’ ongoing cognitive processing of incoming stimulus information. As reviewed by Simony et al. (2016), one way of homing in on these stimulusdriven neural patterns is to compare activity patterns across individuals (e.g., using ISFC analyses). In particular, neural patterns will be similar across individuals to the extent that the neural patterns under consideration are stimulus-driven, and to the extent that the corresponding cognitive representations are reflected in similar spatial patterns across people. Following this logic, we used an across-participant temporal decoding test developed by Manning et al. (2018) to assess the degree to which different neural patterns reflected ongoing stimulus-driven cognitive processing across people. The approach entails using a subset of the data to train a classifier to decode stimulus timepoints (i.e., moments in the story participants listened to) from neural patterns. We use decoding (forward inference) accuracy on held-out data, from held-out participants, as a proxy for the extent to which the inputted neural patterns reflected stimulus-driven cognitive processing in a similar way across individuals.

Forward inference and decoding accuracy

We used an across-participant correlation-based classifier to decode which stimulus timepoint matched each timepoint’s neural pattern. We first divided the participants into two groups: a template group, 𝒢template, and a to-be-decoded group, 𝒢decode. We used Equation 7 to compute a DISFC matrix for each group (Embedded Image and Embedded Image, respectively). We then correlated the rows of Embedded Image and Embedded Image to form a number-of-timepoints by number-of-timepoints decoding matrix, Λ. In this way, the rows of Λ reflected timepoints from the template group, while the columns reflected timepoints from the to-be-decoded group.

We used Λ to assign temporal labels to each row Embedded Image using the row of Embedded Image with which it was most highly correlated. We then repeated this decoding procedure, but using 𝒢decode as the template group and 𝒢template as the to-be-decoded group. Given the true timepoint labels (for each group), we defined the decoding accuracy as the average proportion of correctly decoded timepoints, across both groups. We defined the relative decoding accuracy as the difference between the decoding accuracy and chance accuracy (i.e.,Embedded Image).

Feature weighting and testing

We sought to examine which types of neural features (i.e., activations, first-order dynamic correlations, and higher-order dynamic correlations) were informative to the temporal decoders. Using the notation above, these features correspond to X0, X1, X2, X3, and so on.

One challenge to fairly evaluating high-order correlations is that if the kernel used in Equation 4 is wider than a single timepoint, each repeated application of the equation will result in further temporal blur. Because our primary assessment metric is temporal decoding accuracy, this unfairly biases against detecting meaningful signal in higher-order correlations (relative to lower-order correlations). We attempted to mitigate temporal blur in estimating each Xn by using a Dirac d function kernel (which places all of its mass over a single timepoint; Fig. 3b) to compute each lower-order correlation (X1, X2, …, Xn−1). We then used a new (potentially wider, as described below) kernel to compute Xn from Xn−1. In this way, temporal blurring was applied only in the last step of computing Xn. We note that, because each Xn is a low-dimensional representation of the corresponding Yn, the higher-order correlations we estimated reflect true correlations in the data with lower-fidelity than estimates of lower-order correlations. Therefore, even after correcting for temporal blurring, our approach is still biased against finding meaningful signal in higher-order correlations.

After computing each X1, X2, …, Xn−1 for each participant, we divided participants into two equally sized groups (±1 for odd numbers of participants): 𝒢train and 𝒢test. We then further subdivided 𝒢train into 𝒢train1 and 𝒢train2. We then computed Λ (temporal correlation) matrices for each type of neural feature, using 𝒢train1 and 𝒢train2. This resulted in n + 1 Λ matrices (one for the original timeseries of neural activations, and one for each of n orders of dynamic correlations). Our objective was to find a set of weights for each of these Λ matrices such that the weighted average of the n + 1 matrices yielded the highest decoding accuracy. We used quasi-Newton gradient ascent (Nocedal & Wright, 2006), using decoding accuracy (for Embedded Image and Embedded Image) as the objective function to be maximized, to find an optimal set of training data-derived weights, ϕ0,1,…,n, where Embedded Image and where ϕi ≥ 0∀i ∈ [0, 1, …, n].

After estimating an optimal set of weights, we computed a new set of n + 1 Λ matrices correlating the DISFC patterns from 𝒢train and 𝒢test at each timepoint. We use the resulting decoding accuracy of 𝒢test timepoints (using the weights in ϕ0,1,…,n to average the Λ matrices) to estimate how informative the set of neural features containing up to nth order correlations were.

We used a permutation-based procedure to form stable estimates of decoding accuracy for each set of neural features. In particular, we computed the decoding accuracy for each of 10 random group assignments of 𝒢train and 𝒢test. We report the mean accuracy (along with 95% confidence intervals) for each set of neural features.

Identifying robust decoding results

The temporal decoding procedure we use to estimate which neural features support ongoing cognitive processing is governed by several parameters. In particular, Equation 4 requires defining a kernel function, which can take on different shapes and widths. For a fixed set of neural features, each of these parameters can yield different decoding accuracies. Further, the best decoding accuracy for a given timepoint may be reliably achieved by one set of parameters, whereas the best decoding accuracy for another timepoint might be reliably achieved by a different set of parameters, and the best decoding accuracy across all timepoints might be reliably achieved by still another different set of parameters. Rather than attempting to maximize decoding accuracy, we sought to discover the trends in the data that were robust to classifier parameters choices. Specifically, we sought to characterize how decoding accuracy varied (under different experimental conditions) as a function of which neural features were considered.

To identify decoding results that were robust to specific classifier parameter choices, we repeated our decoding analyses after substituting into Equation 4 each of a variety of kernel shapes and widths. We examined Gaussian (Fig. 3c), Laplace (Fig. 3d), and Mexican Hat (Fig. 3e) kernels, each with widths of 5, 10, 20, and 50 samples. We then report the average decoding accuracies across all of these parameter choices. This enabled us to (partially) factor out performance characteristics that were parameter-dependent, within the set of parameters we examined.

Reverse inference

The dynamic patterns we examined comprise high-dimensional correlation patterns at each timepoint. To help interpret the resulting patterns in the context of other studies, we created summary maps by computing the across-timepoint average pairwise correlations at each order of analysis (first order, second order, etc.). We selected the 10 strongest (absolute value) correlations at each order. Each correlation is between the dynamic activity patterns (or patterns of dynamic high-order correlations) measured at two RBF nodes (see Hierarchical Topographic Factor Analysis). Therefore, the 10 strongest correlations involved up to 20 RBF nodes. Each RBF defines a spatial function whose activations range from 0 to 1. We constructed a map of RBF components that denoted the endpoints of the 10 strongest correlations (we set each RBF to have a maximum value of 1). We then carried out a meta analysis using Neurosynth (Rubin et al., 2017) to identify the 10 terms most commonly associated with the given map. This resulted in a set of 10 terms associated with the average dynamic correlation patterns at each order.

Results

We sought to understand whether high-level cognition is supported by dynamic patterns of high-order correlations. To that end, we developed a computational framework for estimating the dynamics of stimulus-driven high-order correlations in multivariate timeseries data (see Dynamic inter-subject functional connectivity (DISFC) and Dynamic higher-order correlations). We evaluated the efficacy of this framework at recovering known patterns in several synthetic datasets (see Synthetic data). We then applied the framework to a public fMRI dataset collected as participants listened to an auditorally presented story, listened to a temporally scrambled version of the story, or underwent a resting state scan (see Functional neuroimaging data collected during story listening). We used the relative decoding accuracies of classifiers trained on different sets of neural features to estimate which types of features reflected ongoing cognitive processing.

Recovering known dynamic correlations from synthetic data

We generated synthetic datasets that differed in how the underlying correlations changed over time. For each dataset, we applied Equation 4 with a variety of kernel shapes and widths. We assessed how well the true underlying correlations at each timepoint matched the recovered correlations (Fig. 4). For every kernel and dataset we tested, our approach recovered the correlation dynamics we embedded into the data. However, the quality of these recoveries varied across different synthetic datasets in a kernel-dependent way.

Figure 4:
  • Download figure
  • Open in new tab
Figure 4: Recovering known dynamic correlations from synthetic data.

Each panel displays the average correlations between the vectorized upper triangles of the recovered correlation matrix at each timepoint and either the true underlying correlation at each timepoint or a reference correlation matrix. (The averages are taken across 10 different randomly generated synthetic datasets of the given category.) Error ribbons denote 95% confidence intervals (taken across datasets). Different colors denote different kernel shapes, and the shading within each color family denotes the kernel width parameter. For a complete description of each synthetic dataset, see Synthetic data. a. Constant correlations. These datasets have a stable (unchanging) underlying correlation matrix. b. Random correlations. These datasets are generated using a new independently drawn correlation matrix at each new timepoint. c. Ramping correlations. These datasets are generated by smoothly varying the underlying correlations between the randomly drawn correlation matrices at the first and last timepoints. The left panel displays the correlations between the recovered dynamic correlations and the underlying ground truth correlations. The middle panel compares the recovered correlations with the first timepoint’s correlation matrix. The right panel compares the recovered correlations with the last timepoint’s correlation matrix. d. Event-based correlations. These datasets are each generated using five randomly drawn correlation matrices that each remain stable for a fifth of the total timecourse. The left panel displays the correlations between the recovered dynamic correlations and the underlying ground truth correlations. The right panels compare the recovered correlations with the correlation matrices unique to each event.

In general, wide monotonic kernel shapes (Laplace, Gaussian), and wider kernels (within a shape), performed best when the correlations varied gradually from moment-to-moment (Figs. 4a, c, and d). In the extreme, as the rate of change in correlations approaches 0 (Fig. 4a), an infinitely wide kernel would exactly recover the Pearson’s correlation (e.g., compare Eqns. 1 and 4).

When the correlation dynamics were unstructured in time (Fig. 4b), a Dirac d kernel (infinitely narrow) performed best. This is because, when every timepoint’s correlations are independent of the correlations at every other timepoint, averaging data over time dilutes the available signal. Following a similar pattern, holding kernel shape fixed, narrower kernel parameters better recovered randomly varying correlations.

Cognitively relevant dynamic high-order correlations in fMRI data

We used across-participant temporal decoders to identify cognitively relevant neural patterns in fMRI data (see Forward inference and decoding accuracy). The dataset we examined (collected by Simony et al., 2016) comprised four experimental conditions that exposed participants to stimuli that varied systematically in how cognitively engaging they were. The intact experimental condition had participants listen to an audio recording of a 10-minute story. The paragraph-scrambled experimental condition had participants listen to a temporally scrambled version of the story, where the paragraphs occurred out of order (but where the same total set of paragraphs were presented over the full listening interval). All participants in this condition experienced the scrambled paragraphs in the same order. The word-scrambled experimental condition had participants listen to a temporally scrambled version of the story where the words in the story occurred in a random order. All participants in the word condition experienced the scrambled words in the same order. Finally, in a rest experimental condition, participants lay in the scanner with no overt stimulus, with their eyes open (blinking as needed). This dataset provided a convenient means of testing our hypothesis that different levels of cognitive processing and engagement are supported by different orders of brain activity dynamics.

In brief, we computed timeseries of dynamic high-order correlations that were similar across participants in each of two randomly assigned groups: a training group and a test group. We then trained classifiers on the training group’s data to match each sample from the test group with a stimulus timepoint. Each classifier comprised a weighted blend of neural patterns that reflected up to nth-order dynamic correlations (see Feature weighting and testing). We repeated this process for n ∈ {0, 1, 2, …, 10}. Our examinations of synthetic data suggested that none of the kernels we examined were “universal” in the sense of optimally recovering underlying correlations regardless of the temporal structure of those correlations. We found a similar pattern in the (real) fMRI data, whereby different kernels yielded different decoding accuracies, but no single kernel emerged as the clear “best.” In our analyses of neural data, we therefore averaged our decoding results over a variety of kernel shapes and widths in order to identify results that were robust to specific kernel parameters (see Identifying robust decoding results).

Our approach to estimating dynamic high-order correlations entails mapping the high-dimensional feature space of correlations (a T by O(K2) matrix) onto a lower-dimensional T by K matrix. We carried out two sets of analyses that differed in how this mapping was computed. The first set of analyses used PCA to find a low-dimensional embedding of the original dynamic correlation matrices (Fig. 5a,b). The second set of analyses characterized correlations in dynamics of each feature’s eigenvector centrality, but did not preserve the underlying activity dynamics (Fig. 5c,d).

Figure 5:
  • Download figure
  • Open in new tab
Figure 5: Across-participant decoding accuracy varies with correlation order and cognitive engagement.

a. Decoding accuracy as a function of order: PCA. Order (x-axis) refers to the maximum order of dynamic correlations that were available to the classifiers (see Feature weighting and testing). The reported across-participant decoding accuracies are averaged over all kernel shapes and widths (see Identifying robust decoding results). The y-values are displayed relative to chance accuracy Embedded Image. The error ribbons denote 95% confidence intervals across cross-validation folds (i.e., random assignments of participants to the training and test sets). The colors denote the experimental condition. Arrows denote sets of features that yielded reliably higher (upwards facing) or lower (downward facing) decoding accuracy than the mean of all other features (via a two-tailed test, thresholded at p < 0.05). Figure 6 displays additional comparisons between the decoding accuracies achieved using different sets of neural features. The circled values represent the maximum decoding accuracy within each experimental condition. b. Normalized decoding accuracy as a function of order: PCA. This panel displays the same results as Panel a, but here each curve has been normalized to have a maximum value of 1 and a minimum value of 0 (including the upper and lower bounds of the respective 95% confidence intervals). Panels a and b used PCA to project each high-dimensional pattern of dynamic correlations onto a lower-dimensional space. c. Decoding accuracy as a function of order: eigenvector centrality. This panel is in the same format as Panel a, but here eigenvector centrality has been used to project the high-dimensional patterns of dynamic correlations onto a lower-dimensional space. d. Normalized decoding accuracy as a function of order: eigenvector centrality. This panel is in the same format as Panel b, but here eigenvector centrality has been used to project the high-dimensional patterns of dynamic correlations onto a lower-dimensional space.

Both sets of temporal decoding analyses yielded qualitatively similar results for the auditory (non-rest) conditions of the experiment (Fig. 5: pink, green, and teal lines; Fig. 6: three leftmost columns). The highest decoding accuracy for participants who listened to the intact (unscrambled) story was achieved using high-order dynamic correlations (PCA: second-order; eigenvector-centrality: fourth-order). Scrambled versions of the story were best decoded by lower-order correlations (PCA/paragraph: first-order; PCA/word: order zero; eigenvector centrality/paragraph: order zero; eigenvector centrality/word: order zero). The two sets of analyses yielded different decoding results on resting state data (Fig. 5: purple lines; Fig. 6: rightmost column). We note that while the resting state times could be decoded reliably, the accuracies were only very slightly above chance. We speculate that the decoders might have picked up on attentional drift, boredom, or tiredness; we hypothesize that these all increased throughout the resting state scan. The decoders might be picking up on aspects of these loosely defined cognitive states that are common across individuals. The PCA-based approach achieved the highest resting state decoding accuracy using order zero features (non-correlational, activation-based), whereas the eigenvector centrality-based approach achieved the highest resting state decoding accuracy using second-order correlations. Taken together, these analyses indicate that high-level cognitive processing (while listening to the intact story) is reflected in the dynamics of high-order correlations in brain activity, whereas lower-level cognitive processing (while listening to scrambled versions of the story that lack rich meaning) is reflected in the dynamics of lower-order correlations and non-correlational activity dynamics. Further, these patterns are associated both with the underlying activity patterns (characterized using PCA) and also with the changing relative positions that different brain areas occupy in their associated networks (characterized using eigenvector centrality).

Figure 6:
  • Download figure
  • Open in new tab
Figure 6: Statistical summary of decoding accuracies for different neural features.

Each column displays decoding results for one experimental condition (intact, paragraph, word, and rest). We considered dynamic activity patterns (order 0) and dynamic correlations at different orders (order > 0). We used two-tailed t-tests to compare the distributions of decoding accuracies obtained using each pair of features. The distributions for each feature reflect the set of average decoding accuracies (across all kernel parameters), obtained for each random assignment of training and test groups. In the upper triangles of each map, warmer colors (positive t-values) indicate that the neural feature indicated in the given row yielded higher accuracy than the feature indicated in the given column. Cooler colors (negative t-values) indicate that the feature in the given row yielded lower decoding accuracy than the feature in the given column. The lower triangles of each map denote the corresponding p-values for the t-tests. The diagonal entries display the relative average optimized weight given to each type of feature, in a decoder that included all feature types (see Feature weighting and testing).

Having established that patterns of high-order correlations are informative to decoders, we next won-dered which specific networks of brain regions contributed most to these patterns. As a representative example, we selected the kernel parameters that yielded decoding accuracies that best matched the average accuracies across all of the kernel parameters we examined. Using Figure 5c as a template, the best-matching kernel was a Laplace kernel with a width of 50 (Fig. 3d). We used this kernel to compute a single K by K nth-order DISFC matrix for each experimental condition. We then used Neurosynth (Rubin et al., 2017) to compute the terms most highly associated with the most strongly correlated pairs of regions in each of these matrices (Fig. 7; see Reverse inference).

Figure 7:
  • Download figure
  • Open in new tab
Figure 7: Top terms associated with the endpoints of the strongest correlations.

Each color corresponds to one order of inter-subject functional correlations. The inflated brain plots display the locations of the endpoints of the 10 strongest (absolute value) correlations at each order, thresholded at 0.999, and projected onto the cortical surface (Combrisson et al., 2019). The lists of terms on the right display the top five Neurosynth terms (Rubin et al., 2017) decoded from the corresponding brain maps for each order. Each row displays data from a different experimental condition. Additional maps and their corresponding Neurosynth terms may be found in the Supplementary materials (intact: Fig. S1; paragraph: Fig. S2; word: Fig. S3; rest: Fig. S4).

For all of the story listening conditions (intact, paragraph, and word), we found that first- and second-order correlations were most strongly associated with auditory and speech processing areas. During intact story listening, third-order correlations reflected integration with visual areas, and fourth-order correlations reflected integration with areas associated with high-level cognition and cognitive control, such as the ventrolateral prefrontal cortex. However, during listening to temporally scrambled stories, these higher-order correlations instead involved interactions with additional regions associated with speech and semantic processing. By contrast, we found a much different set of patterns in the resting state data. First-order resting state correlations were most strongly associated with regions involved in counting and numerical understanding. Second-order resting state correlations were strongest in visual areas; third-order correlations were strongest in task-positive areas; and fourth-order correlations were strongest in regions associated with autobiographical and episodic memory. We carried out analogous analyses to create maps (and decode the top associated Neurosynth terms) for up to fifteenth-order correlations (Figs. S1, S2, S3, and S4). Of note, examining fifteenth-order correlations between 700 nodes using conventional methods would have required storing roughly Embedded Image floating point numbers– assuming single-precision (32 bits each), this would require roughly 32 times as many bits as there are molecules in the known universe! Although these fifteenth-order correlations do appear (visually) to have some well-formed structure, we provide this latter example primarily as a demonstration of the efficiency and scalability of our approach.

Discussion

We tested the hypothesis that high-level cognition is supported by high-order brain network dynamics (e.g., see Reimann et al., 2017; Solomon et al., 2019). We examined high-order network dynamics in functional neuroimaging data collected during a story listening experiment. When participants listened to an auditory recording of the story, participants exhibited similar high-order brain network dynamics. By contrast, when participants instead listened to temporally scrambled recordings of the story, only lower-order brain network dynamics were similar across participants. Our results indicate that higher orders of network interactions support higher-level aspects of cognitive processing (Fig. 8).

Figure 8:
  • Download figure
  • Open in new tab
Figure 8: Proposed high-order network dynamics underlying high-level cognition during story listening.

Higher orders of network interactions support higher-level aspects of cognitive processing. When tasks evoke richer, deeper, and/or higher-level processing, this is reflected in higher-order network interactions.

The notion that cognition is reflected in (and possibly mediated by) patterns of first-order network dynamics has been suggested by or proposed in myriad empirical studies and reviews (e.g., Chang & Glover, 2010; Demertzi et al., 2019; Fong et al., 2019; Gonzalez-Castillo et al., 2019; Liégeois et al., 2019; Lurie et al., 2018; Park et al., 2018; Preti et al., 2017; Roy et al., 2019; Turk-Browne, 2013; Zou et al., 2019). Our study extends this line of work by finding cognitively relevant higher-order network dynamics that reflect ongoing cognition. Our findings complement other work that uses graph theory and topology to characterize how brain networks reconfigure during cognition (e.g., Bassett et al., 2006; Betzel et al., 2019; McIntosh & Jirsa, 2019; Reimann et al., 2017; Sizemore et al., 2018; Toker & Sommer, 2019; Zheng et al., 2019).

An open question not addressed by our study pertains to how different structures integrate incoming information with different time constants. For example, one line of work suggests that the cortical surface comprises a structured map such that nearby brain structures process incoming information at similar timescales. Low-level sensory areas integrate information relatively quickly, whereas higher-level regions integrate information relatively slowly (Baldassano et al., 2017; Chien & Honey, 2019; Hasson et al., 2015, 2008; Honey et al., 2012; Lerner et al., 2014, 2011). Other related work in human and mouse brains indicates that the temporal response profile of a given brain structure may relate to how strongly connected that structure is with other brain areas (Fallon et al., 2019). Further study is needed to understand the role of temporal integration at different scales of network interaction, and across different anatomical structures.

Another potential limitation of our approach relates to recent work suggesting that the brain undergoes rapid state changes, for example across event boundaries (e.g., Baldassano et al., 2017). Shappell et al. (2019) used hidden semi-Markov models to estimate state-specific network dynamics (also see Vidaurre et al., 2018). Our general approach might be extended by considering putative state transitions. For example, rather than weighting all timepoints using a similar kernel (Eqn. 4), the kernel function could adapt on a timepoint-by-timepoint basis such that only timepoints determined to be in the same “state” were given non-zero weight.

Identifying high-order network dynamics associated with high-level cognition required several important methods advances. First, we used kernel-based dynamic correlations to extended the notion of (static) inter-subject functional connectivity (Simony et al., 2016) to a dynamic measure of inter-subject functional connectivity (DISFC) that does not rely on sliding windows, and that may be computed at individual timepoints. This allowed us to precisely characterize stimulus-evoked network dynamics that were similar across individuals. Second, we developed a computational framework for efficiently and scalably estimating high-order dynamic correlations. Our approach uses dimensionality reduction algorithms and graph measures to obtain low-dimensional embeddings of patterns of network dynamics. Third, we developed an analysis framework for identifying robust decoding results by carrying out our analyses using a range of parameter values and then identifying which results were robust to specific parameter choices.

Concluding remarks

The complex hierarchy of dynamic interactions that underlie our thoughts is perhaps the greatest mystery in modern science. Methods for characterizing the dynamics of high-order correlations in neural data provide a window into the neural basis of cognition. By showing that high-level cognition is reflected in high-order network dynamics, we have elucidated the next step on the path towards understanding the neural basis of cognition.

Author contributions

Concept: J.R.M. Implementation: T.H.C., L.L.W.O., and J.R.M. Analyses: L.L.W.O. and J.R.M. Writing: L.L.W.O. and J.R.M.

Acknowledgements

We acknowledge discussions with Luke Chang, Vassiki Chauhan, Hany Farid, Paxton Fitzpatrick, Andrew Heusser, Eshin Jolly, Aaron Lee, Qiang Liu, Matthijs van der Meer, Judith Mildner, Gina Notaro, Stephen Satterthwaite, Emily Whitaker, Weizhen Xie, and Kirsten Ziman. Our work was supported in part by NSF EPSCoR Award Number 1632738 to J.R.M. and by a sub-award of DARPA RAM Cooperative Agreement N66001-14-2-4-032 to J.R.M. The content is solely the responsibility of the authors and does not necessarily represent the official views of our supporting organizations.

Footnotes

  • https://github.com/ContextLab/timecorr-paper

  • https://timecorr.readthedocs.io

References

  1. ↵
    Allen, E. A., Damaraju, E., Plis, S. M., Erhardt, E. B., Eichele, T., & Calhoun, V. D. (2012). Tracking whole-brain connectivity dynamics in the resting state. Cerebral Cortex, 24(3), 663–676.
    OpenUrlCrossRefPubMed
  2. ↵
    Alvarez-Hamelin, I., Dall’Asta, L., Barrat, A., & Vespignani, A. (2005). k-corr decomposition: a tool for the visualiztion of large scale networks. arxiv:cs/0504107v2.
  3. ↵
    Baldassano, C., Chen, J., Zadbood, A., Pillow, J. W., Hasson, U., & Norman, K. A. (2017). Discovering event structure in continuous narrative perception and memory. Neuron, 95(3), 709–721.
    OpenUrl
  4. ↵
    Barthélemy, M. (2004). Betweenness centrality in large complex networks. European Physical Journal B, 38, 163–168.
    OpenUrl
  5. ↵
    Bassett, D., Meyer-Lindenberg, A., Achard, S., Duke, T., & Bullmore, E. (2006, December). Adaptive reconfiguration of fractal small-world human brain functional networks. Proceedings of the National Academy of Sciences, USA, 103(51), 19518–23.
    OpenUrlAbstract/FREE Full Text
  6. ↵
    Betzel, R. F., Byrge, L., Esfahlani, F. Z., & Kennedy, D. P. (2019). Temporal fluctuations in the brain’s modular architecture during movie-watching. bioRxiv, doi.org/10.1101/750919.
    OpenUrlAbstract/FREE Full Text
  7. ↵
    Bonacich, P. (2007). Some unique properties of eigenvector centrality. Social Networks, 29(4), 555–564.
    OpenUrlCrossRefWeb of Science
  8. ↵
    Bullmore, E., & Sporns, O. (2009). Complex brain networks: graph theoretical analysis of structural and functional systems. Nature Reviews Neuroscience, 10(3), 186–198.
    OpenUrlCrossRefPubMedWeb of Science
  9. ↵
    Capota, M., Turek, J., Chen, P.-H., Zhu, X., Manning, J. R., Sundaram, N., … Shin, Y. S. (2017). Brain imaging analysis kit. Retrieved from https://doi.org/10.5281/zenodo.59780
  10. ↵
    Chang, C., & Glover, G. H. (2010). Time-frequency dynamics of resting-state brain connectivity measured with fMRI. NeuroImage, 50, 81–98.
    OpenUrlCrossRefPubMedWeb of Science
  11. ↵
    Chien, H.-Y. S., & Honey, C. J. (2019). Constructing and forgetting temporal context in the human cerebral cortex. bioRxiv, doi.org/10.1101/761593.
    OpenUrlAbstract/FREE Full Text
  12. ↵
    Christakis, N. A., & Fowler, J. H. (2010). Social network sensors for early detection of contagious outbreaks. PLoS One, 5(9), e12948.
    OpenUrlCrossRefPubMed
  13. ↵
    Combrisson, E., Vallat, R., O’Reilly, C., Jas, M., Pascarella, A., l Saive, A., … Jerbi, K. (2019). Visbrain: a multi-purpose GPU-accelerated open-source suite for multimodal brain data visualization. Frontiers in Neuroinformatics, 13(14), 1–14.
    OpenUrl
  14. ↵
    Comon, P., Jutten, C., & Herault, J. (1991). Blind separation of sources, part II: Problems statement. Signal Processing, 24(1), 11 – 20.
    OpenUrl
  15. ↵
    Demertzi, A., Tagliazucchi, E., Dehaene, S., Deco, G., Barttfeld, P., Raimondo, F., … Sitt, J. D. (2019). Human consciousness is supported by dynamic complex patterns of brain signal coordination. Science Advances, 5(2), eaat7603.
    OpenUrlFREE Full Text
  16. ↵
    Estrada, E., & Rodríguez-Velázquez, J. A. (2005). Subraph centrality in complex networks. Physical Review E, 71(5), 056103.
    OpenUrl
  17. ↵
    Etzel, J. A., Gazzola, V., & Keysers, C. (2009). An introduction to anatomical ROI-based fMRI classification. Brain Research, 1282, 114–125.
    OpenUrlCrossRefPubMedWeb of Science
  18. ↵
    Fallon, J., Ward, P., Parkes, L., Oldham, S., Arnatkevičiūtė, A., Fornito, A., & Fulcher, B. D. (2019). Timescales of spontaneous activity fluctuations relate to structural connectivity in the brain. bioRxiv, doi.org/10.1101/655050.
    OpenUrlAbstract/FREE Full Text
  19. ↵
    Fong, A. H. C., Yoo, K., Rosenberg, M. D., Zhang, S., Li, C.-S. R., Scheinost, D., … Chun, M. M. (2019). Dynamic functional connectivity during task performance and rest predictsindividual differences in attention across studies. NeuroImage, 188, 14–25.
    OpenUrl
  20. ↵
    Freeman, L. C. (1977). A set of measures of centrality based on betweenness. Sociometry, 40(1), 35–41.
    OpenUrlCrossRefWeb of Science
  21. ↵
    Geisberger, R., Sanders, P., & Schultes, D. (2008). Better approximation of betweenness centrality. Proceedings of the meeting on Algorithm Engineering and Experiments, 90–100.
  22. ↵
    Gershman, S., Blei, D., Pereira, F., & Norman, K. (2011). A topographic latent source model for fMRI data. NeuroImage, 57, 89–100.
    OpenUrlCrossRefPubMed
  23. ↵
    Gonzalez-Castillo, J., Caballero-Gaudes, C., Topolski, N., Handwerker, D. A., Pereira, F., & Bandettini, P. A. (2019). Imaging the spontaneous flow of thought: distinct periods of cognition contribute to dynamic functional connectivity during test. NeuroImage, 202(116129).
  24. ↵
    Halu, A., Mondragón, R. J., Panzarasa, P., & Bianconi, G. (2013). Multiplex PageRank. PLoS One, 8(10), e78293.
    OpenUrl
  25. ↵
    Hasson, U., Chen, J., & Honey, C. J. (2015). Hierarchical process memory: memory as an integral component of information processing. Trends in Cognitive Science, 19(6), 304–315.
    OpenUrlCrossRefPubMed
  26. ↵
    Hasson, U., Yang, E., Vallines, I., Heeger, D. J., & Rubin, N. (2008). A hierarchy of temporal receptive windows in human cortex. Journal of Neuroscience, 28(10), 2539–2550. doi: 10.1523/JNEUROSCI.5487-07.2008
    OpenUrlAbstract/FREE Full Text
  27. ↵
    Haxby, J. V., Gobbini, M. I., Furey, M. L., Ishai, A., Schouten, J. L., & Pietrini, P. (2001). Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science, 293, 2425–2430.
    OpenUrlAbstract/FREE Full Text
  28. ↵
    Hinton, G. E., & Salakhutdinov, R. R. (2006). Reducing the dimensionality of data with neural networks. Science, 313(5786), 504–507.
    OpenUrlAbstract/FREE Full Text
  29. ↵
    Honey, C. J., Kötter, R., Breakspear, M., & Sporns, O. (2007). Network structure of cerebral cortex shapes functional connectivity on multiple time scales. Proceedings of the National Academy of Science USA, 104(24), 10240–10245.
    OpenUrlAbstract/FREE Full Text
  30. ↵
    Honey, C. J., Thesen, T., Donner, T. H., Silbert, L. J., Carlson, C. E., Devinsky, O., … Hasson, U. (2012). Slow cortical dynamics and the accumulation of information over long timescales. Neuron, 76, 423–434. doi: 10.1016/j.neuron.2012.08.011
    OpenUrlCrossRefPubMedWeb of Science
  31. ↵
    Huth, A. G., de Heer, W. A., Griffiths, T. L., Theunissen, F. E., & Gallant, J. L. (2016). Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532, 453–458.
    OpenUrlCrossRefPubMed
  32. ↵
    Huth, A. G., Nisimoto, S., Vu, A. T., & Gallant, J. L. (2012). A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron, 76(6), 1210–1224.
    OpenUrlCrossRefPubMedWeb of Science
  33. ↵
    Jutten, C., & Herault, J. (1991). Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture. Signal Processing, 24(1), 1–10.
    OpenUrlCrossRefWeb of Science
  34. ↵
    Kamitani, Y., & Tong, F. (2005). Decoding the visual and subjective contents of the human brain. Nature Neuroscience, 8, 679–685.
    OpenUrlCrossRefPubMedWeb of Science
  35. ↵
    Landau, E. (1895). Zur relativen Wertbemessung der Turnierresultate. Deutsches Wochenschach, 11, 366–369.
    OpenUrl
  36. ↵
    Lee, D. D., & Seung, H. S. (1999). Learning the parts of objects by non-negative matrix factorization. Nature, 401, 788–791.
    OpenUrlCrossRefPubMedWeb of Science
  37. ↵
    Lerner, Y., Honey, C. J., Katkov, M., & Hasson, U. (2014). Temporal scaling of neural responses to compressed and dilated natural speech. Journal of Neurophysiology, 111, 2433–2444.
    OpenUrlCrossRefPubMed
  38. ↵
    Lerner, Y., Honey, C. J., Silbert, L. J., & Hasson, U. (2011). Topographic mapping of a hierarchy of temporal receptive windows using a narrated story. Journal of Neuroscience, 31(8), 2906–2915. doi: 10.1523/JNEUROSCI.3684-10.2011
    OpenUrlAbstract/FREE Full Text
  39. ↵
    Liégeois, R., Li, J., Kong, R., Orban, C., De Ville, D. V., Ge, T., … Yeo, B. T. T. (2019). Resting brain dynamics at different timescales capture distinct aspects of human behavior. Nature Communications, 10(2317), doi.org/10.1038/s41467-019-10317-7.
    OpenUrlCrossRef
  40. ↵
    Lin, J. (2009). Divergence measures based on the Shannon entropy. IEEE Transactions on Information Theory, 37(1), 145–151.
    OpenUrl
  41. ↵
    Lohmann, G., Margulies, D. S., Horstmann, A., Pleger, B., Lepsien, J., Goldhahn, D., … Turner, R. (2010). Eigenvector centrality mapping for analyzing connectivity patterns in fMRI data of the human brain. PLoS One, 5(4), e10232.
    OpenUrlCrossRefPubMed
  42. ↵
    Lurie, D., Kessler, D., Bassett, D., Betzel, R., Breakspear, M., Keilholz, S., … Calhoun, V. (2018). On the nature of time-varying functional connectivity in resting fMRI. PsyArXiv, doi.org/10.31234/osf.io/xtzre.
    OpenUrlCrossRef
  43. ↵
    Mack, M. L., Preston, A. R., & Love, B. C. (2017). Medial prefrontal cortex compresses concept representations through learning. bioRxiv, doi.org/10.1101/178145.
    OpenUrlAbstract/FREE Full Text
  44. ↵
    Mairal, J., Ponce, J., Sapiro, G., Zisserman, A., & Bach, F. R. (2009). Supervised dictionary learning. Advances in Neural Information Processing Systems, 1033–1040.
  45. ↵
    Mairal, J. B., Bach, F., Ponce, J., & Sapiro, G. (2009). Online dictionary learning for sparse coding. Proceedings of the 26th annual international conference on machine learning, 689–696.
  46. ↵
    Manning, J. R., Ranganath, R., Norman, K. A., & Blei, D. M. (2014). Topographic factor analysis: a Bayesian model for inferring brain networks from neural data. PLoS One, 9(5), e94914.
    OpenUrl
  47. ↵
    Manning, J. R., Zhu, X., Willke, T. L., Ranganath, R., Stachenfeld, K., Hasson, U., … Norman, K. A. (2018). A probabilistic approach to discovering dynamic full-brain functional connectivity patterns. NeuroImage, 180, 243–252.
    OpenUrl
  48. ↵
    McInnes, L., & Healy, J. (2018). UMAP: Uniform manifold approximation and projection for dimension reduction. arXiv, 1802(03426).
  49. ↵
    McIntosh, A. R., & Jirsa, V. K. (2019). The hidden repertoire of brain dynamics and dysfunction. Network Neuroscience, doi.org/10.1162/netna00107.
    OpenUrlCrossRef
  50. ↵
    Mitchell, T., Shinkareva, S., Carlson, A., Chang, K., Malave, V., Mason, R., & Just, M. (2008). Predicting human brain activity associated with the meanings of nouns. Science, 320(5880), 1191.
    OpenUrlAbstract/FREE Full Text
  51. ↵
    Newman, M. E. J. (2005). A measure of betweenness centrality based on random walks. Social Networks, 27, 39–54.
    OpenUrlCrossRefWeb of Science
  52. ↵
    Newman, M. E. J. (2008). The mathematics of networks. The New Palgrave Encyclopedia of Economics, 2, 1–12.
    OpenUrl
  53. ↵
    Nishimoto, S., Vu, A., Naselaris, T., Benjamini, Y., Yu, B., & Gallant, J. (2011). Reconstructing visual experience from brain activity evoked by natural movies. Current Biology, 21, 1 – 6.
    OpenUrlCrossRefPubMedWeb of Science
  54. ↵
    Nocedal, J., & Wright, S. J. (2006). Numerical optimization. New York: Springer.
  55. ↵
    Norman, K. A., Newman, E., Detre, G., & Polyn, S. M. (2006). How inhibitory oscillations can train neural networks and punish competitors. Neural Computation, 18, 1577–1610.
    OpenUrlCrossRefPubMedWeb of Science
  56. ↵
    Opsahl, T., Agneessens, F., & Skvoretz, J. (2010). Node centrality in weighted networks: generalizing degree and shortest paths. Social Networks, 32, 245–251.
    OpenUrlCrossRefWeb of Science
  57. ↵
    Park, H.-J., Friston, K. J., Pae, C., Park, B., & Razi, A. (2018). Dynamic effective connectivity in resting state fMRI. NeuroImage, 180, 594–608.
    OpenUrlCrossRef
  58. ↵
    Pearson, K. (1901). On lines and planes of closest fit to systems of points in space. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science, 2, 559–572.
    OpenUrlCrossRef
  59. ↵
    Pereira, F., Lou, B., Pritchett, B., Ritter, S., Gershman, S. J., Kanwisher, N., … Fedorenko, E. (2018). Toward a universal decoder of linguistic meaning from brain activation. Nature Communications, 9(963), 1–13.
    OpenUrl
  60. ↵
    Preti, M. G., Bolton, T. A. W., & Van De Ville, D. (2017). The dynamic functional connectome: state-of-the-art and perspectives. NeuroImage, 160, 41–54.
    OpenUrl
  61. ↵
    Rao, C. R. (1982). Diversity and dissimilarity coeficients: a unified approach. Theoretical Population Biology, 21(1), 24–43.
    OpenUrlCrossRefWeb of Science
  62. ↵
    Reimann, M. W., Nolte, M., Scolamiero, M., Turner, K., Perin, R., Chindemi, G., … Markram, H. (2017). Cliques of neurons bound into cavities provide a missing link between structure and function. Frontiers in Computational Neuroscience, 11(48), 1–16.
    OpenUrl
  63. ↵
    Ricotta, C., & Szeidl, L. (2006). Towards a unifying approach to diversity measures: Bridging the gap between the Shannon entropy and Rao’s quadratic index. Theoretical Population Biology, 70(3), 237–243.
    OpenUrlCrossRefPubMedWeb of Science
  64. ↵
    Roy, D. S., Park, Y.-G., Ogawa, S. K., Cho, J. H., Choi, H., Kamensky, L., … Tonegawa, S. (2019). Brain-wide mapping of contextual fear memory engram ensembles supports the dispersed engram complex hypothesis. bioRxiv, doi.org/10.1101/668483.
    OpenUrlAbstract/FREE Full Text
  65. ↵
    Rubin, T. N., Kyoejo, O., Gorgolewski, K. J., Jones, M. N., Poldrack, R. A., & Yarkoni, T. (2017). Decoding brain activity using a large-scale probabilistic functional-anatomical atlas of human cognition. PLoS Computational Biology, 13(10), e1005649.
    OpenUrl
  66. ↵
    Rubinov, M., & Sporns, O. (2010). Complex network measures of brain connectivity: uses and interpretations. NeuroImage, 52, 1059–1069.
    OpenUrlCrossRefPubMedWeb of Science
  67. ↵
    Schreiber, T. (2000). Measuring information transfer. Physical Review Letters, 85(2), 461–464.
    OpenUrlCrossRefPubMedWeb of Science
  68. ↵
    Shappell, H., Caffo, B. S., Pekar, J. J., & Lindquist, M. A. (2019). Improved state change estimation in dynamic functional connectivity using hidden semi-Markov models. NeuroImage, 191, 243–257.
    OpenUrl
  69. ↵
    Simony, E., Honey, C. J., Chen, J., & Hasson, U. (2016). Uncovering stimulus-locked network dynamics during narrative comprehension. Nature Communications, 7(12141), 1–13.
    OpenUrl
  70. ↵
    Sizemore, A. E., Giusti, C., Kahn, A., Vettel, J. M., Betzel, R. F., & Bassett, D. S. (2018). Cliques and cavities in the human connectome. Journal of Computational Neuroscience, 44(1), 115–145.
    OpenUrl
  71. ↵
    Solomon, S. H., Medaglia, J. D., & Thompson-Schill, S. L. (2019). Implementing a concept network model. Behavior Research Methods, 51, 1717–1736.
    OpenUrl
  72. ↵
    Spearman, C. (1904). General intelligence, objectively determined and measured. Americal Journal of Psychology, 15, 201–292.
    OpenUrl
  73. ↵
    Sporns, O., & Honey, C. J. (2006). Small worlds inside big brains. Proceedings of the National Academy of Science USA, 103(51), 19219–19220.
    OpenUrlFREE Full Text
  74. ↵
    Thompson, W. H., Richter, C. G., Plavén-Sigray, P., & Fransson, P. (2018). Simulations to benchmark time-varying connectivity methods for fMRI. PLoS Computational Biology, 14(5), e1006196.
    OpenUrl
  75. ↵
    Tipping, M. E., & Bishop, C. M. (1999). Probabilistic principal component analysis. Journal of Royal Statistical Society, Series B, 61(3), 611–622.
    OpenUrl
  76. ↵
    Toker, D., & Sommer, F. T. (2019). Information integration in large brain networks. PLoS Computational Biology, 15(2), e1006807.
    OpenUrl
  77. ↵
    Tong, F., & Pratte, M. S. (2012). Decoding patterns of human brain activity. Annual Review of Psychology, 63, 483–509.
    OpenUrlCrossRefPubMedWeb of Science
  78. ↵
    Turk-Browne, N. B. (2013). Functional interactions as big data in the human brain. Science, 342, 580–584.
    OpenUrlAbstract/FREE Full Text
  79. ↵
    van der Maaten, L. J. P., & Hinton, G. E. (2008). Visualizing high-dimensional data using t-SNE. Journal of Machine Learning Research, 9, 2579–2605.
    OpenUrlCrossRef
  80. ↵
    Vidaurre, D., Abeysuriya, R., Becker, R., Quinn, A. J., Alfaro-Almagro, F., Smith, S. M., & Woolrich, M. W. (2018). Discovering dynamic brain neworks from big data in rest and task. NeuroImage, 180, 646–656.
    OpenUrl
  81. ↵
    Zar, J. H. (2010). Biostatistical analysis. Prentice-Hall/Pearson.
  82. ↵
    Zheng, M., Allard, A., Hagmann, P., & Serrano, M. A. (2019). Geometric renormalization unravels self-similarity of the multiscale human connectome. arxiv:1904.11793.
  83. ↵
    Zou, Y., Donner, R. V., Marwan, N., Donges, J. F., & Kurths, J. (2019). Complex network approaches to nonlinear time series analysis. Physics Reports, 787, 1–97.
    OpenUrl
Back to top
PreviousNext
Posted September 10, 2019.
Download PDF

Supplementary Material

Data/Code
Email

Thank you for your interest in spreading the word about bioRxiv.

NOTE: Your email address is requested solely to identify you as the sender of this article.

Enter multiple addresses on separate lines or separate them with commas.
High-level cognition during story listening is reflected in high-order dynamic correlations in neural activity patterns
(Your Name) has forwarded a page to you from bioRxiv
(Your Name) thought you would like to see this page from the bioRxiv website.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Share
High-level cognition during story listening is reflected in high-order dynamic correlations in neural activity patterns
Lucy L. W. Owen, Thomas H. Chang, Jeremy R. Manning
bioRxiv 763821; doi: https://doi.org/10.1101/763821
Digg logo Reddit logo Twitter logo Facebook logo Google logo LinkedIn logo Mendeley logo
Citation Tools
High-level cognition during story listening is reflected in high-order dynamic correlations in neural activity patterns
Lucy L. W. Owen, Thomas H. Chang, Jeremy R. Manning
bioRxiv 763821; doi: https://doi.org/10.1101/763821

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Subject Area

  • Neuroscience
Subject Areas
All Articles
  • Animal Behavior and Cognition (4095)
  • Biochemistry (8793)
  • Bioengineering (6495)
  • Bioinformatics (23409)
  • Biophysics (11772)
  • Cancer Biology (9173)
  • Cell Biology (13304)
  • Clinical Trials (138)
  • Developmental Biology (7426)
  • Ecology (11392)
  • Epidemiology (2066)
  • Evolutionary Biology (15128)
  • Genetics (10419)
  • Genomics (14030)
  • Immunology (9154)
  • Microbiology (22133)
  • Molecular Biology (8797)
  • Neuroscience (47472)
  • Paleontology (350)
  • Pathology (1424)
  • Pharmacology and Toxicology (2487)
  • Physiology (3712)
  • Plant Biology (8073)
  • Scientific Communication and Education (1434)
  • Synthetic Biology (2217)
  • Systems Biology (6023)
  • Zoology (1251)