Abstract
Recent advances in connectomic and neurophysiological tools make it possible to probe whole-brain mechanisms in the mouse that underlie cognition and behavior. Based on experimental data, we developed a large-scale model of the mouse brain for a cardinal cognitive function called working memory, the brain’s ability to internally hold and process information without sensory input. In the model, interregional connectivity is constrained by mesoscopic connectome data. The density of parvalbumin-expressing interneurons in the model varies systematically across the cortex. We found that the long-range cell type-specific targeting and density of cell classes define working memory representations. A core cortical subnetwork and the thalamus produce distributed persistent activity, and the network exhibits numerous attractor states. Novel cell type-specific graph theory measures predicted the activity patterns and core subnetwork. This work highlights the need for cell type-specific connectomics, and provides a theory and tools to interpret large-scale recordings of brain activity during cognition.
Introduction
In contrast to our substantial knowledge of local neural computation, such as orientation selectivity in the primary visual cortex or spatial map of grid cells in the medial entorhinal cortex, much less is understood about distributed processes in multiple interacting brain regions underlying cognition and behavior. This has recently begun to change, as advances in new technologies enable neuroscientists to probe neural activity at single-cell resolution and on a large-scale by electrical recording or calcium imaging of behaving animals (Jun et al. 2017; Steinmetz et al. 2019; Stringer et al. 2019; Musall et al. 2019; Steinmetz et al. 2021), ushering in a new era of neuroscience investigating distributed neural dynamics and brain functions (Wang 2022).
To be specific, consider a core cognitive function called working memory, the ability to temporally maintain information in mind without external stimulation (Baddeley 2012). Working memory has long been studied in neurophysiology using delay-dependent tasks, where stimulus-specific information must be stored in working memory across a short time period between a sensory input and a memory-guided behavioral response (Fuster and Alexander 1971; Funahashi et al. 1989; Goldman-Rakic 1995; Wang 2001). Delay-period mnemonic persistent neural activity has been observed in multiple brain regions, suggesting distributed working memory representation (Suzuki and Gottlieb 2013; Leavitt et al. 2017; Christophel et al. 2017; Xu 2017; Dotson et al. 2018). Connectome-based computational models of the macaque cortex found that working memory activity depends on interareal connectivity (Murray et al. 2017; Jaramillo et al. 2019), macroscopic gradients of synaptic excitation (Wang 2020; Mejias and Wang 2022) and dopamine modulation (Froudist-Walsh et al. 2021a).
Mnemonic neural activity during a delay period is also distributed in the mouse brain (Liu et al. 2014; Schmitt et al. 2017; Guo et al. 2017; Bolkan et al. 2017; Gilad et al. 2018). The new recording and imaging techniques as well as optogenetic methods for causal analysis (Yizhar et al. 2011), that are widely applicable to behaving mice, hold promise for elucidating the circuit mechanism of distributed brain functions in rodents. Recurrent synaptic excitation represents a neural basis for the maintenance of persistent neural firing (Goldman-Rakic 1995; D. J. Amit 1995; Wang 2021). In the monkey cortex, the number of spines (sites of excitatory synapses) per pyramidal cell increases along the cortical hierarchy, consistent with the idea that mnemonic persistent activity in association cortical areas including the prefrontal cortex is sustained by recurrent excitation stronger than in early sensory areas. Such a macroscopic gradient is lacking in the mouse cortex (Gilman et al. 2017), raising the possibility that the brain mechanism for distributed working memory representations may be fundamentally different in mice and monkeys.
In this paper we report a cortical mechanism of distributed working memory that does not depend on a gradient of synaptic excitation. We developed an anatomically-based model of the mouse brain for working memory, built on the recently available mesoscopic connectivity data of the mouse thalamocortical system (Oh et al. 2014; Gămănuţ et al. 2018; Harris et al. 2019; Kim et al. 2017). Our model is validated by capturing large-scale neural activity observed in recent mouse experiments (Guo et al. 2017; Gilad et al. 2018). Using this model, we found that a decreasing gradient of synaptic inhibition mediated by parvalbumin (PV) positive GABAergic cells (Kim et al. 2017; Fulcher et al. 2019; Wang 2020) shapes the distributed pattern of working memory representation.
A focus of this work is to examine whether anatomical connectivity can predict the emergent large-scale neural activity pattern underlying working memory. Interestingly, traditional graph-theory measures of inter-areal connections, which ignore cell types of projection targets, are uncorrelated with activity patterns. We propose new cell type-specific graph theory measures to overcome this problem, and differentiate contributions of cortical areas in terms of their distinct role in loading, maintaining, and reading out the content of working memory. Through computer-simulated perturbations akin to optogenetic inactivations, a core subnetwork was uncovered for the generation of persistent activity. This core subnetwork can be predicted based on the cell type-specific interareal connectivity, highlighting the necessity of knowing the cell type targets of interareal connections in order to relate anatomy with physiology and behavior. This work provides a computational and theoretical platform for cross-scale understanding of cognitive processes across the mouse cortex.
Results
A decreasing gradient of PV interneuron density from sensory to association cortex
Our large-scale circuit model of the mouse cortex uses inter-areal connectivity provided by anatomical data within the 43-area parcellation in the common coordinate framework v3 atlas (Oh et al. 2014) (Fig. 1A, Fig 1 - supplement 1A). The model is endowed with area-to-area variation of parvalbumin-expressing interneurons (PV) in the form of a gradient measured from the qBrain mapping platform (Fig 1 - supplement 1B) (Kim et al. 2017). The PV cell density (the number of PV cells per unit volume) is divided by the total neuron density, to give the PV cell fraction, which better reflects the expected amount of synaptic inhibition mediated by PV neurons (Fig. 1B-C, neuron density is shown in Fig 1 - supplement 1C). Cortical areas display a hierarchy defined by mesoscopic connectome data acquired using anterograde fluorescent tracers (Oh et al. 2014) (Fig. 1D-E). In Fig. 1F, the PV density is plotted as a function of the cortical hierarchy, which shows a moderate negative correlation between the two. Therefore, primary sensory areas have a higher density of PV interneurons than association areas, although the gradient of PV densities does not align perfectly with the cortical hierarchy.
A whole-mouse cortex model with a gradient of interneurons
In our model, each cortical area is described by a local circuit (Fig. 2A), using a mean-field reduction (Wong and Wang 2006) of a spiking neural network (Wang 2002). A minimal version of it has two neural pools selective for different stimuli and a shared inhibitory neural pool. The model makes the following assumptions. First, local inhibitory strength is proportional to PV interneuron density across the cortex. Second, the inter-areal long-range connection matrix is given by the anterograde tracing data (Oh et al. 2014; Knox et al. 2018; Wang et al. 2020). Third, targeting is biased onto inhibitory cells for top-down compared with bottom-up projections, therefore feedforward connections have a greater net excitatory effect than feedback connections (counterstream inhibitory bias, CIB) (Mejias and Wang 2022; Javadzadeh and Hofer 2022; Wang 2022).
Distributed working memory activity depends on the gradient of inhibitory neurons and the cortical hierarchy
We simulated the large-scale network to perform a simple visual delayed response task that requires one of two stimuli to be held in working memory. We shall first consider the case in which the strength of local recurrent excitation is insufficient to generate persistent activity when parcellated areas are disconnected from each other. Consequently, the observed distributed mnemonic representation must depend on long-range interareal excitatory connection loops. Later in the paper we will discuss the network model behavior when some local areas are capable of sustained persistent firing in isolation.
The main question is: when distributed persistent activity emerges after a transient visual input (a 500 ms current pulse to a selective excitatory population) is presented to the primary visual cortex (VISp), what determines the spatial pattern of working memory representation? After we remove the external stimulus, the firing rate in area VISp decreases rapidly to baseline. Neural activity propagates throughout the cortex after stimulus offset (Fig. 2B). Neural activities in the higher visual cortical areas (e.g. VISrl and VISpl) show similar dynamics to VISp. In stark contrast, many frontal and lateral areas (including prelimbic (PL), infralimbic (ILA), secondary motor (MOs) and ventral agranular insula (AIv) areas) sustained a high firing rate during the delay period (Fig. 2B). Areas that are higher in the cortical hierarchy show elevated activity during the delay period (Fig. 2C). This persistent firing rate could last for more than 10 seconds and is a stable attractor state of the network (Inagaki et al. 2019).
The cortical hierarchy and PV fraction predict the delay period firing rate of each cortical area (Fig. 2C-E). Thus the activity pattern of distributed working memory depends on both local and large-scale anatomy. The delay activity pattern has a stronger correlation with hierarchy (r = 0.91) than with the PV fraction (r = −0.43). The long-range connections thus play a predominant important role in defining the persistent activity pattern.
Activity in early sensory areas such as VISp displays a rigorous response to the transient input but returns to a low firing state after stimulus withdrawal. In contrast, many frontal areas show strong persistent activity. When the delay period firing rates are plotted versus hierarchy, we observe a gap in the distribution of persistent activity (Fig. 2D) that marks an abrupt transition in the cortical space. This leads to the emergence of a subnetwork of areas capable of working memory representations.
We also used our circuit model to simulate delayed response tasks with different sensory modalities (Fig 2 - supplement 1), by stimulating primary somatosensory area SSp-bfd and primary auditory area AUDp. The pattern of delay period firing rates for these sensory modalities is similar to the results obtained for the visual task: sensory areas show transient activity, while frontal and lateral areas show persistent activity after stimulus withdrawal. Moreover, the cortical hierarchy could predict the delay period firing rate of each cortical area well (r = 0.89, p < 0.05). Our model thus predicts that working memory may share common activation patterns across sensory modalities, which is partially supported by cortical recordings during a memory-guided response task (Inagaki et al. 2018).
Thalamocortical interactions maintain distributed persistent activity
To investigate how thalamocortical interactions affect the large-scale network dynamics, we designed a thalamocortical network similar to the cortical network (Fig. 3A). Several studies have shown that thalamic areas are also involved in the maintenance of working memory (Bolkan et al. 2017; Guo et al. 2017; Schmitt et al. 2017). However, the large-scale thalamocortical mechanisms underlying memory maintenance are unknown. We set the strength of connections between the thalamus and cortex using data from the Allen Institute (Oh et al. 2014) (Fig 3 - supplement 1). All thalamocortical connections in the model are mediated by AMPA synapses. There are no recurrent connections in the thalamus within or across thalamic nuclei (Jones 2007). The effect of thalamic reticular nucleus neurons was included indirectly as a constant inhibitory current to all thalamic areas. Similarly to cortical areas, the thalamus is organized along a measured hierarchy (Harris et al. 2019). For example, the dorsal part of the lateral geniculate nucleus (LGd) is lower than the cortical area VISp in the hierarchy, consistent with the fact that LGd sends feedforward inputs to VISp. Thalamocortical projections in the model are slightly more biased toward excitatory neurons in the target area if they are feedforward projections and towards inhibitory neurons if they are feedback.
Here, we weakened the strength of cortical interareal connections as compared to the cortex model of Fig. 2. Now, persistent activity can still be generated (Fig. 3B, blue) but is maintained with the help of the thalamocortical loop, as observed experimentally (Guo et al. 2017). Indeed, in simulations where the thalamus was inactivated, the cortical network no longer showed sustained activity (Fig. 3B, red).
In the thalamocortical model, the delay activity pattern of the cortical areas correlates with the hierarchy, again with a gap in the firing rate separating the areas engaged in persistent activity from those that do not (Fig. 3B, Fig. 3C). Sensory areas show a low delay firing rate, and frontal areas show strong persistent firing. Unlike the cortex, the firing rate of thalamic areas continuously increases along the hierarchy (Fig. 3E). On the other hand, cortical dynamics in the thalamocortical and cortical models show many similarities. Early sensory areas do not show persistent activity in either model. Many frontal and lateral areas show persistent activity and there is an abrupt transition in cortical space in the thalamocortical model, like in the cortex only model. Quantitatively, the delay firing pattern of the cortical areas is correlated with the hierarchy and the PV fraction (Fig. 3C, Fig. 3D). Furthermore, the delay period firing rate of cortical areas in the thalamocortical model correlates well with the firing rate of the same areas in the cortical model (Fig. 3F). This comparison suggests that the cortical model captures most of the dynamical properties in the thalamocortical model; therefore in the following analyses, we will mainly focus on the cortex-only model for simplicity.
Cell type-specific connectivity measures predict distributed persistent firing patterns
Structural connectivity constrains large-scale dynamics (Mejias and Wang 2022; Froudist-Walsh et al. 2021a; Cabral et al. 2011). However, we found that standard graph theory measures could not predict the pattern of delay period firing across areas. There is no significant correlation between input strength and delay period firing rate (r = 0.25, p = 0.25, Fig. 4A(i), A(ii)) and input strength cannot predict which areas show persistent activity (prediction accuracy = 0.51, Fig. 4A(iii)). We hypothesized that this is because currently available connectomic data used in this model do not specify the type of neurons targeted by the long-range connections. For instance, when two areas are strongly connected with each other, such a loop would contribute to the maintenance of persistent activity if projections are mutually excitatory, but not if one of the two projections predominantly targets inhibitory PV cells. Therefore, cell type-specificity of interareal connections must be taken into account in order to relate the connectome with the whole-brain dynamics and function. To examine this possibility, we introduced a cell type projection coefficient (see Calculation of network structure measures in the Methods), which is smaller with a higher PV cell fraction in the target area (Fig 4 - supplement 1). The cell type projection coefficient also takes cell type targets of long range connections into account, which, in our model, is quantified by counterstream inhibitory bias (CIB). As a result, the modified cell type-specific connectivity measures increase if the target area has a low density of PV interneurons and/or if long-range connections predominantly target excitatory neurons in the target area.
We found that cell type-specific graph measures accurately predict delay-period firing rates. The cell type-specific input strength of the early sensory areas is weaker than the raw input strength (Fig. 4B(i)). The firing rate across areas is positively correlated with cell type-specific input strength (Fig. 4B(ii)). Cell type-specific input strength also accurately predicts which areas show persistent activity (Fig. 4B(iii)). Similarly, we found that the cell type-specific eigenvector centrality, but not standard eigenvector centrality (Newman 2018), was a good predictor of delay period firing rates (Fig. 4 - supplement 2).
A core subnetwork for persistent activity across the cortex
Many areas show persistent activity in our model. However, are all active areas equally important in maintaining persistent activity? When interpreting large-scale brain activity, we must distinguish different types of contribution to working memory. For instance, inactivation of an area like VISp impairs performance of a delay-dependent task because it is essential for a (visual) “input” to access working memory; on the other hand a “readout” area may display persistent activity only as a result of sustained inputs from other areas that form a “core”, which are causally important for maintaining a memory representation.
We propose four types of areas related to distributed working memory: input, core, readout, and nonessential (Fig. 5A). External stimuli first reach input areas, which then propagate activity to the core and non-essential areas. Core areas form recurrent loops and support distributed persistent activity across the network. By definition, disrupting any of the core areas would affect persistent activity globally. The readout areas also show persistent activity. Yet, inhibiting readout areas has little effect on persistent activity elsewhere in the network. We can assign the areas to the four classes based on three properties: a) the effect of inhibiting the area during stimulus presentation on delay activity in the rest of the network; b) the effect of inhibiting the area during the delay period on delay activity in the rest of the network; c) the delay activity of the area itself on trials without inhibition.
In search of a core working memory subnetwork in the mouse cortex, in model simulations we inactivated each area either during stimulus presentation or during the delay period, akin to optogenetic inactivation in mice experiments. The effect of inactivation was quantified by calculating the decrement in the firing rate compared to control trials for the areas that were not inhibited (Fig. 5B). The VISp showed a strong inhibition effect during the stimulus period, as expected for an Input area. We identified eight areas with a substantial inhibition effect during the delay period (Fig. 5C), which we identify as a core for working memory. Core areas are distributed across the cortex. They include frontal areas PL, ILA, medial part of the orbital area (ORBm), which are known to contribute to working memory (Liu et al. 2014; Bolkan et al. 2017). Other associative and sensory areas (AId, VISpm, ectorhinal area (ECT), perihinal area (PERI), gustatory area (GU)) are also in the core. Similarly, we used the above criteria to classify areas as Readout or Non-essential (Fig. 5D).
The core subnetwork can be identified by the presence of strong excitatory loops
Inhibition protocols across many areas are computationally costly. We sought a structural indicator that is easy to compute and is predictive of whether an area is engaged in working memory function. Such an indicator could also guide the interpretation of large-scale neural recordings in experimental studies. In the dynamical regime where individual cortical areas do not show persistent activity independently, distributed working memory patterns must be a result of long-range recurrent loops across areas. We thus introduced a quantitative measurement of the degree to which each area is involved in long-range recurrent loops (Fig. 6A).
The core subnetwork can be identified by the presence of strong loops between excitatory cells. Here we focus on length-2 loops (Fig. 6A); the strength of a loop is the product of two connection weights for a reciprocally connected pair of areas; and the loop strength measure of an area is the sum of the loop strengths of all length-2 loops that the area is part of. Results were similar for longer loops (Fig. 6B, also see Fig. 6 - supplement 1 for results of longer loops). The raw loop strength had no positive statistical relationship to the core working memory subnetwork (Fig. 6C(i), Fig. 6C(ii)). We then defined cell type-specific loop strength (see Methods). The cell type-specific loop strength is the loop strength calculated using connectivity multiplied by the cell type projection coefficient. The cell type-specific loop strength, but not the raw loop strength, predicts which area is a core area with high accuracy (Fig. 6D(i), Fig. 6D(ii), prediction accuracy = 0.93). This demonstrates that traditional connectivity measures are informative but not sufficient to explain dynamics during cognition in the mouse brain. Cell type-specific connectivity, and new metrics that account for such connectivity, are necessary to infer the role of brain areas in supporting large-scale brain dynamics during cognition.
Multiple attractor states emerge from the mouse mesoscopic connectome and local recurrent interactions
Different tasks lead to dissociable patterns of internally sustained activity across the brain, described as separate attractor states. We developed a protocol to identify other attractor states, then analyzed the relationship between network properties and the attractor states (Fig. 7A-C). For different parameters, the number of attractors and the attractor patterns change. Two parameters are especially relevant here. These are the long-range connection strength (µEE) and local excitatory connection strength (gE,self). These parameters affect the number of attractors in a model of the macaque cortex (Mejias and Wang 2022). Increasing the long range connection strength decreases the number of attractors (Fig. 7D). Stronger long-range connections implies that the coupling between areas is stronger. If areas are coupled with each other, the activity state of an area will be highly correlated to that of its neighbors. This leads to less variability and fewer attractors.
To quantify how the patterns of attractors change for different parameters, two quantities are introduced. The attractor fraction is the fraction of all detected attractor states to which an area belongs. An area “belongs” to an attractor state if it is in a high activity state in that attractor. The attractor size is defined by the number of areas belonging to that attractor. As we increased the long-range connection strength, the attractor size distribution became bimodal. The first mode corresponded to large attractors, with many areas. The second mode corresponded to small attractors, with few areas (Fig. 7D).
When the local excitatory strength is increased, the number of attractors increased as well (Fig. 7E). In this regime some areas are endowed with sufficient local reverberation to sustain persistent activity even when decoupled from the rest of the system, therefore the importance of long-range coupling is diminished and a greater variety of attractor states is enabled. This can be understood by a simple example of two areas 1 and 2, each capable of two stimulus-selective persistent activity states; even without coupling there are 2 × 2 = 4 attractor states with elevated firing. Thus, local and long-range connection strength have opposite effects on the number of attractors.
The cell type-specific input strength predicted firing rates across many attractors. In an example parameter regime (µEE = 0.04 nA and gE,self = 0.44 nA), we identified 143 attractors. We correlated the input strength and cell type-specific input strength with the many attractor firing rates (Fig. 7F). The raw input strength is weakly correlated with activity patterns. The cell type-specific input strength is strongly correlated with activity across attractors. This shows that the cell type-specific connectivity measures are better at predicting the firing rates in many scenarios. These results further prove the importance of having cell type-specific connectivity for modeling brain dynamics.
Different attractor states rely on distinct subsets of core areas. In one example attractor, we found 5 areas that show persistent activity: VISa, VISam, FRP, MOs and ACAd (Fig. 7G) (parameter regime, µEE = 0.03 nA and gE,self = 0.44 nA). We repeated the previous inhibition analysis to identify core areas for this attractor state. Inhibiting one area, MOs, during the delay had the strongest effect on delay activity in the other parts of the attractor (Fig. 7H). MOs also showed strong persistent activity during delay period. This is consistent with its role in short-term memory and planning (Li et al. 2015; Inagaki et al. 2019). According to our definition, MOs is a core area for this attractor. To calculate a loop strength that was specific to this attractor, we only examined connections between these five areas. The cell type-specific loop strength was strongest in area MOs (Fig. 7I). Thus, we can identify likely core areas for individual attractor states from cell type-specific structural measures. This also demonstrates that different attractor states can be supported by distinct core areas.
Discussion
We developed a connectome-based dynamical model of the mouse brain. The model was capable of internally maintaining sensory information across many brain areas in distributed activity in the absence of any input. To our knowledge this is the first biologically-based model of the entire mouse cortex and the thalamocortical system for a cognitive function. Together with our recent work (Mejias and Wang 2022; Froudist-Walsh et al. 2021a; Froudist-Walsh et al. 2021b), it provides a study of contrast between mice and monkeys.
Our main findings are threefold. First, mnemonic activity pattern is shaped by the differing densities of PV interneurons across cortical areas. Areas with high PV density encoded information only transiently. Those with low PV density sustained activity for longer periods. Thus, the gradient of PV cells (Kim et al. 2017) has a definitive role in separating rapid information processing in sensory areas from sustained mnemonic information representation in associative areas of the mouse cortex. This is consistent with the view that each local area operates in the “inhibition-stabilizing regime” where recurrent excitation alone would lead to instability but the local network is stabilized by feedback inhibition even in the primary visual cortex (R. J. Douglas et al. 1995; Murphy and Miller 2009). Second, we deliberately considered two different dynamical regimes: when local recurrent excitation is not sufficient to sustain persistent activity and when it does. In the former case, distributed working memory must emerge from long-range interactions between parcellated areas, thereby the concept of synaptic reverberation (Lorente de Nó 1933; P. S. Goldman-Rakic 1995; Wang 2001; Wang 2021) is extended to the large-scale global brain. Note that currently it is unclear whether persistent neural firing observed in a delay dependent task is generated locally or depends on long-distance reverberation among multiple brain regions. Our work made the distinction explicit and offers specific predictions to be tested experimentally. Third, presently available connectomic data are not sufficient to account for neural dynamics and distributed cognition, and we propose cell type-specific connectomic measures that are shown to predict the observed distributed working memory representations.
We found that the cortical structures form recurrent loops with the thalamus, and the thalamocortical loops aided in sustaining activity throughout the delay period. The specific pattern of cortico-cortical connections was also critical to working memory. However, standard graph theory measures based on the connectome were unable to predict the pattern of working memory activity. By focusing on cell type-specific interactions between areas, we were able to reveal a core of cortical areas. The core is connected by excitatory loops, and is responsible for generating a widely distributed pattern of sustained activity. Outside the core, we identified “readout” areas that inherited activity from the core. Readout areas could use this information for further computations. This clarifies the synergistic roles of the connectome and gradients of local circuit properties in producing a distributed cognitive function. This additionally highlights the need for a cell type-specific connectome.
Previous large-scale models of the human and macaque cortex have replicated the functional connectivity (Deco et al. 2014; Demirtaş et al. 2019; Honey et al. 2007; Schmidt et al. 2018; Shine et al. 2018; Cabral et al. 2011; Wang et al. 2019) and propagation of information along the cortical hierarchy (Chaudhuri et al. 2015; Joglekar et al. 2018; Diesmann et al. 1999). Recently, large-scale models of brain activity during cognitive tasks have been developed (Mejias and Wang 2022; Froudist-Walsh et al. 2021a; Klatzmann et al. 2022). In these models, the number of dendritic spines per pyramidal cell increases along the hierarchy. This enables associative regions of the cortex to maintain information in working memory. The increase of dendritic spines along the hierarchy is a robust feature of primate cortical organisation (Elston 2007), which does not exist in the mouse cortex (Gilman et al. 2017). Yet, in the mouse cortex, other properties do vary along the cortical hierarchy (Kim et al. 2017; Fulcher et al. 2019). We took advantage of the recent discovery of a gradient of PV interneurons in the mouse cortex (Kim et al. 2017), and implemented it directly in our large-scale model. We demonstrated how the increasing gradient of excitation along the primate cortical hierarchy and the decreasing gradient of PV inhibition in the mouse cortex could serve a similar role. Both gradients enable sustained activity to emerge in associative areas. Thus, while the neural activity underlying working memory may be widely distributed in both rodents and primates, the circuit-level mechanisms may differ. This should be considered when interpreting studies of working memory in rodent models of cognition and disease.
In the macaque, long-range connectivity is a strong predictor of the working memory activity (Mejias and Wang 2022; Froudist-Walsh et al. 2021a). Thus, at least some of the functional specialization of brain areas is due to differences in interareal connections. In contrast, we found that traditional graph theory metrics of connectivity were unable to predict the working memory activity in the mouse brain. This may be due to the almost fully connected pattern of interareal connectivity in the mouse cortex (Gămănuţ et al. 2018). This implies that, qualitatively, all areas have a similar set of cortical connections. In our model, we allowed the cell type target of interareal connections to change according to the relative position of the areas along the cortical hierarchy. Feedforward connections had a greater net excitatory effect than feedback connections. This hypothesis (Mejias and Wang 2022) has received some recent experimental support (Yoo et al. 2021; Huang et al. 2019; Javadzadeh and Hofer 2022).
By introducing cell type-specific graph theory metrics, we were able to predict the pattern and strength of delay period activity with high accuracy. Connectome databases are an invaluable resource for basic neuroscience. However, they may be insufficient for constraining models of brain activity. In the future, connectome databases should be supplemented by cell type-specific information.
We demonstrated how cell type-specific graph-theory measures can accurately identify the core subnetwork, which can also be identified independently using a simulated large-scale optogenetic experiment. We found a core subnetwork of areas that, when inhibited, caused a substantial drop in activity in the remaining cortical areas. This core working memory subnetwork included frontal cortical areas with well documented patterns of sustained activity during working memory tasks, such as prelimbic (PL), infralimbic (ILA) and medial orbitofrontal cortex (ORBm) (Schmitt et al. 2017; Liu et al. 2014; Wu et al. 2020). However, the core subnetwork for the visual working memory task we assessed was distributed across the cortex. It also included temporal and higher visual areas, suggesting that long-range recurrent connections between the frontal cortex and temporal and visual areas are responsible for generating persistent activity and maintaining visual information in working memory in the mouse.
The core visual working memory subnetwork generates activity that is then inherited by many readout areas. Readout areas also exhibit persistent activity. However, inhibiting readout areas does not significantly affect the activity of other areas (Figure 5). Readout areas can use the stored information for further computations or to affect behavior. The readout areas in our model were a mixture of higher visual areas, associative areas and premotor areas of cortex. Notably, we classified the secondary motor cortex (MOs), which contains the anterior lateral motor (ALM) area, as a readout area despite its high level of persistent activity. ALM has received a lot of attention in mouse studies of working memory and motor preparation (Guo et al. 2017; Guo et al. 2014; Inagaki et al. 2019; Li et al. 2015; Wu et al. 2020; Voitov and Mrsic-Flogel 2022). If ALM is indeed a readout area for sensory working memory tasks, (e.g., Schmitt et al. 2017), then the following prediction arises. Inhibiting ALM should have a relatively small effect on sustained activity in core areas (such as PL) during the delay period. In contrast, inhibiting PL and other core areas may disrupt sustained activity in ALM. Even if ALM is not part of the core for sensory working memory, it could form part of the core for motor preparation tasks (Figure 7G). We found a high cell type-specific loop strength for area ALM, like that in core areas, which supports this possibility (Figure 7I). Furthermore, we found some attractor states for which the MOs was classified as a core area, while PL was not even active during the delay period. Our result is supported by a recent study that found no behavioral effect after PL inhibition in a motor planning task (Wang et al. 2021). Therefore, the core subnetwork required for generating persistent activity is likely task-dependent. Outside of this core subnetwork, there is a large array of readout brain areas that can use the stored information to serve behavior. Future modeling work may help elucidate the biological mechanisms responsible for switching between attractor landscapes for different tasks.
Neuroscientists are now observing task-related neural activity at single-cell resolution across much of the brain (Stringer et al. 2019; Steinmetz et al. 2019). This makes it important to identify ways to distinguish the core areas for a function from those that display activity that serves other purposes. We show that a large-scale inhibition protocol can identify the core subnetwork for a particular task. We further show how this core can be predicted based on the interareal loops that target excitatory neurons. Were such a cell type-specific interareal connectivity dataset available, it may help interpretation of large-scale recording experiments. This could also focus circuit manipulation on regions most likely to cause an effect on the larger network activity and behavior. Our approach identifies the brain areas that are working together to support working memory. It also identifies those that are benefiting from such activity to serve other purposes. Our simulation and theoretical approach is therefore ideally suited to understand the large-scale anatomy, recording and manipulation experiments which are at the forefront of modern systems neuroscience.
Neuroscience has rapidly moved into a new era of investigating large-scale brain circuits. Technological advances have enabled the measurement of connections, cell types and neural activity across the mouse brain. We developed a model of the mouse brain and theory of working memory that is suitable for the large-scale era. Previous reports have emphasized the importance of gradients of dendritic spine expression and interareal connections in sculpting task activity in the primate brain (Mejias and Wang 2022; Froudist-Walsh et al. 2021a). Although these anatomical properties from the primate cortex are missing in the mouse brain (Gămănuţ et al. 2018; Gilman et al. 2017), other properties such as interneuron density (Kim et al. 2017) may contribute to areal specialization. Indeed, our model clarifies how gradients of interneurons and cell type-specific interactions define large-scale activity patterns in the mouse brain during working memory, which enables sensory and associative areas to have complementary contributions. Future large-scale modeling studies can leverage cell type-specific connectivity to study other important cognitive computations beyond working memory, including learning and decision making (Abbott et al. 2017; Abbott et al. 2020).
Author contributions
XJW: designed the research, worked with the other authors throughout the project and co-wrote the paper; XYD: carried out all the computer simulations and analysis of simulation data, and co-wrote the paper; SFW, JJ and JJJ: contributed to all aspects of this project in interactions with XYD and co-wrote the paper.
Declaration of Interests
No competing interests declared.
Methods
Anterograde tracing, connectivity data
We used the mouse connectivity map from Allen institute (Oh et al. 2014) to constrain our large-scale circuit model of the mouse brain. The Allen Institute measured the connectivity among cortical and subcortical areas using an anterograde tracing method. In short, they injected virus and expressed fluorescent protein in source areas and performed fluorescent imaging in target areas to measure the strength of projections from source areas. Unlike retrograde tracing methods used in other studies (Markov et al. 2014b), the connectivity strength measured using this method does not need to be normalized by the total input or output strength. This means that connectivity strength between any two areas is comparable. The entries of the connectivity matrix from the Allen Institute can be interpreted as proportional to the total number of axonal fibers projecting from unit volume in one area to unit volume in another area. Before incorporating the connectivity into our model, we normalized the data as follows. In each area, we model the dynamics of an “average” neuron, assuming that the neuron receive inputs from all connected areas. Thus, we multiplied the connectivity matrix by the volume V olj of source area j and divided by the average neuron density di in target area i: where Wraw,ij is the raw, i.e., original, connection strength from unit volume in source area j to unit volume in target area i, V olj is the volume of source area j (Wang et al. 2020), and di is the neuron density in source area i (Erö et al. 2018). Wnorm,ij is the matrix that we use to set the long rang connectivity in our circuit model. We can define the cortico-thalamic connectivity Wct,norm,ij and thalamo-cortical connectivity Wtc,norm,ij in a similar manner.
Interneuron density along the cortex
Kim and colleagues measured the density of typical interneuron types in the brain (Kim et al. 2017). They expressed fluorescent proteins in genetically labeled interneurons and counted the number of interneurons using fluorescent imaging. We took advantage of these interneuron density data and specifically used the PV density to set local and long-range inhibitory weights. We first normalized the PV density in each area: PVraw,i is the PV interneuron density of all layers in area i, and PVi is a normalized value of PVraw, which will be used in subsequent modeling.
Hierarchy in the cortex
The concept of hierarchy is important for understanding the cortex. Hierarchy can be defined based on mapping corticocortical long range connections onto feedforward or feedback connections (Felleman and Essen 1991; Markov et al. 2014a; Harris et al. 2019). Harris and colleagues measured the corticocortical projections and target areas in a series of systematic experiments in mice (Harris et al. 2019). Projection patterns were clustered into multiple groups and the label “feedforward” or “feedback” was assigned to each group. Feedforward and feedback projections were then used to determine relative hierarchy between areas. For example, if the projections from area A to area B are mostly feedforward, then area B has a higher hierarchy than area A. This optimization process leads to a quantification of the relative hierarchy of cortical areas hraw,i. We defined the normalized hierarchy value hi as where hraw,i is the raw, i.e., original hierarchical ordering from (Harris et al. 2019). Due to data acquisition issues, 6 areas did not have a hierarchy value assigned to them (SSp-un, AUDv, GU, VISC, ECT, PERI) (Harris et al. 2019). We estimated hierarchy through a weighted sum of the hierarchy value of 37 known areas, while the weight is determined through the connectivity strength. The parameters αh and βh are selected so that hi,estimate are close to hi for areas with known hierarchy. For the thalamocortical model, we also used the hierarchy value for thalamic areas (Harris et al. 2019). The hierarchy of thalamic areas are comparable to cortical areas, so in order to use it in the model, we also normalized them. To estimate the hierarchy value of thalamic areas with missing values, we used the known hierarchy value of the thalamic area next to the missing one as a replacement.
Description of the local circuit
Our large-scale circuit model includes 43 cortical areas. Each area includes two excitatory populations, labeled A and B, and one inhibitory population, C. The two excitatory populations are selective to different stimuli. The synaptic dynamics between populations are based on previous firing rate models of working memory (Wang 1999; Wong and Wang 2006). The equations that define the dynamics of the synaptic variables are where SA and SB are the NMDA synaptic variables of excitatory populations A and B, while SC is the GABA synaptic variable of the inhibitory population C. rA, rB and rC are the firing rates of populations A, B and C, respectively. τN and τG are the time constants of NMDA and GABA synaptic conductances. γ and γI are the parameters used to scale the contribution of presynaptic firing rates. The total currents received Ii (i = A, B, C) are given by In these equations, gE,self, gE,cross denote the connection strength between excitatory neurons with same or different selectivity, respectively. These connection strengths are the same for different areas, since there is no significant gradient for excitatory strength in mice. gIE are the connection strengths from excitatory to inhibitory neurons, while gEI, and gII are connection strengths from inhibitory to excitatory neurons and from inhibitory to inhibitory neurons, respectively. These connections will be scaled by PV density in the corresponding area. We will discuss the details in the next section. I0i (i = A, B, C) are constant background currents to each population. ILR,i (i = A, B, C) are the long range (LR) currents received by each population. The term xi(t) where i = A, B, C represents noisy contributions from neurons external to the network. It is modeled as an Ornstein-Uhlenbeck process: where ζi(t) is Gaussian white noise, τnoise describes the time constant of external AMPA synapses and σi sets the strength of the noise for each population. σA = σB = 5pA while σC = 0pA.
The steady state firing rate of each population is calculated based on a transfer function ϕi(I) of input current received by each population Ii (i = A, B, C) given by Note that the transfer functions ϕi(t) are the same for two excitatory populations. x+ denotes the positive part of the function x. The firing rate of each population follows equations:
Interneuron gradient and local connections
We scaled local interneuron connectivity with the interneuron density that was obtained using fluorescent labeling (Kim et al. 2017). Specifically, local I-I connections and local I-E connections are scaled by the interneuron density by setting the connection strength gk,i(k = EI, II) as a linear function of PV density PVi in area i. where Jk,min (k = EI, II) is the intercept and Jk,scaling (k = EI, II) is the slope.
Hierarchy and long range connections
Long range (LR) connections between areas are scaled by connectivity data from the Allen Institute (Oh et al. 2014). We consider long-range connections that arise from excitatory neurons because most long-range connections in the cortex correspond to excitatory connections (Petreanu et al. 2009). Long-range connections will target excitatory populations in other brain areas with the same selectivity (Zandvakili and Kohn 2015) and will also target inhibitory neurons. These long-range connections are given by the following equations: where WE is the normalized long-range connectivity to excitatory neurons, and WI is the normalized long-range connectivity to inhibitory neurons. µEE and µIE are coefficients scaling the long-range E to E and E to I connection strengths, respectively.
Here, we assume that the long-range connections will be scaled by a coefficient that is based on the hierarchy of source and target area. To quantify the difference between long-range feedforward and feedback projections, we introduce mij to measure the “feedforwardness” of projections between two areas. According to our assumption of counterstream inhibitory bias (CIB), long-range connections to inhibitory neurons are stronger for feedback connections and weaker for feedforward connections, while the opposite holds for long range connections to excitatory neurons. Following this hypothesis, we define mij as a sigmoid function of the difference between the hierarchy value of source and target areas. For feedforward projections, mij > 0.5; for feedback projections, mij < 0.5. Excitatory and inhibitory long-range connection strengths are implemented by multiplying the long-range connectivity strength Wij by mij and (1 − mij), respectively: with The normalized connectivity Wnorm,ij is then rescaled to translate the broad range of connectivity values (over five orders of magnitude) to a range more suitable for our firing rate models. kscale is the coefficient used for this scaling. kscale < 1 effectively makes the range much smaller than the original normalized connectivity Wnorm,ij.
Thalamocortical network model
Corticothalamic connectivity
We introduced thalamic areas in the network to examine their effect on cortical dynamics. Each thalamic area includes 2 excitatory populations, A and B, with no inhibitory population. These two populations share the same selectivity with the corresponding cortical areas. Unlike cortical areas, there are no recurrent connections between thalamic neurons (Sherman 2007). Thalamic currents have the following contributions (tc stands for thalamocortical connections and ct for corticothalamic connections): where Ith,i (i = A, B) is the total current received by each thalamic population, Ict,i (i = A, B) is the long range current from cortical areas to target thalamic area, Ith,0,i (i = A, B) is the background current for each population, and Ith,noise,i (i = A, B) is the noise input to thalamic population A and B, which we set to 0 in our simulations. Ict,i (i = A, B) has the following form: where Wct,E,ij is the LR connectivity to thalamic neurons, and Sk,j is the synaptic variable of population k (k = A, B) in cortical area j. Since all thalamic neurons are excitatory, we model corticothalamic projections as in the previous section: where Wct,norm,ij is the normalized connection strength from cortical area j to thalamic area i. mct,ij is the coefficient quantifying how the long range connections target excitatory neurons based on cortical hierarchy hj and thalamic hierarchy hth,i.
The thalamic firing rates are described by: with the activation function for thalamic neurons given by: Thalamic neurons are described by AMPA synaptic variables (Jaramillo et al. 2019):
Thalamocortical connectivity
The connections from thalamic neurons to cortical neurons follow these equations and connectivity and connectivity matrix The thalamocortical input is added to the total input current of each cortical population.
Calculation of network structural measures
We considered three types of structural measures. The first one is input strength. Input strength of area i is the summation of the connection strengths onto node i. It quantifies the total external input onto area i. The second one is eigenvector centrality (Newman 2018). Eigenvector centrality of area i is the ith element of the leading eigenvector of the connectivity matrix. It quantifies how many areas are connected with the target area i and how important these neighbors are. The third structural measure is loop strength, which quantifies how each area is involved in strong recurrent loops. We first define the strength of a single loop k and then the loop strength of a single area Ai We now focus on cell type-specific structural measures. Cell type specificity is introduced via a coefficient kcell that scales all long range connection strengths (cell type projection coefficient): Thus, we can define cell type-specific input strength as: Similarly, cell type-specific eigenvector centrality is defined as and cell type-specific loop strength:
Stimulation protocol and inhibition analysis
We simulate a working memory task by applying an external current Istim to one of the excitatory populations. The external current is a pulsed input with start time Ton and offset time Toff. Without losing generality, we assume that the external input is provided to population A. In most of the simulations in this study, we simulate a visual working memory task, with the external applied to VISp. The simulation duration is Ttrial and we used a time step of dt.
We apply inhibition analysis to understand the robustness of attractors and, more importantly, to investigate which areas play an important role in maintaining the attractor state. Excitatory input was applied to the inhibitory population I to simulate opto-genetic inhibition. The external input Iinh is strong as compared to Istim and results in an elevated firing rate of the inhibitory population, which in turn decreases the firing rate of the excitatory populations. Usually the inhibition is applied to a single area. When inhibition is applied during the stimulus period, its start and end times are equal to Ton and Toff, respectively. When inhibition is applied during delay period, its start time is later than Toff to allow the system settle to a stable state. Thus, the onset of inhibition starts 2 seconds after Toff and lasts until the end of trial. In the case of thalamocortical network simulations, we inhibit thalamic areas by introducing a hyperpolarizng current to both excitatory populations, since we do not have inhibitory populations in thalamic areas in the model.
To quantify the effect of single area or multiple areas inhibition, we calculate the average firing rate of areas that satisfy two conditions: i) the area shows persistent activity before inhibition and ii) the area does not receive inhibitory input. The ratio between such average firing rate after inhibition and before inhibition is used to quantify the overall effect of inhibition. If the ratio is lower than 100%, this suggests that inhibiting certain area(s) disrupts the maintenance of the attractor state. Note that the inhibition effect is typically not very strong, and only in rare cases, inhibition of a single area leads to loss of activity of other areas (Fig. 5B, Fig. 5C). To quantify such differences, we use a threshold of 5% to differentiate them. We will use (relatively) “weak inhibition effect” and “strong inhibition effect” to refer to them afterwards.
We used the three measures to classify areas into 4 types (Fig. 5D): i) inhibition effect during delay period, ii) inhibition effect during stimulus period, and iii) delay period firing rate. Areas with strong inhibition effect during stimulus period are classified as input areas; areas with strong inhibition effect during delay period and strong delay period firing rate are classified as core areas; areas with weak inhibition effect during delay period but strong firing rate are classified as readout areas; areas with weak inhibition effect during delay period and weak firing rate during delay period are classified as nonessential areas.
Simulation of multiple attractors
Multiple attractors coexist in the network and its properties and number depends on the connectivity and dynamics of each node. In this study we did not try to capture all the possible attractors in the network, but rather compare the number of attractors for different networks. Here we briefly describe the protocol used to identify multiple attractors in the network. We first choose k areas and then generate a subset of areas as the stimulation areas. We cover all possible subsets, which means we run 2k simulations in total. The external stimulus is given to all areas in the subset simultaneously with same strength and duration. The delay period activity is then quantified using a similar protocol as the standard simulation protocol. The selection of k areas corresponds to a qualitative criterion. First we choose the areas with small PV fraction or high hierarchy, since these areas are more likely to show persistent activity. Second, the number of possible combination grows exponentially as we increase k, and if we use k = 43, the number of combinations is around 8.8e+12, which is beyond our simulation power. As a trade-off between the simulation power and coverage of areas, we choose k = 18, which correspond to 2.6e+5 different combinations of stimulation. For each parameter setting, we run 2.6e+5 simulations to capture possible attractor patterns. For each attractor pattern, a binary vector is generated by thresholding delay firing rate using a firing rate threshold of 5Hz. An attractor pattern is considered distinct if and only if the binary vector is different from all identified attractors. In these way we can identify different attractors in the simulation. We also apply same simulation pipeline to identify attractors for different parameters. Specifically we change the long range connectivity strength µEE and local excitatory connections gE,self.
Acknowledgements
We thank Daniel P. Bliss and Ulises Pereira for support with analysis tools at the beginning of the project, and members of the Wang Lab at New York University for discussions related to the project. This work was funded by US National Institutes of Health (NIH) grant R01MH062349, Office of Naval Research (ONR) grant N00014, National Science Foundation (NSF) NeuroNex grant 2015276, Simons Foundation grant 543057SPI, and NIH U19NS123714.
Footnotes
↵* co-first authors
Update on the author info. Minor update on the manuscript.