A multi-level account of hippocampal function from behaviour to neurons

A complete neuroscience requires multi-level theories that address phenomena ranging from higher-level cognitive behaviors to activities within a cell. A levels-of-mechanism approach that decomposes a higher-level model of cognition and behavior into component mechanisms provides a coherent and richer understanding of the system than any level alone. Toward this end, we decomposed a cognitive model into neuron-like units using a neural flocking approach that parallels recurrent hippocampal activity. Neural flocking coordinates units that collectively form higher-level mental constructs. The decomposed model suggested how brain-scale neural populations coordinate to form assemblies encoding concept and spatial representations, and why so many neurons are needed for robust performance at the cognitive level. This multi-level explanation provides a way to understand how cognition and symbol-like representations are supported by coordinated neural populations (assemblies) formed through learning.


Introduction
Neuroscience is a multi-level enterprise. Its target of explanation ranges from behavioral to molecular phenomena. Satisfying and complete explanations of the mind and brain will necessarily be multi-level (1)(2)(3). In multi-level componential (or constitutive) explanations, each component at a higher level can be decomposed into its own lower-level mechanism (1). For example, the circulatory system's capacity to deliver oxygen and energy to the body can be decomposed into lower-level mechanisms including the heart's blood pumping and kidney's blood filtering mechanism, which together supports the function of the circulatory system. These mechanisms themselves can be decomposed into their components, such as the muscle contractions of the heart or filtering units of the kidney, which in turn can be further decomposed as desired (1).
In neuroscience, the highest-level mechanism could be a cognitive model that captures behaviour. The components of this cognitive model could be related to neural measures and further decomposed. What is a component at a higher-level is a mechanism at a lower-level that can itself be decomposed into components to account for additional findings and make new predictions. The power of this approach is that mechanisms at different levels are not unrelated efforts aiming to explain different types of data. Instead, multi-level explanations can be integrated and offer a more complete understanding of the domain in which one can "zoom in or out" to the level of granularity desired for the current question.
Constructing theories and testing their predictions at multiple levels provides a more comprehensive account of the system of interest. Although neuroscience is guilty of a bias toward lower-level explanations (4,5), higher-level mechanisms are crucial in multi-level explanation because they offer explanatory concepts not available at lower levels (3). For example, the heart's contractions make little sense without considering the function of the circulatory system, and the hippocampal synaptic weights and activity patterns make little sense without notions of memory and learning. In neuroscience, it is common to construct a specific theory or model to fit the specific data at hand, which unfortunately leads to a disconnected patch-work of theories for individual phenomena. Multi-level theories can weave this patch-work together into a coherent and complete account in which each level is situated within mechanisms that lie above and below. As one descends levels, additional phenomena can be addressed, whereas as one ascends the function of the mechanism within the overall system becomes clearer.
Neuroscience has very few multi-level theories of this variety that can bridge between cognitive constructs and neuronal activity -multi-level explanations from behavior to neurons. For example, how does the brain implement a symbol? Specifically, how do brain systems coordinate neural populations to form symbol-like representations, such as highly selective neural assemblies (6)(7)(8) that encode concepts (9) or specific spatial locations (10)? One suggestion is that the hippocampal formation represents physical space (11)(12)(13) as well as abstract spaces for encoding concepts and abstract variables (14)(15)(16), constructing cognitive maps (17) for mental navigation in these spaces. However, it is unclear how populations of similarly-tuned neurons in the hippocampus acquire their highly-selective tuning properties to concepts or spatial locations. While tantalising, identifying a cell tuned to particular concept, such as Jennifer Aniston (9), does not specify the supporting mechanisms leaving such findings as curiosities that invite explanation.
Attempts have been made to offer multi-level theories in neuroscience but critical explanatory gaps remain. In our own work, we have developed a cognitive model, SUSTAIN, of how people learn category from examples. SUSTAIN addresses a number of behavioural findings (18,19) and aspects of the model have been related to the brain (19)(20)(21)(22). The hippocampus was related to SUSTAIN's clustering mechanism, which bundles together relevant information in memory during learning, and this account was verified by a number of brain imaging studies (19)(20)(21). The goal-directed attentional mechanism in SUSTAIN was linked to the ventral medial prefrontal cortex and this account was verified by brain imaging (22) and animal lesion studies (23). These same mechanisms provide a good account of place and grid cell activity in spatial tasks (16).
Although wildly successful in addressing both behaviour and accompanying brain activity, this line of work, like almost all work in model-based neuroscience, is limited to (1) proposing correspondences between model components and brain regions and (2) evaluating these correspondences in terms of model-brain correlates. But how do we move beyond mere neural correlates to a lower-level mechanistic explanation that unpacks the higher-level theory? Cognitive models, such as SUSTAIN, come with abstract constructs such as clusters, and it is left entirely open how they could be decomposed into the neural populations that give rise to behavior. It is insufficient to state that each cognitive construct (e.g., a cluster) is instantiated by a number of neurons, just as it is unsatisfying to state that Jennifer Aniston is somehow represented by multiple neurons, the spreadsheet on a computer relies on a number of transistors, and so forth. How do neurons coordinate to give rise to the higher-level construct? This is the key question that needs to be addressed to move beyond mere neural correlates toward a mechanistic understanding of how the brain implements cognition.
We aim to address this explanatory gap by decomposing aspects of a cognitive model, SUS-TAIN, into a mechanism consisting of neuron-like units. Critically, the aggregate action of these neuron-like units give rise to virtual structures akin to the higher-level cognitive constructs in SUSTAIN (i.e., clusters) while retaining SUSTAIN's account of behavior (Fig. 1). By taking a key component of a cognitive model that addresses behaviour and decomposing it to the level of a neuron, we offer an account of how concept cells and spatially-tuned cell assemblies can arise in the brain.
One of the main challenges is how to bridge from abstract constructs such as clusters to neurons, whilst retaining the high-level behavior of the model. How do single neurons track a concept? How does the brain coordinate the activity of many neurons during learning, as there are thousands of neurons, but only a few clusters are required to represent a concept at the cognitive level? In other words, how does a select population of neurons learn and become tuned to similar features in the world such as in concept cells and place cells, rather than independently develop their own tuning? To implement a higher-level construct, such as a cluster or symbol, neurons must somehow solve this coordination problem.
Inspired by algorithms that capture flocking behavior in birds and fish (24), we propose the hippocampus may exhibit neural flocking, where coordinated activity -virtual clusters or flocks -arise from local rules (Fig. 1). This coordination could be achieved by recurrent processing in the hippocampus (e.g., (25)), which we formalise in learning rules in which neuron-like units that are highly active in response to a stimulus (e.g., a photo of Jennifer Aniston) both adjust their tuning toward that stimulus and to each other. The latter learning update is the key to flocking as members of the neural flock coordinate by moving closer to each other. That simple addition to standard learning rules is sufficient to solve the coordination problem and give rise to virtual clusters as well as hippocampal place or concept cell (9) assemblies (7).
Gazing at the neuron-like units forming the model, one will not see clusters just as one will not see clusters nor symbols by peering into the grey goo of the brain. Nevertheless, the coordinated activity of these neuron-like units can be described as supporting these higherlevel constructs that behave in aggregate like the higher-level cognitive model, SUSTAIN. In the model specification and simulations that follow, we aim to close the aforementioned explanatory gap and make the case for multi-level explanation in neuroscience. Additionally, by decomposing the higher-level model we can consider how the brain benefits in terms of fault and noise tolerance by implementing clusters in a large neuronal flock. For instance, the mental representation of a concept is preserved when one neuron, or even a set of neurons in the neural assembly, dies, as well as when there is significant synaptic turnover over relatively short time scales (typical in the hippocampus; (26)). Finally, we consider how the model can be extended and further decomposed to account for differences in processing across anterior-posterior hippocampus axis.

Multi-neuron clustering model
Our multi-neuron clustering model, SUSTAIN-d, is a decomposition of the SUSTAIN model of category learning. Whereas prototype models always form one unit in memory for each category and exemplar models store one unit for each episode, SUSTAIN moves between these two extremes depending on the nature of the learning problem. SUSTAIN assumes the environment is regular and clusters similar experiences together in memory until there is a surprising prediction error, such as encountering a bat and wrongly predicting that it is a bird based on its similarity to an existing bird cluster. When such an error occurs, a new cluster is recruited.
Notable episodes (e.g., encountering a bat for the first time) can transition to concepts over time (e.g., other similar bats are stored in the same cluster). The hippocampus is critical in supporting this form of learning (27). SUSTAIN has other elements that are not the focus of this contribution, such as an attentional mechanism that determines which aspects of stimuli are most relevant to categorisation.
Here, we decompose SUSTAIN into more neuron-like units while retaining its overall function. We will refer to this model as SUSTAIN-d for SUSTAIN decomposed (see Methods for formal model description). Rather than recruit cognitive units like clusters, SUSTAIN-d, like the hippocampus, has an existing pool of neuron-like computing that enter each task with some tuning (i.e., preferentially activated for particular stimuli). Unlike SUSTAIN, which will recruit a handful of clusters for a learning problem, SUSTAIN-d can consist of an arbitrarily large num-ber of computing units (see Results for brain-scale simulations where the number of computing units is equal to the number of neurons in the hippocampus).
Despite these striking differences, SUSTAIN-d's thousands of neuron-like computing units show the same aggregate behaviour as SUSTAIN. This is accomplished by solving the aforementioned coordination problem by what we refer to as neural flocking (Fig. 1c). The key to neural flocking is that units that are highly activated by a stimulus both adjust their tunings toward the stimulus and each other. This double update leads to virtual clusters forming that can consist of thousands of neuron-like computing units. In general, the number of clusters SUS-TAIN recruits will match the number of neural flocks that arise in SUSTAIN-d, which leads to the models providing equivalent behavioural accounts (Fig. 1A). Whereas SUSTAIN associates clusters with a category or response, SUSTAIN-d's individual neuron-like units form connection weights (Fig. 1B). In summary, SUSTAIN-d is formulated absent of cognitive constructs like clusters, but nevertheless its neuron-like units in aggregate manifest the same principles and can account for the same behaviours, which provides an account of how cognitive constructs can be implemented in the brain.

Formation of concept and spatial representations by neural flocking
SUSTAIN-d's neural flocking mechanism can explain how concept cells ( Fig. 2A-B) and spatially-tuned place and grid cells arise (Fig. 2C-D). Whereas our previous work (16,18) relied on higher-level mental constructs (i.e., clusters) to account for such phenomena, here we show how a neural population can coordinate to virtually form such structures via the flocking learning rule. In the spatial domain, we simulated an agent (e.g., rodent) exploring a square environment (free foraging), which leads to virtual clusters akin to place cells distributed in a grid-like pattern (Fig. 2C), which in turn leads to cells that monitor these units' activity display-ing a grid-like firing pattern (Fig. 2D). In the conceptual domain, where the representation space is not as uniformly sampled, virtual clusters are clumpier (  Neural population-based model retains high-level cognitive model properties and captures concept learning behavior One major challenge for our multi-level proposal is to account for complex behaviours that hitherto were the sole province of cognitive models. Can SUSTAIN-d with its neuron-like units account for the same behaviors that SUSTAIN does by relying on cognitive constructs? We evaluate whether SUSTAIN-d can account for human learning performance on Shepard et al.'s six classic learning problems (Fig. 3). To provide a true multi-level theory, we aim for SUSTAIN-d's solution in terms of virtual clusters arising from neural flocking to parallel SUSTAIN's clustering solutions, which provide a good correspondence to hippocampal activity patterns (21).
SUSTAIN-d was trained in a manner analogous to human participants, learning through trial-and-error in these supervised learning tasks. On each trial, SUSTAIN-d, updated its neuronlike units' positions in representational space, attention weights, and connections weights from its neuron-like to category responses (see Methods for details). All the learning updates were error driven and local, as opposed to propagating errors overs multiple network layers as in deep learning models. Whereas SUSTAIN forms a new cluster in response to a surprising error (e.g., learning that a bat is not a bird), SUSTAIN-d recruits the k nearest unconnected neuron-like units to the current stimulus, which is in accord with the intuition that the brain repurposes existing resources during learning. The neural flocking learning rule leads to these recruited units forming a virtual cluster.
SUSTAIN-d captured the difficulty ordering of the human learning curves (Fig. 3B, right) and its solutions paralleled those of SUSTAIN in terms of the modal number of clusters recruited  Like the human brain, SUSTAIN-d is resistant to minor insults and faults. Each neural assembly or flock can consist of many neuron-like units (Fig. 3C, Fig. S2), not all of which are needed for the remaining units to function as a virtual cluster. SUSTAIN can be viewed as SUSTAIN-d when the number of highly activated units in response to the stimulus is 1 (i.e., k = 1). As k or total number of units increase, lesioning a subset of SUSTAIN-d's units has negligible effects on aggregate behavior (Fig. 3D, Fig. S3A). This robustness through redundancy is consistent with operation of the brain where multiple neurons and synapses with similar properties offset the effects of damage and natural turnover of dendritic spines (26,(32)(33)(34).
Having multiple units combined with SUSTAIN-d's recurrent update can also counteract noise (Fig. 3E). In these simulations, we added noise to SUSTAIN-d's neuron-like units, which will lead to units from other assemblies (or neural flocks) becoming highly active, which can lead to incorrect decisions and disrupt learning. SUSTAIN's recurrent update (Fig. 1C-D) ameliorates these effects of noise by pulling individual units' responses towards the mean of flock (Fig. 3E, Fig. S3B). The same self-organizing learning rules that enables SUSTAIN-d's neuron-like units to behave in aggregate like a cognitive model also make it more robust to noise and damage.
Further decomposing to capture differential function in anterior and posterior hippocampus In a multi-level mechanistic account, model components can be further decomposed to capture findings that require more fine-grain mechanisms. SUSTAIN-d decomposed SUSTAIN's clusters into neuron-like units. Here, we further decompose SUSTAIN-d's neuron-like units into two pools to capture functional differences between anterior and posterior hippocampus.
Anterior place fields tend to be broader and lower granularity than posterior hippocampal fields (35), and this appears to be a general principle at the population level (36,37). For category learning studies, one prediction is that learning problems with a simpler structure that promotes broad generalization would be more anterior whereas learning problems that have a complex irregular structure would be better suited to posterior hippocampus. Indeed, this pattern holds across studies (20, 21) (Fig. 4A). Here, we simulate Shepard's six learning problems, which order from what should be most anterior (Type I) to most posterior (Type VI).
Type I can be solved by focusing and generalizing based on one stimulus feature, whereas Type VI requires memorizing the details of each item.   only linked in that they both aimed to correctly predict the category label. Thus, the pool that was more suited to a learning problem will take over by developing larger connection weights to the output units indicating the response. In effect, both pools of units compete to guide the overall response (see Methods). Thus, the anterior pool of units should guide the overall system response when a problem is very simple (e.g., Type I) whereas the posterior pool should take over when a problem is highly irregular and complex (e.g., Type VI).
As predicted, the anterior pool guided the overall response for simple problems like the Type I problem whereas the posterior pool did so for complex problems, like the Type VI problem

Discussion
Neuroscience is inherently a multilevel enterprise that requires multilevel theories to move from a patchwork understanding of the mind and brain toward more integrative and encompassing theories. One solution, which we advance, is to adopt a levels of mechanism approach (cf. With a neural population and a recurrent mechanism, the model also naturally captures the brain's tendency to encode the same information in many neurons (i.e., redundancy; (32-34)), which makes the system more tolerant to neuronal damage, natural synaptic turnover, or noise.
Notably, there could be more than one recurrent update, as there are multiple recurrent loops in the brain. Future work could introduce more recurrent steps or different forms of recurrence such as constraining it by known anatomical pathways.
It is worth noting that the model assumes a sparse representation where most neurons are dormant but active neurons are highly active for a small group of stimuli, which parallels sparse coding in the hippocampus. Furthermore, the sparse code of a fixed proportion of highlyactivated winners in a neural population is consistent with recent hippocampal neurophysiological evidence showing that the overall activity of place cells within an environment is constant (38). This work shows how localist coding models (i.e., with sparse, highly-tuned neurons) can be implemented in a neural population (39)(40)(41)(42). One way for related approaches to show mixed selectivity is for different stimulus dimensions to be relevant in different contexts. For example, a neuron may respond to wool socks in a clothing context and to tart flavours in a food context with the relevant dimensions or features in each context being non-overlapping.
One benefit of multi-level theories is that model components can be decomposed into their own mechanisms as desired. In effect, one can selectively zoom in to consider the aspects of the mechanism of interest at a finer grain. In the last set of simulations, we further decom- We found that the bank of units best suited to the task dominated learning. For learning problems with a simple structure, the broader receptive fields of the anterior bank dominated. When the learning problem had an irregular structure that required memorization, the posterior bank dominated. By varying the broadness of the receptive fields, we introduced a generally applicable framework in which modules compete with one another to guide behavior but are ultimately complementary in terms of accomplishing the goals of the overall system. This cross-region competitive framework captures the common finding across cognitive neuroscience where brain regions that have the more appropriate or useful representations for the task at hand are more strongly activated and contribute more to behavior.
In the future, we plan to extend the method to include connections across modules (e.g., excitatory/inhibitory) based on known anatomy and functional properties, and to model more interacting brain regions. The neuron-like units themselves could be further decomposed and elaborated to behave more like biological neurons. For our present purposes, we did not require spiking neurons with complex dynamics. However, just as SUSTAIN was decomposed into SUSTAIN-d, so too can SUSTAIN-d be decomposed as desired, all the while retaining the overarching principles and behavior of the higher-level models.
In sum, cognitive neuroscience can benefit from multi-level explanations by exploring and bridging mechanisms across levels. We have many cognitive models that characterize behavior successfully, but are in need to be decomposed into a set of mechanistic processes that could be implemented in the brain. In recent years, neuroscience is finally putting more emphasis on behaviour (5), but we suggest that for a complete account of the cognitive function of interest, a successful high-level explanation of the behavior (e.g., through a cognitive model) that can be decomposed into the relevant lower-level mechanisms is key.

Methods
Overview and motivation of the model As discussed in the main text, SUSTAIN-d is a decomposition of SUSTAIN (18), a cognitive model of concept learning that has captured behaviour (18,19) in a number of tasks and brainactivity patterns including in the hippocampus and medial temporal lobe structures (16,20,21) (also see (43)). The formal specification of SUSTAIN is included in the aforementioned papers. Whereas SUSTAIN contains clusters (a cognitive construct) that are recruited in response to surprising events, SUSTAIN-d decomposes the notion of cluster into neuron-like units that coordinate to form virtual clusters through a flocking learning rule as described below.

Model Learning Implementation details
The where ζ is a positive scaling parameter which controls the steepness of the tuning curve, and dist i is the attention-weighted distance between neuron i's position pos i and the stimulus x in the R n representational space they inhabit: where r is set to 1 for the city-block distance (for separable-dimension stimuli) and the nonnegative a j attention weights sum to 1. The n attention weights correspond to each stimulus feature dimension, and their magnitude reflects the importance of each feature. The attention weighting also corresponds to each unit's receptive field, which is centered on its position along each feature dimension (Fig. 1B, attention weighting). ζ controls the steepness of the receptive field and will be used to model the different tuning properties of the anterior and posterior hippocampus (see below). As such, units most strongly tuned to the current stimulus input are activated and the activity is propagated forward to produce the categorization decision. Only winners have non-zero output that contribute to the category decision: and evidence for category decisions propagates from the output: where w ij is the connection weight (Fig. 1B, cyan and pink connections) between unit i and decision j and m is the number of units. Finally, the probability of making decision j is computed by: where z is the number of possible decisions and φ is a non-negative decision certainty parameter, or inverse temperature (see (18,44) for related formulations). If there are no connected units (e.g. first trial) or fewer than k winners, then no units respond or fewer than k units respond, respectively.
During learning, the k winners update twice, once toward the stimulus and a second time toward each other, which supports neural flocking. We view unit updating as a continuous process through time relying on recurrent connections, which we simplify here to two simple updates. In the first update, the k winners update their positions toward the current stimulus' position on each trial according to the Kohonen learning rule: where η pos is the learning rate, x is the current stimulus' location and pos i is to unit i's position in representational space (Fig. 1C, 1st update). Bold type is reserved for vectors. This is the first learning step, where the initial forward pass of stimulus information occurs and the first clustering update is applied. In the second update, the k winners perform an additional recurrent step which re-codes the stimulus based on their activity, updating their positions toward the centroid of the k winners' positions: where η group is the learning rate for the recurrent update and group is the centroid (mean position) of all the winners (Fig. 1C, 2nd update). With this double-update rule, neurons update their tuning profiles to activate more to stimuli that inhabit that part of space in a coordinated fashion, and over the course of learning will typically stabilize into a portion of the space adaptive for the task. As co-activated units cluster together and become more similar to each other in tuning, this group naturally lead to a virtual cluster (i.e., many units that are similar tuned to a concept) similar to a cluster (18) and akin to to a hippocampal place or concept cell assembly where assemblies of neurons show similar tuning and co-activate to similar stimuli or environmental features (7)(8)(9).
To update attention weights over learning, the model applies gradient ascent on the summed activation of winning units minus the summed activation of the non-winner units (i.e., connected but not a winner) with learning rate η attn . Note that this learning rule is local (i.e., no backpropagation).
Connection weights are updated using descent on error (based on the output prob j ) using cross-entropy loss with one-hot target vector (i.e., one category is correct) on each trial, with learning rate η cweights . To display the output of the model and for fitting to the behavioural data, we plot 1 − prob j (where j is the correct category decision), averaged for each block to match the error learning curves as done in prior work.
As study. We present a brief overview of this classic study here. Participants learned to categorize eight stimuli that varied on three binary feature dimensions (shape, size, and color) into two categories. The concept structure was one of six possible logical structures from Shepard et al. (Fig. 3A). On each trial, participants categorized each stimulus into a category and was are the data we will fit (Fig. 3B, left).
Task blocks consisted of 16 stimuli presented in a randomized order. To obtain error curves for each parameter set, a random stimulus sequence for each problem type was generated 25 times, and the error curves were produced by taking the mean across those iterations. To maintain consistency, each iteration was seeded with a specific number, so that the 25 sequences were the same across the different parameters.
Model learning curves were fit to display the human pattern of results. We performed a hierarchical grid search across the parameters. For the standard model (i.e. no separation of brain regions or modules), there were 6 free parameters: ζ, φ, and four learning rates (attention weights η attn , connection weights η cweights , Kohonen update η pos , recurrent update η group ). To fit the model that includes a separate anterior and posterior bank of units, separate tuning (ζ) parameters were used for each bank with the constraint that the anterior bank should have a broader tuning than the posterior bank of units (12 free parameters).

Robustness to failure modes: noise and lesion experiments
To demonstrate the beneficial effect of having a population of neurons (rather than a single unit or cluster in cognitive models) and the recurrent update, we simulated Shepard's problems with different failure modes during the learning process and how robust the model was to these perturbations.
To simulate noise in the learning process, we added noise to the units' positions. For each trial, noise was sampled from a n dimensional Gaussian distribution (corresponding to n features) with zero mean and standard deviation of 0, 0.5, or 1.0 which was added to the update.
By adding noise to the unit's position in representational space, this causes potential problems for 1) selecting the appropriate k neurons as winners, 2) appropriate updating of the attention weights, and 3) appropriate updating of the connection weights.

Unsupervised learning on spatial tasks
To simulate a rodent foraging in an environment, we placed an agent in a two-dimensional square environment and produced a randomly-generated 500,000 set of steps with the restriction that the agent could not step out of the environment. On each trial, it was able to move left, right, up, or down in steps of 0, 0.025, 0.05, or 0.075. The environment was a square that spanned from 0 to 1 on the horizontal and vertical dimensions.
For unsupervised learning, the model could recruit units like SUSTAIN does by relying on a surprise signal. Here, we further simplify as in (16) and assume all units in the population are relevant to the current context. Unit positions were updated according to the learning rules specified above. While a rodent's actual environment contains many features, we assumed these features effectively reduce to a two-dimensional space corresponding to coordinates within the agent's enclosure.
On each trial, the agent moved a step (randomly selected over four directions and four step sizes; one trial), and the model updated the k winners with the Kohonen learning rule as before, with an annealed learning rate so that the units would eventually settle and stabilize into a particular location: where η t is the learning rate at time t. The learning rate followed an annealing schedule: where η 0 is the initial learning rate and ρ is the annealing rate set to 4 × 10 −12 (see (16)). The recurrence update learning rate η group was fixed at 1.0, though smaller values such as 0.8 and 0.6 produce similar results.
To compute grid scores at the end of learning, activation maps were produced by generating a new movement trajectory with 250,000 steps (as above), and computing the unit activations based on their positions at the end of learning (i.e. freezing the positions; see (16)). For each value of k, we ran 100 simulations and computed the grid scores at the end of learning. The activation maps were binned in 40x40 bins (original 100x100), then normalized by the number of visits to each binned location (normalized activation map). Grid scores were calculated based on (45). Briefly, the spatial autocorrelogram of the activation maps were calculated as defined in (46), and gridness was computed using the expanding gridness method, where a cicular annulus with a radius of eight bins was placed on the center of the autocorrelogram, with the central peak removed. The annulus was rotated in 30°steps, and the Pearson correlation between the rotated and unrotated version of the spatial autocorrelogram was recorded. The highest correlation value for 30°, 90°, and 150°rotations was subtracted from the lowest correlation value at 0°, 60°, and 120°to give an interim grid score. This was repeated expanding the annuls by two bins, up to 20. The final grid score was the highest interim grid score.