Elsevier

Neurocomputing

Volume 70, Issues 7–9, March 2007, Pages 1130-1138
Neurocomputing

Learning sensory representations with intrinsic plasticity

https://doi.org/10.1016/j.neucom.2006.11.006Get rights and content

Abstract

Intrinsic plasticity (IP) refers to a neuron's ability to regulate its firing activity by adapting its intrinsic excitability. Previously, we showed that model neurons combining a model of IP based on information theory with Hebbian synaptic plasticity can adapt their weight vector to discover heavy-tailed directions in the input space. In this paper we show how a network of such units can solve a standard non-linear independent component analysis (ICA) problem. We also present a model for the formation of maps of oriented receptive fields in primary visual cortex and compare our results with those from ICA. Together, our results indicate that intrinsic plasticity that tries to locally maximize information transmission at the level of individual neurons may play an important role for the learning of efficient sensory representations in the cortex.

Introduction

Computational models of unsupervised learning of sensory representations in the brain abound. Frequently, they fall into one of two categories: mechanistic models or functional models. Mechanistic models start from neuroscientific data about the structure of cortical networks and cortical plasticity mechanisms (cell types, connection patterns, plasticity rules, etc.) which are distilled into simplified models. These models are trained on actual sensory data or noise patterns and the learned representations can be compared to neurophysiological observations. If the resulting representations are similar to those found in the brain then this provides evidence that the processes in the brain have been accurately captured, but it does not clarify why the brain operates this way or in what sense the brain's solution may be optimal. An example of a model of this kind is by Linsker [15], where V1-style orientation columns are learned from random prenatal visual noise through Hebbian learning. Later Miller extended this work to learn many of the various map-structures in V1, and used model neurons that were somewhat more plausible [19].

Functional models focus on the abstract computational goal of the problem. For the case of learning sensory representations they start by asking: what is the optimal way to represent sensory stimuli such as natural images, where optimality is usually defined with respect to certain statistical criteria (e.g. sparseness, independence, temporal coherence, etc.) and additional constraints. Algorithms are derived to learn the optimal solution to the problem, which can again be compared to neuroscientific data. If the found solution resembles the biological solution, then this provides evidence that the brain may in fact be trying to optimize a similar objective function. Through what mechanisms the brain may achieve this goal is typically not answered, however. Some examples of such an approach will be given below.

Both mechanistic and functional models have their merits, but for a comprehensive understanding of sensory coding in the cortex we arguably have to develop models that bridge functional and mechanistic levels of description. Such models should explain how the physiological mechanisms contribute to optimizing the system's information processing properties in a meaningful way. In the following, we develop a model that can be viewed as a step in this direction.

A central idea in many functional models of the development of sensory representations is information maximization [1], [3], [16], [20], [22]. According to some formulations of this idea, individual neurons should maximize the entropy of their firing rate distribution. If the firing rate is constrained to lie in a fixed interval between zero and the neuron's maximum firing rate, then entropy maximization means that the neuron should use all its firing rate levels equally often. In order to achieve this, it should spread out its responses in dense regions of the input space and compress responses in sparse regions such that it maps the distribution of its inputs to a uniform distribution of its outputs, maximizing entropy. Biological evidence for this idea comes from Laughlin, who showed that blowfly large monopolar cells have been adapted so that their input/output transfer functions nearly optimally represent the contrast statistics of the blowfly's visual environment [13].

Information maximization may not be the only important objective, however, and energy considerations may also play an important role for sensory coding in the brain, e.g. [14]. In particular, Baddeley et al. found that neurons in different visual cortical areas of cats and monkeys show exponential distributions of their firing rate. They have argued that this maximizes a neuron's information transfer given a fixed energy budget [2]. This is because the exponential distribution has the maximum entropy among all distributions of a positive random variable (the firing rate) with a fixed mean. This and other reasons suggest that sparse representations, where individual units are highly active only rarely, may be an important principle of sensory coding [10].

On the modeling side, Olshausen and Field showed that localized, oriented, and bandpass receptive fields similar to those observed in primary visual cortex (V1) arise when optimizing image reconstruction error subject to lifetime sparseness constraints [21]. They imposed a sparse prior on the contribution of each basis function in a generative model with the intuition that among the space of possible sources of an image, each one is present only rarely. In a closely related approach, Bell and Sejnowski showed that the information maximization principle can be applied to the independent component analysis (ICA) problem. They applied their technique to natural images and also found localized, oriented, and bandpass sources [4].

Most work on the learning of sensory representations has focused on synaptic plasticity as the only mechanism for learning efficient codes. But it is becoming increasingly clear that biological neurons also regulate their pattern of firing by adapting their intrinsic excitability through the modification of voltage-gated channels in their membrane. Such intrinsic plasticity (IP) seems to be a ubiquitous phenomenon in the brain [30]. For example, Desai et al. showed that neurons that had been prevented from spiking for two days increased their response to current injection [6]. Consistent with this finding, it is frequently assumed that IP contributes to the homeostasis of a neuron's mean firing activity. A few computational models do in fact incorporate a mechanism for regulating the mean activity level of a unit by controlling a “threshold” parameter [7], [8], [9]. But it is also plausible that IP may help to optimize the encoding and transmission of information in a more sophisticated fashion. Concretely, it has been speculated that IP may be instrumental in achieving approximately exponential firing rate distributions in cortical neurons [23]. More recently, we have shown that an IP mechanism that drives a neuron to exhibit an exponential firing rate distribution can synergistically interact with Hebbian learning at the synapses. The two processes lead to the discovery of heavy-tailed directions in the input space [24], [26].

In this paper we extend these results to networks of neurons with IP and Hebbian learning. Our specific goal is to explore the potential role of IP for learning efficient map-like representations for sensory stimuli. The model we present in the following attempts to bridge the gap between mechanistic and functional models. On the one hand, it has a clear connection to the idea of information maximization and energy efficient coding [28]. On the other hand, it has a mechanistic formulation that is biologically viable because the learning mechanisms make use of information that is local in time and space. While similar bridges have been attempted before, e.g. [5], [8], our model is distinguished by utilizing an IP model derived from information theory as a mechanism for the learning of efficient sensory representations.

Section snippets

Network model with intrinsic plasticity

We consider a network of units learning to represent a sensory input vector x. The activity of unit i in the network in response to input x is given byyi(hi)=[1+exp(-aihi-bi)]-1withhi=x·wi,where wi is the neuron's weight vector, “·” denotes the inner or dot product, and ai and bi are adjustable parameters of the neuron's transfer function that are controlled by IP (compare Fig. 1a). In particular, ai and bi are adapted in such a way that the unit's output yi assumes an approximately exponential

The “bars” problem

As a first test bed for studying the learning of sensory representations with networks of units with intrinsic plasticity we consider the “bars” problem. This is a standard non-linear ICA problem introduced by Földiák [9]. Horizontal and vertical bars are presented on a retina of R-by-R pixels. The presence or absence of a bar is independent of that of any other bars. The unsupervised learning problem is to learn filters that correspond to the individual independent components, i.e. the bars.

Modeling the emergence of orientation maps

Receptive fields of simple cells in primary visual cortex (V1) are oriented, localized, and bandpass. In addition, neighboring neurons in V1 will have a similar orientation preference, giving rise to smooth orientation maps. For modelling the emergence of orientation maps, we consider the neurons in our network to be located on a two-dimensional sheet, with neuron i at grid position (j,k)iN×N after the fashion of a self-organizing map (SOM). The most active unit exhibits a center-surround

Discussion

Different forms of plasticity are involved in shaping sensory representations in the brain and it is important to understand how these different mechanisms interact. In [24], [25] we developed model neurons that maintain sparse lifetime distributions of their individual activities through IP. We showed that when IP is combined with various forms of Hebbian learning at the synapses, a single unit will discover heavy-tailed directions in its input [24], [26]. Here we constructed networks of such

Acknowledgments

The authors thank Cornelius Weber, Erik Murphy-Chutorian, and three anonymous reviewers for comments on earlier drafts. This work was supported by the Hertie foundation.

Nicholas Butko is a graduate student in the department of Cognitive Science at the University of California in San Diego. His research interest is developing novel algorithms to allow machines to perceive and act in real-world situations.

References (30)

  • A.J. Bell et al.

    The independent components of scenes are edge filters

    Vision Res.

    (1997)
  • P. Lennie

    The cost of cortical computation

    Curr. Biol.

    (2003)
  • B.A. Olshausen et al.

    Sparse coding with an overcomplete basis set: a strategy employed by V1?

    Vision Res.

    (1997)
  • F. Attneave

    Some informational aspects of visual perception

    Psychol. Rev.

    (1954)
  • R. Baddeley et al.

    Responses of neurons in primary and inferior temporal visual cortices to natural scenes

    Proc. R. Soc. London B

    (1998)
  • H.B. Barlow

    Possible principles underlying the transformation of sensory messages

  • L.N. Cooper et al.

    Theory of Cortical Plasticity

    (2004)
  • N.S. Desai et al.

    Plasticity in the intrinsic excitability of cortical pyramidal neurons

    Nature Neurosci.

    (1999)
  • D. DeSieno, Adding a conscience to competitive learning, IEEE Proceedings of the International Conference on Neural...
  • M.S. Falconbridge et al.

    A simple Hebbian/anti-Hebbian network learns the sparse, independent components of natural images

    Neural Comput.

    (2005)
  • P. Földiák

    Forming sparse representation by local anti-hebbian learning

    Biol. Cybern.

    (1990)
  • P. Földiák

    Sparse coding in the primate cortex

  • A. Hyvärinen et al.

    Emergence of phase- and shift-invariant features by decomposition of natural images into independent feature subspaces

    Neural Comput.

    (2000)
  • A. Hyvärinen et al.

    Topographic independent component analysis

    Neural Comput.

    (2001)
  • S. Laughlin

    A simple coding procedure enhances a neuron's information capacity

    Z. Naturforsch

    (1981)
  • Cited by (19)

    • Sparse coding with a somato-dendritic rule

      2020, Neural Networks
      Citation Excerpt :

      This usually involves a winner-take-all system (Kohonen, 1990), or inhibitory connections between the coding neurons (Marshall, 1990, 1992) — an organisation which is equivalently called lateral, recurrent or mutual inhibition. Starting with Földiák (1990), these two heuristics have been applied in a variety of sparse coding networks with rate-based (Butko & Triesch, 2007; Falconbridge, Stamps, & Badcock, 2006; Lucke, 2007) and then spiking neurons (Ferré, Mamalet, & Thorpe, 2018; King, Zylberberg, & DeWeese, 2013; Savin, Joshi, & Triesch, 2010; Zylberberg, Murphy, & DeWeese, 2011). These networks have in common the use of Hebbian lateral inhibition to decorrelate the output, and of nonlinear Hebbian rules to perform projection pursuit on the feedforward input.

    • Intrinsic plasticity via natural gradient descent with application to drift compensation

      2013, Neurocomputing
      Citation Excerpt :

      This maximizes the neuron's information transmission, related to the high entropy of the target distribution. Since its introduction, the IP-rule has been used to learn sensory representations [2] and enhance the encoding in reservoir networks [3,4], which are of main interest in this paper. A batch-version has been shown to improve extreme learning machines [5] and static reservoirs [6].

    View all citing articles on Scopus

    Nicholas Butko is a graduate student in the department of Cognitive Science at the University of California in San Diego. His research interest is developing novel algorithms to allow machines to perceive and act in real-world situations.

    Jochen Triesch is Assistant Professor of Cognitive Science at UC San Diego and a Fellow of the Frankfurt Institute for Advanced Studies. His research interests span neural computation, human and machine vision, cognitive robotics, and models of human cognitive development.

    View full text