Abstract

Chromaticity diagrams—Cartesian plots that depict the multidimensional subjective relationship between perceived colors, independent of intensity—are widely used in studies of animal color vision, as they allow direct links to be made between color discrimination and color vision–mediated behaviors. However, because these diagrams are invariant to the differences in noise levels between photoreceptor classes that underlie some models of color discrimination thresholds, the distance between colors in chromaticity space is not necessarily related to their perceptual distance. The geometric interpretation of color data from chromaticity diagrams may therefore lead to misleading conclusions. This paper describes a method of plotting visual perceptual distances derived from the noise-corrected model of Vorobyev and Osorio (Vorobyev M, Osorio D. 1998. Receptor noise as a determinant of color thresholds. Proc R Soc B. 265:351–358) such that the perceptual distance between them is exactly preserved. It also describes how both chromatic (color) and achromatic (brightness) contrasts can be visualized within the same diagram. The potential benefits of this approach over commonly used triangular and tetrahedral chromaticity diagrams are discussed with regards to the study of animal visual ecology.

INTRODUCTION

The graphical representation of color is an invaluable technique for describing experimental data and for modeling color thresholds (Chittka 1992; Backhaus 1998; Kelber et al. 2003; Endler and Mielke 2005) and has a long history of use in studies of human color vision (e.g., Maxwell 1860; MacLeod & Boynton 1979; Derrington et al. 1984). More recently, it has also been increasingly used to visualize the relationships between colors of importance to animal receivers (e.g., Osorio et al. 1999; Defrize et al. 2010). Such representations incorporate knowledge of the receiver's visual system into the calculation of a color's spatial position and so add a level of biological realism in the interpretation of color data compared with using visual system-independent analyses (e.g., Andersson and Prager 2006). The most commonly used diagrams, at least in vertebrate studies, are the triangular or tetrahedral chromaticity diagrams described by Kelber et al. (2003) and Endler and Mielke (2005) that allow the visualization of a color in terms of the relative quantum catch of the different photoreceptor classes in the retina, each of which is maximally sensitive to a different wavelength of light. They also allow the geometric description of the spatial relationships between color data and the extraction of spatial coordinates for subsequent analysis (Endler and Mielke 2005). However, because there is no straightforward relationship between the intercolor distance in a chromaticity diagram and the perceptual distance between those colors (Endler and Mielke 2005), in some situations, they may misrepresent the perceptual relationships between colors.

To illustrate this, I consider 1) the distance between pairs of colors in the commonly used triangular chromaticity diagram described in detail by Kelber et al. (2003) and 2) the perceptual distance between those colors determined from the noise-corrected model of Vorobyev and Osorio (1998), as described below. In the examples, I have used data on the visual system of the three-spined stickleback (Gasterosteus aculeatus). Sticklebacks are tetrachromatic, having 4 photoreceptor types with peak sensitivities, λmax, at 605 (denoted L), 530 (M), 445 (S), and 360 nm (U) (Rowe et al. 2004), although in most of the examples below I only consider the L, M, and S cones for simplicity. Photoreceptor spectral sensitivity functions were calculated using the photopigment nomogram of Govardovskii et al. (2000) and incorporate transmission across the preretinal media (Boulcott 2003). Unless otherwise stated, I assume full-sunlight illumination conditions (modeled using a standard daylight-simulating illumination spectrum, D65; Wyszecki and Stiles 1982).

For an eye viewing a color, the quantum catch of a photoreceptor i, Qi, is given by:
Qi=λminλmaxRi(λ)S(λ)I(λ),
(1)
where λ denotes wavelength, λmin and λmax are, respectively, the lower and upper limits of the visible spectrum, Ri(λ) is the spectral sensitivity of receptor i, S(λ) is the reflectance spectrum, and I(λ) is the illumination spectrum (Endler and Mielke 2005).
When color data are plotted in triangular chromaticity space, information on the quantum catch of each of the 3 photoreceptor classes, QL, QM, and QS, is normalized by dividing by the sum of quantum catches over all 3 cone types to yield qL, qM, and qS, such that (qL + qM + qS) = 1. These normalized quantum catches are then used to calculate 2D Cartesian coordinates for each color (Endler and Mielke 2005), as:
x=12(qLqS),y=23(qM(qL+qS)2).
(2)
Colors defined using these coordinates are confined to a triangular region, with maximally chromatic colors (in which a single 1 of the 3 cone classes is stimulated) being at the triangle's apices. This color space therefore uses only information on the quantum catch of each photoreceptor type (and not, for instance, information on the noise in each photoreceptor channel, which may affect the ability of an animal to discriminate particular colors [sensu Vorobyev and Osorio 1998], or information on subsequent neural processing). Nonetheless, the Euclidean distances between 2 colors, A and B, in this color space, ΔST, given by
ΔST=(xAxB)2+(yAyB)2
(3)
provides an indication of the ability of an animal to distinguish between them, such that a greater spatial distance indicates a greater perceptual distance between the 2 colors (Endler and Mielke 2005). As a result, spatial data derived from this chromaticity space have been widely used to explain color vision–mediated behavior in animals (e.g., Osorio et al. 1999; Cassey et al. 2010; Pike et al. 2010).
The model described by Vorobyev and Osorio (1998) allows the calculation of perceptual distances between 2 colors, taking into account the noise levels in different photoreceptor types. This results in a measure of chromatic discriminability in units of just noticeable differences (jnds). Under ideal conditions, differences between 2 colors <1 jnd should be indiscriminable, those ∼1 jnd barely discriminable, and those >1 increasingly discriminable (although in practice the discrimination boundary often lies at slightly more than 1 jnd; Defrize et al. 2010). Specifically, when 3 cone classes are assumed to contribute to the discrimination of chromatic stimuli, the distance ΔSVO between 2 colors, A and B, is calculated (Vorobyev and Osorio 1998) as follows:
ΔSVO=eS2(ΔfLΔfM)2+eM2(ΔfLΔfS)2+eL2(ΔfSΔfM)2(eSeM)2+(eSeL)2+(eLeM)2,
(4)
where Δfi is the difference in response for the cone type i between colors A and B, calculated as:
Δfi=log(QiAQiB),
(5)
and the term ei denotes the standard deviation of noise in receptor mechanism i. The value of ei is not fixed but typically differs both between photoreceptor classes and as a result of changing illumination conditions (Osorio et al. 2004). Under illumination levels typically considered in behavioral studies, photoreceptors suffer from 2 main sources of noise, which predominate at different intensity ranges (Barlow 1964). First, at low light levels, variation in photon catch is likely to be the main source of noise, such that noise is proportional to the square root of the number of absorbed quanta. This means the contrast threshold should fall with the square root of intensity, a relationship known as the DeVries–Rose law (Barlow 1964; Rovamo et al. 2001). Second, under bright illumination (such as under direct sunlight), noise is proportional to signal intensity, leading to contrast thresholds being independent of intensity, as predicted by Weber’s law (Barlow 1972). Under these conditions, noise can be modeled (Vorobyev and Osorio 1998) as:
ei=ωηi,
(6)
where ω is the Weber fraction (the smallest detectible contrast; Wyszecki and Stiles 1982), and ηi describes the relative density of cone class i on the retina. Because photoreceptors with higher values of λmax tend to be more abundant in the retina than those with lower λmax, they will typically suffer from lower noise levels (Osorio et al. 2004). The approach taken by the Vorobyev–Osorio model is intuitively appealing and has therefore been used to identify color thresholds and link discriminability to behavior in a variety of animal species with considerable success (e.g., Siddiqi et al. 2004; Eaton 2005; Loyau et al. 2007; Defrize et al. 2010).

However, there is not necessarily a straightforward relationship between the Euclidean distance between data points in chromaticity space (i.e., ΔST) and the perceptual distance between them according to the Vorobyev–Osorio model (i.e., ΔSVO) (Figure 1a). To illustrate this further, Figure 1b shows 5 data points, labeled A–E, plotted in a stickleback triangular chromaticity space representing the quantum catches of the L, M, and S cones. The Euclidean distance between A and B and between B and C is approximately equal (i.e., ΔST(A,B) ≈ ΔST(B,C) ≈ 0.25 units), although if the Vorobyev–Osorio model is applied to these data (assuming, for simplicity, an equal standard deviation of noise in each photoreceptor class of ei = 0.05 and bright illumination conditions), then it can be shown, using the equations given above, that the noise-corrected levels of discriminability are unequal: ΔSVO(A,B) = 17.9 jnds and ΔSVO(B,C) = 12.6 jnds. The reason for this is that the contour representing an equal number of jnds around each point in chromaticity space is typically not circular, but elliptical, and moreover, the shape and size of a contour changes with the spatial position of a color (see the contours around each point in Figure 1b), generally getting larger and more circular as a color approaches the graph's origin (the achromatic, or “gray” point). The size of a contour is also affected by the level of noise in a photoreceptor (both the inner and the outer contours in Figure 1b assume an equal standard deviation of noise in each photoreceptor; however, the inner contours assumes that ei = 0.05 and the outer contour assumes that ei = 0.1).

Figure 1

(a) Relationship between the perceptual distance (jnds), calculated using the model of Vorobyev and Osorio (1998) (horizontal axis), and the Euclidean distance in triangular chromaticity space (sensu Kelber et al. 2003) (vertical axis) of 1000 randomly generated pairs of colors. (b) Five colors (black points, labeled A–E) plotted in a triangular chromaticity diagram, each surrounded by 2 contours representing the 3 jnds boundary (i.e., colors within a contour would have a perceptual distance from the corresponding color of <3 jnds, as calculated using the Vorobyev–Osorio model): both contours assume an equal standard deviation of noise in each photoreceptor of ei = 0.05 (inner contour) or ei = 0.1 (outer contour) (for details, see text). Contours were generated by finding the best-fitting ellipses (in the least-squares sense) to 1000 random colors that differed from the central color by 3 ± 0.01 jnds. The achromatic point is denoted by an “x” at the triangle's origin, and the 3 apices represent stimulation of a single 1 of the 3 cone classes contributing to color vision (labeled L, M, and S). Data for the visual system of the three-spined stickleback were used in generating the data in both plots; for full details, see text.

These examples illustrate that whether any 2 points in triangular (or indeed tetrahedral) chromaticity space are perceptually different cannot be reliably inferred directly from the chromaticity diagram itself, and so has implications for the geometric interpretation of color data within a chromaticity diagram. It would therefore be useful to have a chromaticity diagram in which the Euclidean distance between any 2 points, at any place in the chromaticity diagram, is directly proportional to the perceptual distance between them. The approach described here achieves this by deriving Cartesian coordinates for each color such that the Euclidean distance between any 2 points in this coordinate system is exactly equal to their perceptual distance in jnds, as calculated using the Vorobyev–Osorio model.

ALGORITHM

Values obtained using the Vorobyev–Osorio model describe perceptual distances between 2 colors (i.e., a single value of perceptual distance rather than 2 sets of coordinates, one for each color). In order to plot these data, spatial coordinates must be derived for each color based solely on the value of perceptual distance between each color pair. Perceptual distance data calculated by the Vorobyev–Osorio model allow the construction of such coordinates because any 3 colors can be arranged in a triangle such that length of each side is exactly equal to the jnd between the 2 connected colors at the triangle's vertices (i.e., if you know the jnds between colors A and B and between B and C, then the jnds between A and C can be determined geometrically). The algorithm used here determines the positions of each color in space using trilateration (i.e., determining position based on the geometry of triangles and circles or, in 3D space, spheres) and is described below. Matlab (The Mathworks, Natick, MA) and R (R Core Development Team) code for converting perceptual distances matrices into Cartesian coordinates is available as Supplementary Material.

Given a set of N colors (c1, c2, …, cN), represented in terms of the quantum catch of n photoreceptors (where n = 3 or 4), an N × N matrix D can be created in which each entry dij represents the perceptual distance (in jnds) between colors i and j (as determined following Vorobyev and Osorio’s 1998 model). Cartesian coordinates can be determined from the matrix of perceptual distances, D, as follows:

The first color, c1, is defined as having the Cartesian coordinates [x1, y1] = [0, 0]. (In practice, this point could have any arbitrary coordinates but placing it at the origin simplifies the calculations.)

As the distance between the first and second points is known from D, the coordinates of the second color, c2, can be defined as [x2, y2] = [d1,2, 0].

Using [x1, y1] and [x2, y2] as reference coordinates, the coordinates of the third color, c3, can be determined by finding the location in space at which the 2 distances d1,3 and d2,3 are satisfied, as:
x3=d1,32d2,32+d1,222d1,2,y3=±d1,32x32.
(7)

As there are always 2 potential solutions to this problem (because 2 circles centered on [x1, y1] and [x2, y2] will intersect at 2 discrete points), 1 of these 2 sets of coordinates, [x3, y3] or [x3, −y3], is chosen arbitrarily as the location of the third color.

If n = 3 (i.e., a trichromatic system), the data can be visualized in 2 dimensions (as in triangular chromaticity space), and so the 3 sets of coordinates [x1, y1], [x2, y2], and [x3, y3] are sufficient to calculate the position of all other colors in the data set (c4, …, cN), relative to these reference coordinates (using equation (7) and the distance from point [x3, y3] to determine which of the 2 solutions is correct). If n = 4 (i.e., a tetrachromatic system), the data must be visualized in 3 dimensions (as in tetrahedral chromaticity space). To calculate coordinates in 3D space, one can use the 3 known reference coordinates in the xy plane (i.e., [x1, y1, 0], [x2, y2, 0], and [x3, y3, 0]) plus an additional fourth reference point, [x4, y4, z4], which can be calculated from the coordinates of the first 3 points, as:
x4=d1,42d2,42+d1,222d1,2,y4=d1,42d3,42+y32+x322y3x4(x3y3),z4=±d1,42x42y42.
(8)

Again, the final coordinates must be chosen arbitrarily from the 2 possible solutions, [x4, y4, z4] or [x4, y4, −z4]. Simple modifications would also allow coordinates for dichromatic systems to be derived, and in principle, the technique could be expanded to derive coordinates for systems with 5 or more photoreceptors (e.g., cabbage white butterflies, Pieris rapae, have at least 6 cone types making them putative hexachromats or pentachromats; Morehouse and Rutowski 2010), although they could not be visualized in their entirety because they would require more than 3 dimensions (Pike 2011).

In the examples given in this paper (Figures 2 and 3), the points in color space have been translated so that the achromatic point (i.e., stickleback gray) is at the origin and rotated so that they are orientated similarly to the triangular chromaticity diagram in Figure 1b. Neither of these transformations is strictly necessary, as the color data could still be plotted with an arbitrary origin and orientation; however, these transformations not only aid interpretation of the color data and the limitations of visual system that is detecting it (see next section) but also allows direct comparisons to be made with the same data plotted in triangular chromaticity space. (Note that, if additional points in space are needed for construction of a diagram, such as the location of the achromatic point, then this will need to be included in the matrix D, so that the perceptual distance from this point to all the data points can be determined, as this cannot easily be computed retrospectively.)

Figure 2

Color data plotted in the perceptual chromaticity diagram described in this paper. Each subplot contains the same 100 randomly generated colors. Also in each subplot are the 5 data points shown in Figure 1b (black points), each surrounded by a contour representing the 3 jnds boundary (as in Figure 1b), and arrows representing each of the cone classes contributing to the chromatic discrimination (labeled L, M, and S accordingly, as in Figure 1b); the additional arrow in (d) represents the sticklebacks ultraviolet-sensitive (U) cone. Please refer to the text for a full explanation of these plots.

Figure 3

The 5 colors from Figure 1b (labeled A–E), plus an additional color, F, plotted in a perceptual color space that includes both chromatic and achromatic information (black points). The arrows and contours (ellipses) on the xy plane are exactly the same as those depicted in Figure 2c and therefore represent the relative chromatic information contained within each of the 6 colors. The z-axis represents the relative achromatic information contained within each color, such that the distance between colors along this axis is equal to their perceptual achromatic distance in jnds; the stems (dashed lines) simply allow the data points to be visualized more clearly in 3D space. Please refer to the text for a full explanation of this plot.

Caveat

Under the formulation used here, where cone responses are logged prior to the calculation of response differences (i.e., Δfi = log(QiA/QiB) = log(QiA) − log(QiB)), situations where levels of cone stimulation approach zero (e.g., for particularly saturated colors) can cause problems because log(Qi) is undefined when Qi = 0. This causes no problems when all cones receive a moderate amount of stimulation (which is often the situation in practice) and can be overcome completely by using the unlogged differences instead (i.e., Δfi = QiAQiB), although this may not always be appropriate (Osorio et al. 2004).

DESCRIPTION OF THE PERCEPTUAL CHROMATICITY DIAGRAM

Some examples of the proposed perceptual chromaticity diagram are given in Figure 2, modeled using data for three-spined sticklebacks (Rowe et al. 2004; see INTRODUCTION). This chromaticity diagram achieves the goal of preserving perceptual distances between colors (i.e., the Euclidean distance between any 2 colors is exactly equal to their perceptual distance, in jnds). The axes units are therefore in jnds, although the values themselves are necessarily arbitrary since the set of data points can be rotated or translated in any direction while still preserving the Euclidean distance between colors. Note that the range of the x- and y-axes remains constant between subpanels in Figure 2 in order to better illustrate how changing levels of noise in the photoreceptors affect the spatial distribution of color data (see below). Each subpanel also shows the same set of 100 randomly generated colors, as well as the same 5 colors (A–E) that are represented in triangular chromaticity diagram in Figure 1b, each surrounded by a contour representing 3 jnds.

In each subpanel of Figure 2, arrows have been included to represent each of the cone classes contributing to the chromatic discrimination. Colors that lie on along one of these arrows will, in human perceptual terms, have the same hue, and the arrows therefore represent variation in the relative stimulation of a single cone class (i.e., low stimulation of a single given cone class will result in the color point falling near the foot of the relevant arrow; high stimulation will result in the color falling near its head). The actual length of these arrows is arbitrary (here they have been set to point at [0.9, 0.1, 0.1], [0.1, 0.9, 0.1], and [0.1, 0.1, 0.9], where these triplets are the quantum catches of the L, M, and S cones, respectively, normalized to the background irradiance), but what is important is the relative change in the arrows’ lengths and interarrow angles between graphs, as this is directly related to the level and asymmetry of noise in each of the receptor channels, as described below.

If we assume that the standard deviation of noise in each receptor channel is equal (ei = 0.05 in Figure 2a and ei = 0.1 in Figure 2b), then the arrows are of equal length, and the angle between any 2 arrows is equal. This is because all the receptors provide an equal weighting in the determination of perceptual distance. However, as the noise level in the receptors increases (as from Figure 2a to Figure 2b), the relative length of the arrows decreases because the greater level of noise means that it is harder for the visual system to discriminate between 2 given colors (i.e., they are now separated by a smaller jnd), thereby “squeezing” the colors together in space. Note that the contour representing the 3 jnd boundary around a color is circular (cf. the ellipses in Figure 1b), and remains the same size irrespective of changes that occur in the noise of a photoreceptor (e.g., between Figure 2a and Figure 2b), because of the direct relationship between Euclidean and perceptual distances.

It is unlikely that in a given animal, the noise in all photoreceptor channels will in fact be equal, and usually, there will be an asymmetry in the level of noise between photoreceptor classes (Vorobyev and Osorio 1998). In this case, both the relative lengths of the arrows and the angles between them will be unequal. This is illustrated in Figure 2c, where the standard deviation of noise in the S, M, and L cones is assumed (for illustration purposes) to be eS = 0.08, eM = 0.06, and eL = 0.04, respectively. This has the effect of squeezing together (i.e., reducing the number of jnds between) colors along the S cone and, to a lesser extent, the M cone axes, and those colors that fall between these axes, whereas “stretching” (i.e., increasing the number of jnds between) colors that fall between the M and L cone axes. With this distribution of noise levels, it should therefore be easier for a stickleback to distinguish between reds and greens than between greens and blues. The geometry of these axes can therefore be used to infer aspects of the visual system, in terms of the noise in the color channels mediating perception, and can be used to directly compare visual systems experiencing different noise levels (e.g., as a result of changes in ambient light levels; Osorio et al. 2004).

Finally, a 3D version of this color space, for animals with 4 cone classes, can be drawn (Figure 2d), which, like the 2D version, can geometrically represent changes in receptor noise. The interpretation of this plot is in all respects identical to that described above because the z-axis units are also in jnds. The only difference is that there is an additional arrow representing the stickleback's ultraviolet-sensitive (U) cone (Rowe et al. 2004).

INCORPORATING LUMINANCE INFORMATION

When studies require the interpretation of a species’ ability to discriminate both chromatic and achromatic information, these are typically analyzed separately (e.g., Osorio et al. 2004). However, when an animal makes a visual discrimination, both the chromatic and achromatic aspects of a color are viewed and processed simultaneously. This underlies the ability of humans to discriminate between and categorize “colors” that are chromatically similar but differ in terms of luminance (brightness), such as orange and brown (brown is simply a dark orange) and red and pink (pink is a light red with reduced chroma) (e.g., Bartleson 1976). It is possible to add an independent achromatic axis to the perceptual chromaticity diagram described above, allowing relative variation in both the chromatic and the achromatic aspects of a color (or colors) to be visualized simultaneously (Figure 3).

In this example, I modeled luminance discrimination following Siddiqi et al. (2004), in which the difference between 2 stimuli (Δd, in jnds) is given by:
Δd=ΔfLeL,
(9)
where ΔfL is the difference between 2 stimuli for the L cones and eL is the receptor noise of the L cones, as calculated for the color signal (see above). This assumes that the L cone alone contributes to luminance perception in sticklebacks; whether this is the case or not is unknown but provides a plausible means of modeling luminance perception in this species (Osorio and Vorobyev 2005). Because I assume that a single cone type determines luminance perception (although this also applies if luminance were determined by the summed output of 2 or more cones, as in humans; Wyszecki and Stiles 1982), perceptual differences in luminance between colors can be conceptualized as lying along a straight line or, alternatively, as occupying a single axis in an n-dimensional graph (the z-axis in Figure 3, where n = 3). In Figure 3, the position of colors in 2 dimensions (along the xy plane) represents their perceived chromatic differences (in jnds) and contains exactly the same information as plotted in Figure 2c; their relative position along the z-axis represents their perceived achromatic differences (also in jnds); and their position in all 3 dimensions represents a composite of their perceived chromatic and achromatic differences. This is valid because there is no intrinsic curvature in the combined chromatic/achromatic color space. (Note that a value of 0 on the z-axis [e.g., point A] does not mean that a color patch has no luminance; the axis values are arbitrary, and what is important is the distance between points along the z-axis [i.e., the perceptual distance between them in jnds].)

Colors that are chromatically identical but differ in perceived luminance can now be visually differentiated (e.g., points C and F in Figure 3) and the resulting Cartesian coordinates used in statistical models (e.g., Endler and Mielke 2005), providing a test of overall perceptual differences between color stimuli. (Note that exactly the same procedure can be used to incorporate luminance information into a 4-cone perceptual color space, and thereby derive Cartesian coordinates for further analyses, although this information cannot be visualized directly because it requires more than 3 dimensions; Pike 2011.) However, because the achromatic and chromatic components of a color are probably processed via different neural mechanisms and are often used independently (e.g., animals may respond to variation in luminance independently of chromaticity and vice versa) (Osorio and Vorobyev 2005), this combined model may not always be appropriate.

CONCLUSIONS

I have described a simple method for converting perceptual distances between colors, as determined using the popular noise-corrected model of Vorobyev and Osorio (1998), into a graphical representation of colors in a chromaticity diagram that exactly preserves perceptual distances between them. This brings a number of benefits over the more commonly used triangular and tetrahedral chromaticity diagrams (sensu Kelber et al. 2003; Endler and Mielke 2005). First, it allows data that previously existed only as a matrix of perceptual distances to be assigned useful coordinates. This not only allows the data to be plotted and visualized but the coordinates themselves can be used directly in statistical analyses that require Cartesian coordinates as inputs (e.g., Endler and Mielke 2005). Second, it allows color data to be visualized in chromaticity diagrams from which meaningful insights can be gained by looking at the spatial positioning of the data points (e.g., to interpret the “direction” of the difference between 2 colors, which cannot be inferred from the absolute jnd value; Delhey et al. 2010). This allows for a more intuitive way of representing and interpreting color data that enhances the level of biological realism. Moreover, it allows the effects that changing receptor noise levels (e.g., when comparing low and high illumination conditions; Osorio et al. 2004), and the effect this has on chromatic discrimination, to be readily visualized as changes in the geometry of the data points. Third, with modifications, it is possible to visualize both the chromatic and the achromatic aspects of a perceived color within a single diagram, allowing inferences to be drawn regarding their composite perceptual distance from other colors and the derivation of composite spatial coordinates for use in statistical models (e.g., Endler and Mielke 2005).

Perceptual modeling has become very popular in studies of animal color vision and is an invaluable tool for helping to understand color vision–mediated behavior. However, these models only provide abstractions of real visual systems, which are rarely validated against behavioral data and are subject to limitations in our understanding of the particular visual system we are modeling. For example, although we have an increasingly good understanding of photoreceptor sensitivities for a wide range of species, we have a poorer understanding of the levels of noise present in these receptor mechanisms (prerequisites for using the Vorobyev–Osorio model). Care must therefore be taken when interpreting data from perceptual models, and it may be prudent to present a number of different visual scenarios before drawing firm conclusions (e.g., as has been done in studies which explore perception over a range of plausible receptor noise levels; Osorio et al. 2004; Langmore et al. 2009). Being able to graphically represent these different scenarios, using models such as the one described here, would therefore greatly aid their interpretation.

SUPPLEMENTARY MATERIAL

Supplementary material can be found at http://www.beheco.oxfordjournals.org/.

FUNDING

This work was supported by a Natural Environment Research Council fellowship (NE/F016514/2).

References

Andersson
S
Prager
M
Hill
GE
McGraw
KJ
Quantifying colors
Bird coloration volume I: mechanisms and measurements
2006
Cambridge (MS)
Harvard University Press
(pg. 
41
-
89
)
Backhaus
WGK
Backhaus
WGK
Reinhold
K
Werner
JS
Physiological and psychophysical simulations of color vision in humans and animals
Color vision: perspectives from different disciplines
1998
Berlin (Germany)
Walter de Gruyter
(pg. 
45
-
71
)
Barlow, HB. 1964. The physical limits of visual discriminations. In: Giese AC, editor. Photophysiology volume II. New York: Academic Press. p. 163–202
Barlow
GW
The attitude of fish eye-lines in relation to body shape and to stripes and bars
Copeia
1972
, vol. 
1972
 (pg. 
4
-
12
)
Bartleson
CJ
Brown
Color Res Appl
1976
, vol. 
1
 (pg. 
181
-
191
)
Boulcott
PD
Visual perception and ecology of the threespined stickleback [PhD thesis]
2003
Edinburgh (UK)
University of Edinburgh
Cassey
P
Portugal
SJ
Maurer
G
Ewen
JG
Boulton
RL
Hauber
ME
Blackburn
TM
Variability in avian eggshell colour: a comparative study of museum eggshells
PLoS One
2010
, vol. 
5
 pg. 
e12054
 
Chittka
L
The colour hexagon: a chromaticity diagram based on photoreceptor excitations as a generalized representation of colour opponency
J Comp Physiol A
1992
, vol. 
170
 (pg. 
533
-
543
)
Defrize
J
Théry
M
Casas
J
Background colour matching by a crab spider in the field: a community sensory ecology perspective
J Exp Biol
2010
, vol. 
213
 (pg. 
1425
-
1435
)
Delhey
K
Roberts
ML
Peters
A
The carotenoid-continuum: carotenoid-based plumage ranges from conspicuous to cryptic and back again
BMC Ecol
2010
, vol. 
10
 pg. 
13
 
Derrington
AM
Krauskopf
J
Lennie
P
Chromatic mechanisms in lateral geniculate nucleus of macaque
J Physiol
1984
, vol. 
357
 (pg. 
241
-
265
)
Eaton
MD
Human vision fails to distinguish widespread sexual dichromatism among sexually ‘monochromatic’ birds
Proc Natl Acad Sci U S A
2005
, vol. 
102
 (pg. 
10942
-
10946
)
Endler
JA
Mielke
PW
Comparing entire colour patterns as birds see them
Biol J Linn Soc
2005
, vol. 
86
 (pg. 
405
-
431
)
Govardovskii
VI
Fyhrquist
N
Reuter
T
Kuzmin
DG
Donner
K
In search of the visual pigment template
Vis Neurosci
2000
, vol. 
17
 (pg. 
509
-
528
)
Kelber
A
Vorobyev
M
Osorio
D
Animal colour vision: behavioural tests and physiological concepts
Biol Rev
2003
, vol. 
78
 (pg. 
81
-
118
)
Langmore
NE
Stevens
M
Maurer
G
Kilner
RM
Are dark cuckoo eggs cryptic in host nests?
Anim Behav
2009
, vol. 
78
 (pg. 
461
-
468
)
Loyau
A
Gomez
D
Moureau
BT
Thery
M
Hart
NS
Saint Jalme
M
Bennett
ATD
Sorci
G
Iridescent structurally based coloration of eyespots correlates with mating success in the peacock
Behav Ecol
2007
, vol. 
18
 (pg. 
1123
-
1131
)
MacLeod
DIA
Boynton
RM
Chromaticity diagram showing cone excitation by stimuli of equal luminance
J Opt Soc Am
1979
, vol. 
69
 (pg. 
1183
-
1186
)
Maxwell
JC
On the theory of compound colours, and the relations of the colours of the spectrum
Phil Trans R Soc Lond
1860
, vol. 
150
 (pg. 
57
-
84
)
Morehouse
NI
Rutowski
RL
In the eyes of the beholders: female choice and avian predation risk associated with an exaggerated male butterfly color
Am Nat
2010
, vol. 
176
 (pg. 
768
-
784
)
Osorio
D
Smith
AC
Vorobyev
M
Buchanan-Smith
HM
Detection of fruit and the selection of primate visual pigments for color vision
Am Nat
2004
, vol. 
164
 (pg. 
696
-
708
)
Osorio
D
Vorobyev
M
Photoreceptor spectral sensitivities in terrestrial animals: adaptations for luminance and colour vision
Proc R Soc B
2005
, vol. 
272
 (pg. 
1745
-
1752
)
Osorio
D
Vorobyev
M
Jones
CD
Colour vision of domestic chicks
J Exp Biol
1999
, vol. 
202
 (pg. 
2951
-
2959
)
Pike
TW
Generalised chromaticity diagrams for animals with n-chromatic colour vision
J Insect Behav
2011
 
doi:10.1007/s10905-011-9296-2
Pike
TW
Bjerkeng
B
Blount
JD
Lindström
J
Metcalfe
NB
How integument colour reflects its carotenoid content: a stickleback’s perspective
Func Ecol
2010
, vol. 
25
 (pg. 
297
-
304
)
Rovamo
JM
Kankaanpaa
MI
Hallikainen
J
Spatial neural modulation transfer function of human foveal visual system for equiluminous chromatic gratings
Vis Res
2001
, vol. 
41
 (pg. 
1659
-
1667
)
Rowe
MP
Baube
CL
Loew
ER
Phillips
JB
Optimal mechanisms for finding and selecting mates: how threespine stickleback (Gasterosteus aculeatus) should encode male throat colors
J Comp Physiol A
2004
, vol. 
190
 (pg. 
241
-
256
)
Siddiqi
A
Cronin
TW
Loew
ER
Vorobyev
M
Summers
K
Interspecific and intraspecific views of color signals in the strawberry poison frog Dendrobates pumilio
J Exp Biol
2004
, vol. 
207
 (pg. 
2471
-
2485
)
Vorobyev
M
Osorio
D
Receptor noise as a determinant of colour thresholds
Proc R Soc B
1998
, vol. 
265
 (pg. 
351
-
358
)
Wyszecki
G
Stiles
WS
Color science: concepts and methods, quantitative data and formulae
1982
New York
John Wiley & Sons

Supplementary data