Skip to main content
eLife logoLink to eLife
. 2021 Jun 8;10:e63355. doi: 10.7554/eLife.63355

Spherical arena reveals optokinetic response tuning to stimulus location, size, and frequency across entire visual field of larval zebrafish

Florian A Dehmelt 1,, Rebecca Meier 1,, Julian Hinz 1,†,, Takeshi Yoshimatsu 2, Clara A Simacek 1, Ruoyu Huang 1, Kun Wang 1, Tom Baden 2, Aristides B Arrenberg 1,
Editors: Kristin Tessmar-Raible3, Tirin Moore4
PMCID: PMC8233042  PMID: 34100720

Abstract

Many animals have large visual fields, and sensory circuits may sample those regions of visual space most relevant to behaviours such as gaze stabilisation and hunting. Despite this, relatively small displays are often used in vision neuroscience. To sample stimulus locations across most of the visual field, we built a spherical stimulus arena with 14,848 independently controllable LEDs. We measured the optokinetic response gain of immobilised zebrafish larvae to stimuli of different steradian size and visual field locations. We find that the two eyes are less yoked than previously thought and that spatial frequency tuning is similar across visual field positions. However, zebrafish react most strongly to lateral, nearly equatorial stimuli, consistent with previously reported spatial densities of red, green, and blue photoreceptors. Upside-down experiments suggest further extra-retinal processing. Our results demonstrate that motion vision circuits in zebrafish are anisotropic, and preferentially monitor areas with putative behavioural relevance.

Research organism: Zebrafish

Introduction

The layout of the retina, and the visual system as a whole, evolved to serve specific behavioural tasks that animals perform to survive in their respective habitats. A well-known example is the position of the eyes in the head which varies between hunting animals (frontal eyes) and animals that frequently need to avoid predation (lateral eyes) (Cronin et al., 2014). Hunting animals keep the prey within particular visual field regions to maximise behavioural performance (Hoy et al., 2016; Bianco et al., 2011; Yoshimatsu et al., 2020). To avoid predation, however, it is useful to observe a large proportion of visual space, especially those regions in which predators are most likely to appear (Smolka et al., 2011; Zhang et al., 2012). The ecological significance of visual stimuli thus depends on their location within the visual field, and it is paralleled by non-uniform processing channels across the retina. This non-uniformity manifests as an area centralis or a fovea in many species, which is a region of heightened photoreceptor density in the central retina and serves to increase visual performance in the corresponding visual field regions. Photoreceptor densities put a direct physical limit on performance parameters such as spatial resolution (Haug et al., 2010; Merigan and Katz, 1990). In addition to these restrictions mediated by the peripheral sensory circuitry, an animal’s use of certain visual field regions is also affected by behaviour-specific neural pathways, for example pathways dedicated to feeding and stabilisation behaviour. The retinal and extra-retinal circuit anisotropies can in turn effect a dependence of behavioural performance on visual field location (Hoy et al., 2016; Bianco et al., 2011; Murasugi and Howard, 1989; Shimizu et al., 2010; Yang and Guo, 2013; Zimmermann et al., 2018; Baden et al., 2013).

Investigating behavioural performance limits and non-uniformities can offer insights into the processing capabilities and ecological adaptations of animal brains, especially if they can be studied and quantitatively understood at each processing step. The larval zebrafish is a promising vertebrate organism for such an endeavour, since its brain is small and a wide array of experimental techniques are available (Baier and Scott, 2009; McLean and Fetcho, 2011). Zebrafish are lateral-eyed animals and have a large visual field, which increases during early development and in individual 4-day-old larvae has been reported to cover about 163° per eye (Easter and Nicola, 1996), although little is known about interindividual variability. Their retina contains four different cone photoreceptor types (Nawrocki et al., 1985), each distributed differently across the retina. UV photoreceptors are densest in the ventro-temporal retina (area temporalis ventralis), whereas the red, green, and blue photoreceptors cover more central retinal regions (Zimmermann et al., 2018).

Zebrafish larvae perform a wide range of visually mediated behaviours, ranging from prey capture (Trivedi and Bollmann, 2013; Mearns et al., 2020) and escape behaviour (Heap et al., 2018) to stabilisation behaviour (Kubo et al., 2014; Orger et al., 2008); however, the importance of stimulus location within the visual field for the execution of the respective behaviours has only recently been recognized and is still not well understood (Hoy et al., 2016; Zimmermann et al., 2018; Mearns et al., 2020; Kist and Portugues, 2019; Wang et al., 2020; Johnson et al., 2020; Lagogiannis et al., 2020).

During visually mediated stabilisation behaviours, such as optokinetic and optomotor responses, animals move their eyes and bodies, respectively, in order to stabilise the retinal image and/or the body position relative to the visual surround. The optokinetic response (OKR) consists of reflexively executed stereotypical eye movements, in which phases of stimulus ‘tracking’ (slow phase) are interrupted by quick phases (Figure 1a). In the quick phases, eye position is reset by a saccade in the direction opposite to stimulus motion. In humans, optokinetic responses are strongest in the central visual field (Howard and Ohmi, 1984). Furthermore, lower visual field locations of the stimulus evoke stronger OKR than upper visual field locations, which likely represents an adaptation to the rich optic flow information available from the structures on the ground in the natural environments of primates (Murasugi and Howard, 1989; Hafed and Chen, 2016).

Figure 1. Presenting visual stimuli across the visual field.

(a) When presented with a horizontal moving stimulus pattern, zebrafish larvae exhibit optokinetic response (OKR) behaviour, where eye movements track stimulus motion to minimise retinal slip. Its slow phase is interrupted by intermittent saccades, and even if only one eye is stimulated (solid arrow), the contralateral eye is indirectly yoked to move along (dashed arrow). (b) Often, experiments on visuomotor behaviour such as OKR sample only a small part of the visual field, whether horizontally or vertically. As different spatial directions may carry different behavioural importance, an ideal stimulation setup should cover all or most of the animal’s visual field. For zebrafish larvae, this visual field can be represented by an almost complete unit sphere. (c) We arranged 232 LED tiles with 64 LEDs each across a spherical arena, such that 14,484 LEDs (green dots) covered nearly the entire visual field. (d) The same individual positions, shown in geographic coordinates. Each circle represents a single LED. Each cohesive group of eight-by-eight circles corresponds to the 64 LEDs contained in a single tile. (e) To identify LED and stimulus locations, we use Up-East-North geographic coordinates: Azimuth α describes the horizontal angle, which is zero in front of the animal and, when seen from above, increases for rightward position. Elevation β refers to the vertical angle, which is zero throughout the plane containing the animal, and positive above. (f) The spherical arena is covered in flat square tiles carrying 64 green LEDs each. (g) Its structural backbone is made of a 3D-printed keel and ribs. Left and right hemispheres were constructed as separate units. (h) Across 85-90% of the visual field, we can then present horizontally moving bar patterns of different location, frequency and size to evoke OKR.

Figure 1—source data 1. SCAD files for 3D-printing the arena scaffold.

Figure 1.

Figure 1—figure supplement 1. A spherical LED arena to present visual stimuli across the visual field.

Figure 1—figure supplement 1.

(a) LED tiles are arranged in ribbons parallel to the equator and glued in between structural ribs. Gaps at the top and bottom pole of the sphere allow coupling in an optical pathway for infrared illumination and subsequent video recording of eye movement. The fish is placed at the centre of the sphere, facing the observer. See the Supplementary file 1A as well as the mathematical manual in our data repository for more on the angles βk. (b) Optical pathway for eye movement tracking. (c) To minimise obstruction and refraction, the zebrafish larva is immobilised on the tip of a glass triangle (left) using agarose, which is then inserted into the centre of a spherical glass bulb (middle). This bulb is then mounted into a metal holder (right) and thus placed at the centre of the sphere. (d) Image of the two hemispheres and the camera setup. One hemisphere is mounted on a rail to allow opening and closing the arena.

Figure 1—figure supplement 2. Nearest-neighbour distances between LEDs.

Figure 1—figure supplement 2.

Histograms of (a) the angle between an LED and its nearest neighbour, (b) the mean, (c) median and (d) maximum angle between an LED and its eight nearest neighbours, and (e) the mean, (f) median and (g) maximum angle between an LED and its 24 nearest neigbours. Whereas (a) only captures the microstructure of LED tiles, where 8 by 8 LED are arranged in a square pattern, (b-d) and especially (e-g) also capture edge effects and gaps in between tiles.

In zebrafish larvae, OKR behaviour has been used extensively to assess visual function in genetically altered animals (Brockerhoff et al., 1995; Muto et al., 2005; Neuhauss et al., 1999). OKR tuning to velocity, frequency, and contrast of grating stimuli has been measured (Clark, 1981; Cohen et al., 1977; Huang and Neuhauss, 2008; Rinner et al., 2005), and, more recently, zebrafish are used as a model for investigating vertebrate sensorimotor transformations (Kubo et al., 2014; Portugues et al., 2014). While zebrafish can distinguish rotational from translational optic flow to evoke appropriate optokinetic and optomotor responses (Kubo et al., 2014; Naumann et al., 2016; Wang et al., 2019), it is still unclear which regions of the visual field zebrafish preferentially observe in these behaviours. The aquatic lifestyle, in combination with the preferred swimming depths (Lindsey et al., 2010), might cause the lower visual field to contain less relevant information when compared to terrestrial animals. This in turn might result in behavioural biases to other – more informative – visual field regions. A corresponding systematic behavioural quantification in zebrafish or other aquatic species, which would relate OKR behaviour to naturally occurring motion statistics and the underlying neuronal representations in retina and retino-recipient brain structures, has been prevented by technical limitations. Specifically, little is known about (i) the dependence of OKR gain on stimulus location or (ii) on stimulus sizes, (iii) possible interactions between stimulus location, size and frequency, (iv) putative asymmetries between the left and right hemispheres of the visual field, and (v) the relationship between a putative dependence of OKR on stimulus location and zebrafish retinal architecture.

In other species with large visual fields, such as Drosophila, full-surround stimulation setups have been designed and used successfully (Reiser and Dickinson, 2008; Kim et al., 2017; Maisak et al., 2013), but to date, none has been used for fish. This is at least partly due to their aquatic environment and the associated difficulties regarding the refraction of stimulus light at the air-water interface. Such distortions of shape can be partially compensated by pre-emptively altering the shape of the stimulus. However, using regular computer screens or video projection, the resulting luminance profiles remain anisotropic unless compensated, potentially biasing the response toward brighter locations. Additionally, most stimulus arenas cannot easily be combined with the recording of neural activity, for example, via calcium imaging, as stimulus light and calcium fluorescence overlap in both the spectral and time domains. These challenges must be overcome to enable full-field visual stimulation in zebrafish neurophysiology experiments (Figure 1b). At least one existing solution for stimulating and tracking freely moving zebrafish supports unusually large stimuli, although despite its versatility, it still only covers part of the visual field and does not address the remaining issues of total internal reflection and the interoperability with laser-scanning microscopes (Stowers et al., 2017 ).

Here, we present a novel visual stimulus arena for aquatic animals, which covers almost the entire surround of the animal, and use it to characterise the anisotropy of the zebrafish OKR across different visual field locations as well as the tuning to stimulus size, spatial frequency and leftside versus rightside stimulus locations. We find that the OKR is mostly symmetric across both eyes and driven most strongly by lateral stimulus locations. These stimulus locations approximately correspond to a retinal region of increased photoreceptor density. By rotating the experimental setup and/or the animal, our control experiments revealed that additional extra-retinal determinants of OKR drive exist as well. Our characterisation of OKR drive across the visual field will help inform bottom-up models of the vertebrate neural pathways underlying the optokinetic response and other visual behaviour.

Results

Spherical LED arena allows presentation of stimuli across the visual field

By combining 3D printing with electronic solutions developed in Drosophila vision research, we constructed a spherical stimulus arena containing 14,848 individual LEDs covering over 90% of the visual field of zebrafish larvae (Figure 1c, Materials and methods section on coverage). Using infrared illumination via an optical pathway coupled into the sphere (Figure 1—figure supplement 1a–b), we tracked eye movements of larval zebrafish during presentation of visual stimuli (Dehmelt et al., 2018).

To avoid stimulus aberrations at the air-to-water interface, we designed a nearly spherical glass bulb containing fish and medium (Figure 1—figure supplement 1c–d). With this design, stimulus light from the surrounding arena is virtually not refracted (light is orthogonal to the air-to-water interface) and reaches the eyes of the zebrafish larva in a straight line. Thus, no geometric corrections are required during stimulus design (Source code 1), and stimulus luminance is expected to be nearly isotropic across the visual field. We additionally designed the setup to minimise visual obstruction and developed a new embedding technique to immobilise the larva at the tip of a narrow glass triangle (see Materials and methods). In almost all possible positions, fish can thus perceive stimuli without interference (Wang et al., 2021). The distance between most of the adjacent LED pairs is smaller than the photoreceptor spacing in the larval retina (Haug et al., 2010; Tappeiner et al., 2012), resulting in a good spatial resolution across the majority of the spherical arena surface (Figure 1—figure supplement 2, see Materials and methods section on resolution). As flat square LED tiles cannot be perfectly arranged on a spherical surface (Supplementary file 1A), small triangular gaps are unavoidable. More importantly, several gaps in LED coverage, resulting from structural elements of the arena, were restricted mainly to the back, the top, and bottom of the animal. The ‘keel’ behind and in front of the fish supports the horizontal ‘ribs’, and the circular openings in the top and bottom accommodate the optical path for eye tracking or scanning microscopy (also see discussion of arena geometry in our data repository [Dehmelt et al., 2020https://gin.g-node.org/Arrenberg_Lab/spherical_arena]).

Stimulus position dependence of the optokinetic response

Horizontally moving vertical bars reliably elicit OKR in zebrafish larvae (Beck et al., 2004). We used a stimulus which rotated clock- and counter-clockwise with a sinusoidal velocity pattern (velocity amplitude 12.5 degree/s, frequency of the velocity envelope 0.1 Hz, spatial frequency 0.06 cycles/degree, Figure 2a). OKR performance was calculated by measuring the amplitude of the resulting OKR slow-phase eye movements after the saccades had been removed (Figure 2b, Figure 2—figure supplement 1, Source code 1C, Materials and methods). The OKR gain then corresponds to the speed of the slow-phase eye movements divided by the speed of the stimulus (which is equivalent to the ratio of the eye position and stimulus position amplitudes). In addition to traditional full-field stimulation, our arena can display much smaller stimuli in different parts of the visual field. These novel stimuli evoked reliable OKR even at remote stimulus locations (Figure 2c, Figure 2—figure supplement 2), and thus allowed us to investigate previously inaccessible behavioural parameters. We excluded any trials from data analysis that showed other behaviours (e.g. drifts and ongoing spontaneous eye movements) superimposed on OKR (Figure 2—figure supplement 2). In addition to the characteristic eye traces (Figure 2c), Bode plots of the magnitude and phase shift relative to the stimulus qualitatively match previously reported zebrafish response to full-field stimulation (Figure 5—figure supplement 1, Beck et al., 2004), further confirming that the behaviour we observe is indeed OKR.

Figure 2. OKR gain is inferred from a piece-wise fit to the slow phase of tracked eye movements.

(a) We present a single pattern of horizontally moving bars to evoke OKR and crop it by superimposing a permanently dark area of arbitrary size and shape (left). Its velocity follows a sinusoidal time course, repeating every 10 s for a total of 100 s for each stimulus phase (right). (b) OKR gain is the amplitude of eye movement (green trace) relative to the amplitude of the sinusoidal stimulus (grey trace). The OKR gain is often well below 1, e.g. for high stimulus velocities as used here (up to 12.5°/s). (c) Even small stimuli are sufficient to elicit reliable OKR, although gains are low if stimuli appear in a disfavoured part of the visual field. Shown here are responses to a whole-field stimulus (top) and to a disk-shaped stimulus in the extreme upper rear (bottom, Figure 2—figure supplement 2).

Figure 2.

Figure 2—figure supplement 1. Saccades are cropped before OKR gain is measured.

Figure 2—figure supplement 1.

(a) As shown in Figure 2b, optokinetic gain is the ratio between the amplitudes of eye movement and stimulus movement. To extract these, raw eye traces must be processed. (b) OKR eye movements consist of a slow phase, gradually tracking stimulus motion, and intermittent saccades. (c) After pre-processing data to detect and remove saccades, we fit a piece-wise sinusoidal function with a single amplitude to the remaining slow-phase eye traces. The amplitude of the best fit determines OKR gain.
Figure 2—figure supplement 2. Even at remote stimulus locations, fish exhibit reliable OKR behaviour; trials with mixed behaviour are excluded.

Figure 2—figure supplement 2.

(a–d) show 60 s samples of eye position traces obtained during visual stimulation (dark grey traces). Saccades were cropped, as demonstrated in Figure 2—figure supplement 1. Positive angles correspond to more rightward eye positions. Coloured traces show best fit to data (see Materials and methods), along with the resulting OKR gain. Fish were presented with either (a) whole field visual stimuli, or (b–d) smaller disk-shaped stimuli, as in Figure 3. (a) Whole-field stimulation elicits reliable OKR, even though high stimulus frequencies keep its gain below 1. (b) The same is true for disk stimulus in a preferred region of the visual space (stimulus D7, azimuth −90°, elevation +39°, that is lateral and dorsal to the fish), although the smaller stimulus evokes a smaller OKR gain. (c) Disk stimuli presented in a disfavoured part of the visual field (stimulus D3, azimuth −90°, elevation +74°, almost dorsal to the fish) evoke significantly lower gains, but still reliable OKR behaviour. Data in (a,c) correspond to trials shown in Figure 2c. (d) As it is possible to encounter behaviours other than OKR, or in addition to OKR, we inspected all trials before data analysis. Trials in which no pure OKR was present, such as the one shown here, were excluded (stimulus D3, same as in c).

To quantify position tuning, we cropped the presented gratings (Figure 2a) to a disk-shaped area of constant size, centred on one of 38 nearly equidistant parts of the visual field (Figure 3a, Supplementary File 1B, Video 1, Figure 3—video 1). The distribution of positions was symmetric between the left and right, upper and lower, as well as front and rear hemispheres, with some stimuli falling right on the edge between two hemispheres. As permanent asymmetries in a stimulus arena or in its surroundings could affect OKR gain, we repeated our experiments in a second group of larvae after rotating the arena by 180 degrees (Figure 4a–b), then matched the resulting pairs of OKR gains during data analysis (see Materials and methods, Figure 4e–f). Any remaining asymmetries in the OKR distributions should result from biological lateralisation, if indeed there exists such consistent lateralisation across individuals.

Figure 3. OKR gain depends on stimulus location.

(a) The stimulus is cropped to a disk-shaped area 40 degrees in diameter, centred on one of 38 nearly equidistant locations (Supplementary file 1B) across the entire visual field (left), to yield 38 individual stimuli (right). (b–d) Dots reveal the location of stimulus centres D1-D38. Their colour indicates the average OKR gain across individuals and trials, corrected for external asymmetries. Surface colour of the sphere displays the discretely sampled OKR data filtered with a von Mises-Fisher kernel, in a logarithmic colour scale. Top row: OKR gain of the left eye (b), right eye (d), and the merged data including only direct stimulation of either eye (c), shown from an oblique, rostrodorsal angle. Bottom row: same, but shown directly from the front. OKR gain is significantly higher for lateral stimulus locations and lower across the rest of the visual field. The spatial distribution of OKR gains is well explained by the bimodal sum of two von-Mises Fisher distributions. (e) Mercator projections of OKR gain data shown in panels (b–d). White and grey outlines indicate the area covered by each stimulus type. Numbers indicate average gain values for stimuli centred on this location. Red dots show mean eye position during stimulation. Dashed outline and white shading on panels (b, d, e) indicate indirect stimulation via yoking, that is, stimuli not directly visible to either the left or right eye. Data from n = 7 fish for the original configuration and n = 5 fish for the rotated arena.

Figure 3—source data 1. Numerical data and graphical elements of Figure 3b–d.
Figure 3—source data 2. Numerical data and graphical elements of Figure 3e.

Figure 3.

Figure 3—figure supplement 1. Gaps in the arena do not bias OKR behaviour.

Figure 3—figure supplement 1.

(a) Artificial triangular holes setup. (b) Artificial keel setup. (c) Neither triangular nor elongated gaps result in significantly different OKR gains (n = 5 fish).
Figure 3—figure supplement 2. During OKR, the beating field and average eye position are independent of stimulus location.

Figure 3—figure supplement 2.

All data were pooled across fish and trials. One gain value was computed per stimulus presentation. Violin plots show distribution of mean horizontal eye positions across the pooled data; vertical lines indicate 25th percentile, median and 75th percentile of the distribution. Positions are those during presentation of stimulus types (a) D1 to D38 shown in Figure 3a, (b) A1 to A42 shown in Figure 5b, (c) F1 to F42 shown in Figure 5a. Dashed lines in (c) represent axis limits of (a,b).
Figure 3—figure supplement 3. Vertical eye position under upright and upside-down embedding.

Figure 3—figure supplement 3.

(a) Five dpf and six dpf larvae were embedded in agarose with their eyes cut free and placed under a microscope. Using an additional mirror, we recorded simultaneous image time series along both the dorsoventral and mediolateral axes. Vertical eye position was determined geometrically for each individual frame (see Materials and methods). Because some larvae were embedded in such a way that mediolateral axis was not entirely aligned with the true horizon of the environment, we measured vertical eye position (b) relative to both the true environmental horizon, and (c) the mediolateral body axis, and in both cases, compared the left (L) and right eyes (R) of larvae embedded upside-up (uu) or upside-down (ud). To facilitate comparison, positive signs were chosen to roughly correspond to the dorsal hemisphere in both cases. Bars indicate mean after pooling across all frames, error bars show corresponding standard error of the mean (s.e.m.). In summary, larval eyes were almost always inclined towards the dorsum, irrespective of the direction of embedding. The fish do not appear to compensate for their orientation with respect to the gravitational axis.
Figure 3—figure supplement 4. Yoking indices are biased by reflections within the arena.

Figure 3—figure supplement 4.

Yoking indices were computed for experiments using the regular setup as in Figure 4a (light grey), with a rotated arena as in Figure 4b (dark grey), corrected for experimental asymmetries as in Figure 3 (green), and with one side of the glass bulb, contralateral to the stimulus centre, painted black (black). Yoking indices from most experiments are close to zero, indicating similar OKR gains for both eyes regardless of stimulus location. In contrast, yoking indices from the latter control experiment differ markedly from zero, indicating significantly weaker responses by the respectively unstimulated eye. This finding points to reflections at the air-glass interface being visible to the purportedly ‘unstimulated’ eye. Stimulus types are listed in the order given in Supplementary file 1B and Supplementary file 1E: the whole-field stimulus H1, hemifield stimuli H2 to H7, and then disk-shaped stimuli D1 to D38 from front to rear, top to bottom, and left hemisphere to right hemisphere.
Figure 3—figure supplement 5. Across different types of arenas, stimulus reflections affect perceived yoking.

Figure 3—figure supplement 5.

Control experiments conducted in a rectangular stimulus arena. OKR-inducing gratings were shown on all four stimulus screens surrounding the larvae, while additional elements were introduced around either the left eye (LE), the right eye (RE) or neither. Specifically, selected eyes were (a,b) shown stationary stimuli of the same frequency and contrast as the moving stimuli, (c,d) shown a blank white surface, or (e,f) shielded with a fully opaque cover. See Supplementary file 1D for an overview of stimulus combinations. (a,c,e,g) Bars indicate mean OKR gains, and error bars show standard error of the mean. (b,d,f,h) Yoking indices are near zero when both eyes move with identical amplitudes and positive when left eye amplitude exceeds that of the right eye (see Materials and methods). (a,b) In the presence of two conflicting stimuli (moving vs. stationary), yoking between the eyes is reduced by almost half, confirming that the unstimulated eye is yoked to the stimulated eye, albeit with a lower OKR gain. (c,d) When there is no conflicting stimulus, yoking drives OKR of the contralateral eyes, albeit with a lower amplitude as if both eyes were stimulated directly with identical stimulus, (e,f) which is equally true in the presence of shielding. (g,h) To assess the effect of reflections on the difference between directly stimulated and purportedly stimulated eyes, we compare blank stimuli (as in c-d, which could diffusely reflect light) to fully shielded eyes (as in e-f, where no reflections should occur). There are no significant differences, indicating that the larger effect of reflections observed in our spherical arena (Figure 3—figure supplement 4) may be caused by the different stimulus design or by different reflection properties. *Two rectangular arena setups were used for the control experiments. Asterisks indicate data obtained from the second setup, for which balanced illumination was explicitly confirmed via diode photodetector. Data from n = 22 fish for initial setup and n = 10 fish for second setup.
Figure 3—video 1. Animation showcasing short samples of all disk stimuli used to study location dependence, as in Figure 3a.
Download video file (13.8MB, mp4)

Figure 4. The OKR is biased towards upper environmental elevations irrespective of fish orientation.

(a) Regular arena setup. (b) Arena can be tilted 180 degrees so front and rear, upper and lower LED positions are swapped. The bulb holder moves accordingly, so from the perspective of the fish, left and right, upper and lower LEDs are swapped. (c) Upside-down embedding setup. (d) Setup with inverted optical path, including illumination. (e–h) Results in body-centred coordinates, where positive elevations refer to dorsal positions, for the four setups shown in (a–d). As in Figure 3b–e, colour indicates the discretely sampled OKR data filtered with a von Mises-Fisher kernel and follows a logarithmic colour scale. (g,h) Experiments with presentation of a less regularly distributed set of stimuli, cropped to disks of 64 degrees polar angle instead of the 40 degrees used in (a,b,e,f). (g) Fish embedded upside-down exhibit a slight preference for stimuli below the body-centred equator, that is, positions slightly ventral to their body axis. (h) Fish embedded upright, as in (a). To account for environmental asymmetries such as arena anisotropies, we combined the data underlying (e) and (f) to obtain Figure 3b–e (see Materials and methods). Data from (e) n = 7, (f) n = 5, (g) n = 3, (h) n = 10 fish.

Figure 4—source data 1. Numerical data and graphical elements of Figure 4e.
Figure 4—source data 2. Numerical data and graphical elements of Figure 4f.
Figure 4—source data 3. Numerical data and graphical elements of Figure 4g.
Figure 4—source data 4. Numerical data and graphical elements of Figure 4h.

Figure 4.

Figure 4—figure supplement 1. Asymmetries between left and right eye are strongly affected by the environment.

Figure 4—figure supplement 1

Differences between the OKR gains of directly stimulated left eyes and directly stimulated right eyes can be explained by a linear combination of biological and environmental factors, for example, biases of individual animals b3,k or asymmetries of the stimulus arena (see Materials and methods). Comparing data from the regular and rotated setups (Figure 4e–f), as well as data from fish immobilised upside-down (Figure 4g), including additional stimuli covering entire hemispheres (Supplementary file 1E), we can infer the underlying factors via multivariate regression of our linear model. We find that individual biases (grey) vary strongly from fish to fish and are broadly distributed from left to right. There are some constant biases across fish (green), both towards the left side of their visual field (b2) and towards one of the two LED hemispheres (b1); however, these biases are small and, given the large variability of individual biases, might be a result of the limited number of fish studied. Underlying stimulus types are listed in Supplementary file 1B and Supplementary file 1E. (b) Histogram of individual-fish biases shows no collective preference for either side.
Figure 4—figure supplement 2. OKR gain maps for individual larvae.

Figure 4—figure supplement 2.

Mercator projections of OKR gain data as in Figure 3e and Figure 4e–h, but shown here for individual fish in (a) the regular setup, (b) the rotated arena. Data from n = 1 fish per panel. The results of the first fish in (b) are so irregular as to have likely been caused by experimenter error in preprocessing, rather than biologically meaningful differences. We thus excluded this fish from any analyses, such as the those for Figure 3 and Figure 4.
Figure 4—figure supplement 3. Elevation of stimuli evoking strongest OKR remains upward regardless of embedding direction.

Figure 4—figure supplement 3.

See Materials and methods for a detailed description of the permutation test. Showing the actual observation (red) and all 119 non-redundant permutations (blue) that maintain relative group size, p<0.05. ‘Dorsalward’ refers to the dorsal hemisphere in a body-centred reference frame, as opposed to the ‘upper’ hemisphere in an environmental reference frame. Before permutation, label one corresponds to the experiments shown in Figure 4a, whereas label two corresponds to those in Figure 4c. The test statistic is the signed distance from the black diagonal, with results in the white area being less extreme than those of the unpermuted observation. The null-hypothesis is rejected, so the elevation of OKR gain peaks within the visual field is not determined in body-centred coordinates, but instead biased towards upper elevations (‘towards the sun’, world-centred coordinates).

Video 1. Repulsion-based algorithm to numerically distribute stimulus centres across a sphere surface (Source code 1B), with gradual convergence on a logarithmic timescale (blue to green).

Download video file (213.7KB, mp4)

A subset of stimulus centres is held on the equator at zero elevation (orange).

To overcome our spatially discrete sampling, we then fit our data with a bimodal function comprised of two Gaussian-like two-dimensional distributions on the stimulus sphere surface (see Materials and methods, Source code 1D), to determine the location of highest OKR gain evoked by ipsilateral stimuli and contralateral stimuli, respectively. We observed significantly higher OKR gains in response to nearly lateral stimuli, and lower gains across the rest of the visual field (Figure 3b–e). OKR was strongest for stimuli near an azimuth of −82.5 degrees and an elevation of 5.1 degrees for the left side (in body-centred coordinates), as well as 81.7 and 1.6 degrees for the right side – slightly rostral of the lateral meridian, and slightly above the equator. In the nasal visual field (binocular overlap) the OKR gain was relatively high, but still lower than for lateral visual field locations. Note that due to the fast stimulus speeds, the absolute slow phase eye velocities were high, while the OKR gain was relatively low. We chose such high stimulus speeds to minimise the experimental recording time needed to obtain reliable OKR measurements for each visual field location.

As our stimulus arena is not completely covered by LEDs (Figure 1c, Figure 1d), some areas remain permanently dark. These could interfere with the perception of stimuli presented on adjacent LEDs. This is especially relevant as LED coverage is almost perfect for some stimulus positions (near the equator), whereas the size of triangular holes increases at others (towards the poles). We thus performed control experiments comparing the OKR gain evoked by a stimulus in a densely covered part of the arena to the OKR gain evoked by same stimulus, but in the presence of additional dark triangular patches (Figure 3—figure supplement 1a). We found no significant difference in OKR gain (Figure 3—figure supplement 1c, t-test, p>0.05). Additionally, we performed another series of control experiments using a dark shape mimicking the dark structural elements, the front ‘keel’ of the arena (Figure 3—figure supplement 1b). Again, we found no difference in OKR gain (Figure 3—figure supplement 1c, t-test, p>0.05), and thus ruled out that position dependence data was corrupted by incomplete LED coverage. Since the eyes were moving freely in our experiments, the range of eye positions during OKR, or so-called beating field (Schaerer and Kirschfeld, 2000), could have changed with stimulus position. We found that animals instead maintained similar median horizontal eye positions (e.g. left eye: −83.7 ± 1.8 degrees, right eye: 80.3 ± 1.9 degrees, average median ± standard deviation of medians, n = 7 fish, Figure 3—figure supplement 2) even for the most peripheral stimulus positions.

A priori, it is unclear whether the sampling preference originates from the peculiarities of the sensory periphery in the eye, or the behavioural relevance inferred by central brain processing. The former would establish stimulus preference based on its position relative to the eye and, by extension, its representation on specific parts of the retina (i.e. an eye-centered reference frame). The latter would establish stimulus preference in body- or head-centered reference frames or based on stimulus positions relative to environment (a world-centered reference frame). A world-centered reference frame is useful, if stimuli in different environmental locations have different behavioural relevance (such as a predator approaching from the water surface). To start distinguishing these possible effects in the context of OKR, as well as to reveal any stimulus asymmetries accidentally introduced during the experiment, we performed control experiments with larvae embedded upside-down (i.e. with their dorsum towards the lower pole of the arena, Figure 4c). Note that such an upside-down state can occur when the fish loses balance in natural behaviour and would normally be counteracted by the animals’ self-righting reflex. As a result of upside-down embedding, world-centred and fish-centred coordinate systems were no longer identical, in that ‘up’ in one is ‘down’ in the other, and ‘left’ becomes ‘right’. See Supplementary file 1C for a summary of these changes across experiments, and Materials and methods for notes on head- and retina-centred coordinate systems. To facilitate comparisons across embedding types, all positions from here onwards are given relative to the visual field, and thus in fish-centred coordinates. Unexpectedly, the elevation of highest OKR gains relative to the fish changed from slightly above to slightly below the equator of the visual field when comparing upright to inverted fish (Figure 4e, Figure 4g): When upright, azimuths and fish-centred elevations of the peaks of the best fits to data were −78.7° and 8.2° for the left eye, as well as 81.8° and 3.1° for the right eye. When inverted, −88.0° and 1.6° for the left eye, as well as 82.8° and −15.5° for the right eye. These numbers were obtained from the gains of those eyes to which any given stimulus was directly visible. The results from Figure 4e–f were combined to correct Figure 3 for external asymmetries (Materials and methods); this is why the best-fit position reported for Figure 4e alone differs slightly from that reported above for Figure 3b–e. Because the set of visual stimuli presented to inverted fish stemmed from an earlier stimulus protocol with less even sampling of the visual field, a slight scaling of azimuths and elevations is expected. The consistent reduction of the elevation, however, is not. We performed a permutation test in which embedding-direction labels were randomly swapped while stimulus-location labels were maintained, and the Gaussian-type fit to data was then repeated on each permuted dataset. This test confirmed that fish preferred upward (in the environmental reference frame) rather than dorsalward elevations (p<0.05, Source code 1E, Figure 4—figure supplement 3). While the fit peaks capture its centres of mass, the apparent maxima of the OKR spatial distribution are even further apart, with their elevation flipping signs from the dorsal to the ventral hemisphere (shown in Figure 4e,g).

Adjustment by the fish of its vertical resting eye position between the upright and inverted body positions would have been a simple potential explanation for this result. However, time-lapse frontal microscopy images (Materials and methods) ruled this out, since for both upside-up and upside-down embedding the eyes were inclined by an average of about four degrees towards the dorsum (3.5 ± 1.0° for the left eye, 4.9 ± 0.8° for the right eye, mean ± s.e.m., Figure 3—figure supplement 3). We also tested the influence of camera and infrared light (840 nm) positions (Figure 4d) – which in either case should have been invisible to the fish (Shcherbakov et al., 2013) – and found that they could indeed not explain the observed differences (Figure 4h). In summary, the body-centred preferred location only flipped from slightly dorsal to slightly ventral in upside-down embedded fish (Figure 4g), and thus remained virtually unchanged in environmental coordinates across all control experiments.

Eye-, head-, or body-centred reference frames are therefore not sufficient to fully explain the observed optokinetic stimulus location preferences. Instead, the stimulus location preference is additionally related to the environmental reference frame, suggesting that the behavioural relevance of our motion stimuli depends on their environmental elevation positions.

Yoking of the non-stimulated eye

Almost all stimuli were presented monocularly – that is, in a position visible to only one of the two laterally located eyes. Without exception, zebrafish larvae responded with yoked movements of both the stimulated and unstimulated eye. To rule out reflections of stimuli within the arena, we performed a series of experiments in which the unstimulated side of the glass bulb had been covered with a matte, black sheet of plastic. Reflections on the glass-air interface would otherwise cause monocular stimuli (that should only be visible to the ipsilateral eye) to also be seen by the contralateral eye. Yoking indices (YI) were significantly different between the regular monocular setup (YI≈0.2) and the control setup (YI≈0.7) containing the black surface on the side of the unstimulated eye, confirming that yoking indices had been affected by reflections (Figure 3—figure supplement 4, an index of 1 indicating completely monocular eye movements, an index of 0 perfectly conjugate eye movements/yoking). This suggests that previously reported yoking partially depends on reflections of the stimulus pattern at the glass-to-air or water-to-air interface in our spherical setup and other commonly used stimulus arenas. We performed additional control experiments using a previously described setup (Dehmelt et al., 2018) with four flat LCD screens for stimulus presentation in a different room. In these experiments, stimuli (Supplementary file 1C) were presented monocularly or binocularly, and the unstimulated eye was either (i) stimulated with a stationary grating (Figure 3—figure supplement 5a–b), (ii) shielded with a blank, white shield placed directly in front of the displays (Figure 3—figure supplement 5c–d), or (iii) shielded with a matte, black sheet of aluminium foil placed inside the petri dish (control for possible reflections on the Petri dish wall) (Figure 3—figure supplement 5e–f). This experiment showed that yoking was much reduced (YI≈0.3) if the non-stimulated eye saw a stationary grating (i) instead of the white or black shields (ii-iii, YI≈0.1) or a binocular motion stimulus (YI≈0) (Figure 3—figure supplement 5g–h, p<0.05).

Spatial asymmetries

As a few previous studies suggested left-right asymmetries in zebrafish visuomotor processing and behaviour other than OKR (Andrew et al., 2009; Watkins et al., 2004Sovrano and Andrew, 2006) we computed an asymmetry index B (Materials and methods) to reveal whether zebrafish OKR is lateralised in individuals or across the population. We did not observe a general asymmetry between the response of the left and right eyes. Rather, our data is consistent with three distinct sources of asymmetry: individual bias towards one eye, shared bias across individuals, and asymmetries induced by the environment (including the experimental setup and stimulus arena). Through multivariate linear regression, we fit a linear model of asymmetries to our data (Materials and methods), which combined data from fish embedded upside-up (Figure 4e), upside-down (Figure 4g) and data obtained with the arena rotated relative to the fish (Figure 4f), and included whole-field and hemispheric stimuli (Supplementary file 1D). Regression coefficients for external causes of asymmetry were similar to those for biological causes (Figure 4—figure supplement 1), and individual biases from fish to fish were broadly and symmetrically distributed from left to right (mean coefficient 3.710-4±120.010-4 st. dev., n = 15), so that no evidence was found for a strong and consistent lateralisation of OKR behaviour across animals (Figure 4—figure supplement 2).

Our results show that the OKR behaviour is mostly symmetric across both eyes, with individual fish oftentimes having a dominant eye due to seemingly random bias for one eye (lateralisation) across fish. Some of the observed asymmetries are consistent with external factors. Therefore, the OKR gains presented in Figure 3 have been corrected to present only biologically meaningful differences (Materials and methods).

Spatial frequency tuning of the optokinetic response is similar across visual field locations

We investigated the spatial frequency tuning of OKR behaviour across visual field positions by presenting seven different spatial frequencies of the basic stimulus, each cropped into a planar angle of 40 degrees, at different visual field locations (Figure 5a). We held the temporal frequency constant, therefore stimulus velocity decreased whenever spatial frequency increased. These seven disk-shaped stimuli were presented while centred on one of 6 possible locations in different parts of the visual field (Figure 5c), with three locations on each hemisphere: one near the location of highest OKR gain as determined in our experiments on position dependence, one in a nasal location, and one in a lower temporal location. In total, we thus presented 42 distinct types of stimuli (Supplementary file 1F, Figure 5—video 1). For each stimulus location and eye, the highest OKR gain was observed at a spatial frequency of 0.03 to 0.05 cycles/degree (Figure 5d, Figure 5—figure supplement 2). We did not observe any strong modulation of frequency dependence by stimulus location.

Figure 5. OKR tuning to spatial frequency is similar across different visual field locations.

(a) Patterns with seven different frequencies were cropped to disks of a single size. These disks were placed in six different locations for a total of 42 stimuli. cpd: cycles per degree. (b) Patterns with identical spatial frequencies were cropped to disks of seven different sizes. These disks were also placed in six different locations for another set of 42 stimuli. Degrees indicate planar angles subtended by the stimulus outline, so 360° correspond to whole-field stimulation. (a, b) Displaying the entire actual pattern at the size of this figure would make the individual bars hard to distinguish. We thus only show a zoomed-in version of the patterns in which 45 out of 360 degrees azimuth are shown. (c) Coloured dots indicate the six locations on which stimuli from a and b were centred, shown from above (top), from front (middle), and from an oblique angle (bottom). (d) OKR gain is unimodally tuned to a wide range of spatial frequency (measured in cycles per degree). (e) OKR gain increases sigmoidally as the area covered by the visual stimulus increases logarithmically (a stimulus size of 1 corresponds to 100% of the spherical surface). (d–e) Colours correspond to the location of stimulus centres shown in (c). There is no consistent dependence on stimulus location of either frequency tuning or size tuning. Error bars show standard error of the mean. Data from n = 7 fish for frequency dependence and another n = 7 fish for size dependence.

Figure 5—source data 1. Numerical data and graphical elements of Figure 5.

Figure 5.

Figure 5—figure supplement 1. Magnitude and phase shift of eye movements at different frequencies resemble those previously observed for zebrafish OKR.

Figure 5—figure supplement 1.

(A) Bode plots showing OKR gain (solid lines) and phase shift relative to the stimulus (dashed lines). Gains and phase shifts decrease as temporal frequency increases. Each line represents the mean across trials for a specific stimulus location, pooled across all fish, all trials and both eyes. Colours match the stimulus locations identified in Figure 5c. (B) Our observations qualitatively, but not quantitatively, match those reported for traditional full-field stimulus paradigms. Solid blue circles represent OKR phase shifts for approximately 6.4 dpf zebrafish larvae reported by Beck et al., 2004. The blue line represents our own mean phase shift across all stimulus locations shown in (A), and the light blue envelope shows the standard deviation across locations. Interestingly, our phase shifts are more in line with those reported for older fish (approximately 33.8dpf larvae, open blue circles) in Beck et al., 2004. Solid orange circles show scaled OKR gain for approximately 6.4dpf zebrafish larvae, as reported by Beck et al., 2004. Direct comparison to our gain data could be misleading because of the smaller size and higher velocity of our stimuli. Based on our results reported in Figure 5e, the size difference alone should account for a factor of about 5. Orange circles thus show the literature gain data scaled 0.2x. The black line and orange envelope represent our mean OKR gain and standard deviation across all stimulus locations shown in (A). Quantitative changes are consistent with the notion that smaller and faster stimuli evoke equally reliable, but lower-amplitude OKR behaviour (Figure 2c).
Figure 5—figure supplement 2. OKR gain of individual larvae, for different stimulus frequencies.

Figure 5—figure supplement 2.

Same colours as in Figure 5. Error bars show standard error of the mean. Data from n = 7 fish.
Figure 5—figure supplement 3. OKR gain of individual larvae, for different stimulus sizes.

Figure 5—figure supplement 3.

Same colours as in Figure 5. Error bars show standard error of the mean. Data from n = 7 fish.
Figure 5—video 1. Animation showcasing short samples of all disk stimuli used to study frequency dependence, as in Figure 5a.
Download video file (18.6MB, mp4)
Figure 5—video 2. Animation showcasing short samples of all disk stimuli used to study size dependence, as in Figure 5b.
Download video file (52.2MB, mp4)

Size dependence of the optokinetic response

It is unclear to what extent small stimuli are effective in driving OKR. We therefore employed a stimulus protocol with seven OKR stimuli presented in differently sized areas on the sphere (Figure 5b). Spatial and temporal frequencies were not altered, so bars appeared with the same width and velocity profile in all cases. These seven disk-shaped stimuli were presented while centred on one of 6 possible locations, identical to those used to study frequency dependence (Figure 5c), again yielding 42 unique stimuli (Supplementary file 1G, Figure 5—video 2). Stimulus area size was chosen at logarithmic intervals, ranging from stimuli almost as small as the spatial resolution of the zebrafish retina, to stimuli covering the entire arena. Throughout this paper, the term ‘stimulus size’ refers to the fractional area of the sphere surrounding the fish, in which the moving grating stimuli was presented. For instance, with a solid angle of 180°, the stimulus size is 50% and covers half of the surrounding space (Figure 5b). In line with many other psychophysical processes, OKR gain increased sigmoidally with the logarithm of stimulus size (Figure 5e, Figure 5—figure supplement 2). Weak OKR behaviour was already observable in response to very small stimuli of 0.8% (solid angle or ‘stimulus diameter’ of 10.4°), and reached half-maximum performance at a stimulus size of roughly 25% (solid angle: 120°). As was the case for spatial frequency dependence, we did not observe any strong modulation of size dependence by the tested stimulus locations, although OKR gains of the left eye appeared more dependent on stimulus location than those of the right eye. Specifically, the left eye responded more strongly to ipsilateral than contralateral stimuli (ANOVA, p<0.05), whereas the right eye did not (ANOVA, p>0.05).

Optokinetic response gain covaries with retinal density of long-wave sensitive photoreceptors

We hypothesised that the non-uniform distribution of the OKR gain across the visual field is related to the surface density of photoreceptors and investigated this using data from a recent study (Zimmermann et al., 2018; Chouinard-Thuly et al., 2017) on photoreceptor densities in explanted eye cups of 7- to 8-day-old zebrafish larvae. As shown in Figure 6b, ultraviolet receptor density exhibits a clear peak in the upper frontal part of the visual field, whereas red, green, and blue receptors (Figure 6a) are most concentrated across a wider region near the intersection of the equator and lateral meridian, with a bias to the upper visual field (in body coordinates). For comparison, density maps in retinal coordinates, not body coordinates, are shown in Figure 6—figure supplement 1, and Figure 6c shows total photoreceptor density across all types. To register our OKR gain data onto the photoreceptor density maps, we took the average eye position into account, which was located horizontally at −84.8 ± 6.2 degrees azimuth for the left and 80.1 ± 6.5 deg for the right eye (mean ±st.dev., n = 7 fish), and vertically at 3.5 ± 3.2 degrees elevation for the left and 4.9 ± 2.7 deg for the right eye (n = 10 fish). For green, blue, and especially red receptors, the visual space location in which OKR gain is maximal (cf, Figure 3e) coincides with a retinal region of near-maximum photoreceptor density (white ring in Figure 6). For ultraviolet receptors (Figure 6b), there is no strong correlation between photoreceptor density and OKR gain.

Figure 6. Maximum OKR gain is consistent with high photoreceptor densities in the retina.

Contour lines show retinal photoreceptor density determined by optical measurements of explanted eye cups of 7–8 dpf zebrafish larvae, at increments of 10% of maximum density. Data shown in visual space coordinates relative to the body axis, that is, 90° azimuth and 0° elevation corresponds to a perfectly lateral direction. To highlight densely covered regions, densities from half-maximum to maximum are additionally shown in shades of colour. Solid white circles indicate the location of maximum OKR gain inferred from experiments of type D in 5-7dpf larvae (Figure 3). White outlines indicate the area that would be covered by a 40° disk-shaped stimulus centred on this location when the eye is in its resting position. As the eyes move within their beating field during OKR, the actual, non-stationary retinal coverage extends further rostrally and caudally. For (a) red, green, and blue photoreceptors, high densities coincide with high OKR gains. (b) For ultraviolet receptors, there is no clear relationship to the OKR gain. (c) For reference, the summed total density of all receptor types combined. We did not observe a significant shift in the position-dependence of maximum OKR gain between groups of larvae at 5, 6, or 7 dpf of age, consistent with the notion that retinal development is far advanced and the circuits governing OKR behaviour are stable at this developmental stage. U: Up, D: Down, A: Anterior, P: Posterior. 

Figure 6—source data 1. Numerical data and graphical elements of Figure 6.

Figure 6.

Figure 6—figure supplement 1. Maximum OKR gain compared to photoreceptor densities in retinal coordinates.

Figure 6—figure supplement 1.

Same as Figure 6, but showing positions across the retina in Cartesian coordinates, as originally published (Zimmermann et al., 2018 Curr Biol), instead of in geographic visual field coordinates. When plotting photoreceptor densities in Cartesian coordinates, the regions of highest densities appear to be located quite peripheral/eccentric. However, the plot of the densities in visual field coordinates (Figure 6) confirms the coincidence of high densities and high OKR gains. Solid circles indicate the location of maximum OKR gain inferred from experiments of type D in 5-7 dpf larvae (Figure 3), and corrected by the mean eye position over time. U: Up, D: Down, A: Anterior, P: Posterior.

Discussion

We used a spherical visual display to systematically investigate visual space anisotropies of the zebrafish optokinetic response. We show that animals react most strongly to stimuli located laterally and near the equator of their visual space. Across individuals, the OKR appears to be symmetric between both eyes, although individual animals oftentimes have a dominant eye. For small stimuli, the OKR gain depends on the size of the area in which the stimulus was presented in a logarithmic fashion. OKR to our mostly yellow stimuli was tuned to the higher spatial densities of red, green, and blue photoreceptors in the central retina. In addition, further processing appears to affect the preferred OKR stimulus location, as suggested by the experiments in upside-down embedded animals. This could include extra-retinal processing, or hitherto unknown projections to the retina itself.

The spherical arena introduced here covers a large proportion of the surround and therefore lends itself to many other investigations of zebrafish, and of other species with limited visual acuity. In comparison to other feasible technical solutions, such as video projection setups, our spherical LED array stimulus setup provides homogeneous light and contrast across the entire stimulation area. Thereby, stimulus design is much easier because stimulus warping and conditioning becomes unnecessary. When combined with calcium imaging in a scanning microscope, the use of LED arrays provides the additional advantage that the visual stimulus can be controlled with high temporal precision, fast enough to interlace visual stimuli and line scans.

Despite the common notion that OKR is a whole-field gaze stabilisation behaviour, our results show that the OKR can be driven effectively by moving stimuli that cover only small parts of the spherical surface (e.g. a half-maximum OKR gain is observed for a stimulus that covers 25% of the spherical surface). In humans, small-field stimuli are effective in driving OKR as well, and the optokinetic drive and motion percept varies for central and peripheral retinal locations, with central stimulus locations eliciting stronger optokinetic drive (Howard and Ohmi, 1984; Brandt et al., 1973). Our experiment on spatial frequency dependence further demonstrates that the spatial frequency tuning of zebrafish OKR is similar across retinal locations. Since photoreceptors are not equally distributed across the retina, this result suggests that photoreceptor density is not the limiting factor for OKR performance in this frequency range.

Although for OKR and prey capture, animals are thought to respond to stimuli on the left and the right side alike, a few previous reports suggested that the zebrafish visual system is lateralised with the left eye preferentially assessing novel stimuli, while the right eye being associated with decisions to respond (Sovrano and Andrew, 2006; Miklósi and Andrew, 1999). Given the frequent occurrence of lateralised brain structures and behaviours in the animal kingdom (Bisazza et al., 1998; Güntürkün et al., 2020), we therefore investigated whether there are consistent behavioural asymmetries for the OKR. We observed almost no consistent, inter-individual asymmetries in OKR between the left and right hemispheres of the visual field, other than those induced by external conditions. Individual fish, however, show a wide and continuous range of biases towards either hemisphere.

We measured OKR gain in larvae at 5–7 days post fertilisation (dpf) of age, whereas our data on photoreceptor densities corresponds to slightly older, 7–8 dpf larvae. Owing to their rapid development, zebrafish undergo noticeable morphological changes on this timescale, but the zebrafish retina itself is known to be well developed by five dpf (Avanesov and Malicki, 2010) and stable OKR behaviour is exhibited from then on. Crucially, we did not observe a salient age-dependent spatial shift of maximum OKR gain between our 5 dpf and 7 dpf larvae (data not shown).

The qualitative match between red cone retinal photoreceptor densities and the stimulus position driving the highest OKR gains may provide a mechanistic bottom-up explanation of the gradual differences associated with OKR. The correspondence of red photoreceptor density with the visual field map of OKR gain is consistent with the fact that our LEDs emit light at 568 nm peak power, which should have activated the red cones most. Our data is also in agreement with observations in other species, that the OKR drive is strongest when the moving stimulus covers the central visual field (Murasugi and Howard, 1989; Shimizu et al., 2010; Howard and Ohmi, 1984). In a simplistic, additive view of visual processing, increased numbers of receptors would be triggered by incident light, gradually leading to stronger activation of retinal ganglion cells and downstream circuits, eventually driving extraocular eye muscles towards higher amplitudes. Instead, or in addition, the increased resolution offered by denser distributions of photoreceptors could help reduce sensory uncertainty (and increase visual acuity). It is unclear however, how more uncertainty would lead to consistently lower OKR gains instead of a repeated switching between periods of higher and lower gains, or between OKR and other behaviours. If sensory uncertainty were indeed crucial to OKR tuning, presenting blurred or otherwise deteriorated stimuli should reduce OKR gain in disfavoured locations more strongly than those in favoured locations. It is also possible that correlations between OKR gain and cone photoreceptor density are entirely coincidental, as our spatial frequency tuning results for different stimulus locations had implied. Genetic zebrafish variants with altered photoreceptor distributions would thus be a valuable tool for further studies. Recent studies suggest that melanopsin-expressing ganglion cells also contribute to form vision (Allen et al., 2019; Lucas et al., 2020), and melanopsins as well as a plethora of other opsins are expressed in the zebrafish retina and brain (Davies et al., 2015; Fontinha et al., 2021). It is therefore possible that photoreception via opsin-expressing neurons in the inner retina contributes to zebrafish OKR as well.

The pronounced increase in OKR gain for nearly lateral stimulus locations raises questions regarding the top-down behavioural significance of these directions in the natural habitat of larval zebrafish. While reduced OKR gains near the limits of the visual field might be expected, we show that gains are also reduced in the frontal binocular area, as well as in upper and lower visual field locations. Interestingly, when animals were mounted upside-down, they still prefer stimulus locations just above the equator of the environment. This result cannot be explained by shifted resting vertical eye positions in the inverted animal, which we have measured. Instead, it could potentially be explained by multimodal integration, where body orientation appears to influence the preferred OKR stimulus locations via the vestibular system (Lafortune et al., 1990; Pettorossi et al., 2011; Zolotilina et al., 1995).

Furthermore, it seems possible that the unequal distribution of OKR gains across the visual field is related to the optic flow statistics that naturally occur in the habitats of larval zebrafish (Zimmermann et al., 2018; Arunachalam et al., 2013; Engeszer et al., 2007; Parichy, 2015; Spence et al., 2008). For another stabilisation behaviour of zebrafish, the optomotor response (Orger et al., 2008), we have recently shown that the underlying circuits prefer stimulus locations in the lower temporal visual field to drive forward optomotor swimming (Wang et al., 2020). Therefore, the optokinetic and the optomotor response are preferentially driven by different regions in the visual field, suggesting that they occur in response to different types of optic flow patterns in natural habitats. Both the optokinetic and the optomotor response (OKR, OMR) are thought to be mediated by the pretectum (Kubo et al., 2014; Naumann et al., 2016), and we therefore hypothesise that circuits mediating OKR and OMR segregate within the pretectum and form neuronal ensembles with mostly different receptive field centre locations. Future studies on pretectal visual feature extraction in the context of naturalistic stimulus statistics are needed to establish a more complete picture of the visual pathways and computations underlying zebrafish OKR, OMR and other visually mediated behaviours.

Materials and methods

Key resources table.

Reagent type
(species) or resource
Designation Source or reference Identifiers Additional information
Genetic reagent
(Danio rerio)
mitfas184 https://zfin.org/action/genotype/view/ZDB-GENO-080305-2, (Scott et al., 2007) RRID:ZFIN_ZDB-ALT-070629-2 Mutant fish line (melanocyte inducing transcription factor a)
Software, algorithm ZebEyeTrack http://zebeyetrack.org/, (Dehmelt et al., 2018) DOI:10.1038/s41596-018-0002-0 Modular eye tracking software for zebrafish
Software, algorithm spherical arena code https://gin.g-node.org/Arrenberg_Lab/spherical_arena, (Dehmelt et al., 2020) DOI:10.12751/g-node.qergnn Code for stimulus generation, LED position mapping, and arena control
Software, algorithm OpenSCAD https://www.openscad.org/ 3D computer-aided design of arena scaffold
Other LED array Kingbright Electronic Co. TA08-81CGKWA Green light emitting LED tiles, 20 × 20 mm each, peak power at 568 nm
Other hardware controller Joesch et al., 2008 DOI:10.1016/j.cub.2008.02.022 Circuit board designs and code by Alexander Borst (MPI Neurobiol, Martinsried), Väinö Haikala and Dierk Reiff (University of Freiburg)

Animal experiments

Animal experiments were performed in accordance with licenses granted by local government authorities (Regierungspräsidium Tübingen) in accordance with German federal law and Baden-Württemberg state law. Approval of this license followed consultation of both in-house animal welfare officers and an external ethics board appointed by the local government. We used mitfa-/- animals (5–7 dpf) for the experiments, because this strain lacks skin pigmentation that could interfere with eye tracking. One report suggests that mitfa-/- animals show different optomotor responses than wildtype animals (Stowers et al., 2017), so it is possible that optokinetic responses investigated here are also modulated by this genetic background.

Coordinate systems and conventions

To remain consistent with the conventions adopted to describe stimuli and eye positions in previous publications, we adopted an East-North-Up, or ENU, geographic coordinate system. In this system, all positions are relative to the fish itself, and expressed as azimuth (horizontal angle, with positive values to the right of the fish), elevation (vertical angle, with positive values above the fish), and radius (or distance to the fish). The point directly in front of the fish (at the rostrum) is located at [0°, 0°] azimuth and elevation. Azimuth angles cover the range [−180°, 180°] and elevation angles [−90°, 90°]. Azimuth sign is opposite to the conventional mathematical notation of angles when looking top-down onto the fish. In our data repository, we provide a detailed description of the coordinate systems used, including transformations between Cartesian and geographic coordinate systems (Dehmelt et al., 2020https://gin.g-node.org/Arrenberg_Lab/spherical_arena).

Geographic coordinates can be used to identify stimulus position in at least four distinct reference frames: body-centred, head-centred, retina-centred, or world-centred coordinates. Because the heads of our embedded fish remain aligned with the body axis, we do not distinguish between body- and head-centred coordinates, and instead jointly refer to them as ‘fish-centred’. We generally identify stimuli by the coordinates of their centre, so there is a unique stimulus position in both fish-and world-centred coordinates. But because the eyes move during OKR, this stimulus centre will fall not on one but on many different retina-centred positions over time. Therefore, our comparison of the OKR gain map to retinal photoreceptor density distributions is based on the time-averaged eye position during optokinetic stimulation.

Design of the spherical arena

Geometric design of the arena

The overall layout of the spherical arena was optimised to contain the near maximum number of LED tiles that can be driven by our hardware controllers (232 out of a possible 240), and arrange them with minimal gaps in between. Also, care was taken to leave sufficient gaps near the top and bottom poles to insert the optical pathway used to illuminate and record fish behaviour. A further eight LED tiles could be included as optional covers for the top and bottom poles, bringing the total number to 240 out of 240 possible. A detailed walkthrough of the mathematical planning is found in our data repository (Dehmelt et al., 2020https://gin.g-node.org/Arrenberg_Lab/spherical_arena).

Arena elements

The arena consists of a 3D-printed structural scaffold (designed with open-source software OpenSCAD, Figure 1–source data 1); green light emitting LED tiles (Kingbright TA08-81CGKWA, 20x20 mm each, peak power at 568 nm) hot-glued to the scaffold and connected by cable to a set of circuit boards with hardware controllers (Figure 1—figure supplement 1d); 8x8 individual LEDs contained in each tile (Figure 1f); a nearly spherical glass bulb filled with water, into which the immobilised larvae are inserted (Figure 1—figure supplement 1c, middle); a metal rotation mount attached to the scaffold 'keel' of the arena (Figure 1—figure supplement 1c, right), holding the glass bulb in place and allowing corrections of pitch and roll angles; the optical pathway with an infrared light source to illuminate the fish from below (Figure 1—figure supplement 1b), and an infrared-sensitive USB camera for video recording of the transmission image (Figure 1—figure supplement 1d). In the assembled arena, the whole-field stimulus H1 (see Supplementary file 1E) resulted in a luminance of ( 10.94 ± 0.59cdm2) at the arena centre.

Electronics and circuit design

To provide hardware control to the LEDs, we used circuit boards designs and C controller code provided by Alexander Borst (MPI of Neurobiology, Martinsried), Väinö Haikala and Dierk Reiff (University of Freiburg) (Joesch et al., 2008). The electronic and software architecture of stimulus control has originally been designed by Reiser and Dickinson, 2008 (also see the available documentation for a different version of the CAD and code at (https://reiserlab.github.io/Modular-LED-Display/)). Any custom circuit board design and code could be substituted for these, and alternative solutions exist, for example, in Drosophila vision research (Suver et al., 2016). At the front end, these electronics control the 8 × 8 LED matrices, which are multiplexed in time to allow control of individual LEDs with just 8 input and eight output pins.

Optical pathway, illumination, and video recording

A high-power infrared LED was placed outside the stimulus arena and its light diffused by a sheet of milk glass and then guided towards the fish through the top hole of the arena (Figure 1—figure supplement 1b, Figure 1—figure supplement 1d). Non-absorbed IR light exits through the bottom hole, where it is focused onto an IR-sensitive camera. Between the arena and the proximal lens, a neutral density filter (NE13B, Thorlabs, ND 1.3) was inserted half-way (off-axis) into the optic pathway using an optical filter slider (CFH2/M, Thorlabs, positioned in about 5 cm distance of the camera CCD chip) to improve image contrast (oblique detection). We used the 840 nm, 125 degree IR emitter Roschwege Star-IR840-01-00-00 (procured via Conrad Electronic GmbH as item 491118–62) in custom casing, lenses LB1309 and LB1374, mirror PF20-03-P01 (ThorLabs GmbH), and IR-sensitive camera DMK23U618 (TheImagingSource GmbH). Approximate distances between elements are 14.5 cm (IR source to first lens), 12 cm (first lens to centre of glass bulb), 22 cm (bulb centre to mirror centre), 8.5 cm (mirror centre to second lens), 28.5 cm (second lens to camera objective).

Fish mounting device

Larvae were mounted inside a custom-built glass bulb (Figure 1—figure supplement 1c, middle). Its nearly spherical shape minimises reflection and refraction at the glass surface. It was filled with E3 solution, so there was no liquid-to-air boundary distorting visual stimuli. Through an opening on one side, we inserted a glass rod, on the tip of which we immobilise the larva in agarose gel (see description of the embedding procedure below). The fish was mounted in such a way that the head protruded the tip of the narrow triangular glass stage, which ensured that visual stimuli are virtually unobstructed by the glass triangle on their way to the eyes (Figure 1—figure supplement 1c, left). The entire glass structure was held at the centre of the spherical arena by metal parts attached to the arena scaffold itself (Figure 1—figure supplement 1c, right). Care was taken to remove air bubbles and completely fill the glass bulb with E3 medium.

Computer-assisted design and 3D printing

To arrange the square LED tiles across a nearly spherical surface, we 3D-printed a structural scaffold or ‘skeleton’, consisting of a reinforced prime meridian major circle (‘keel’) and several lighter minor circles of latitude (Figure 1g). Available hardware controllers allow for up to 240 LED matrices in parallel, so we chose the exact size of the scaffold (106.5 mm in diameter) to hold as many of these as possible while minimising gaps in between. As individual LEDs are arranged in a rectangular pattern on each of the flat LED tiles, and stimuli defined by true meridians (arcs from pole to pole, or straight vertical lines in Mercator projection), pixelation of the stimulus is inevitable, and stimulus edges become increasing stair-shaped near the poles. Because of the poor visual acuity of zebrafish larvae (see Materials and methods), this should not affect OKR behaviour. Our design further includes two holes necessary for behavioural recordings and two-photon imaging, located at the North and South poles of the sphere. We placed the largest elements of the structural scaffold behind the zebrafish (Figure 1—figure supplement 1d). Given the ~160° azimuth coverage per eye in combination with a slight eye convergence at rest, this minimises the loss of useful stimulation area.

We printed all structures out of polylactide (PLA) filament using an Ultimaker two printer (Ultimaker B.V.). Parts were assembled using a hot glue gun.

Visual field coverage of the arena

We can estimate the fraction of the visual field effectively covered by LEDs based on a projection of LED tiles onto a unit sphere. The area A of a surface segment delimited by the projection of the edges of a single tile onto the sphere centre is given by

AS=dA=-λλdx-λλdyuα×uβ

where uα and uβ are the Cartesian unit vectors spanning the tile itself and (±λ, ±λ) is the Cartesian position of the four edges of another rectangle. This smaller rectangle is the straight projection of the sphere segment onto the tile,

λ=sintan-1D/2RS

where RS=106.5mm is the sphere radius and D=21mm is the length of the edges of the tile. Summing over the number of tiles included in the arena, the equations above can be used to estimate the total coverage of the sphere by its square LED tiles to around 66.5% of the surface area. Using this strict estimate, the small gaps in between LED arrays are counted as not covered, even though we successfully demonstrated that they are small enough not to affect OKR performance, likely due to the low visual acuity of zebrafish larvae. A more meaningful estimate of coverage must take these results into account, and in fact reveals that stimuli presented with our LEDs effectively cover 85.6% of all possible directions:

The top and bottom holes of the arena accommodating the optic path for motion tracking are approximately limited by the circles of +69 and −69 degrees latitude. The fraction of a spherical surface area in between those circles, covered more or less densely by LEDs, corresponds to

F=Asphere2AcapAsphere=12AcapAsphere=1(1cos(9069))=cos21=93.36%

However, in addition to the holes at the top and bottom of the arena, the structural scaffold of our stimulus arena contains a wide ‘keel’ on the prime meridian, caudal to the fish, a narrower rostral keel, also on the prime meridian, as well as several thin structural ‘ribs’ at various latitudes, each without LEDs. The rostral keel spans about 30 degrees azimuth, reducing maximum coverage to 93.36%*(360-30)/360 = 85.58%. Yet this rostral keel is almost always outside the visual field of the zebrafish larvae and thus likely irrelevant for OKR behaviour. Our control experiments further demonstrate that the narrower rostral keel has no discernible effect on OKR gain. We thus assume that the even thinner structural ‘ribs’ have little to no effect, either; and the fraction of the visual field of zebrafish effectively covered by our stimulus arena thus exceeds 90%.

Visual acuity and the spatial resolution of the arena

The spatial resolution of the spherical stimulus arena is not very high, providing space for only about 0.54 LEDs per square degree on average. However, larval zebrafish have poor visual acuity, since each retina of their tiny eyes only contains about 10,000 photoreceptors (Zimmermann et al., 2018), corresponding to about 1–2 red-green double cone photoreceptors per square degree (Haug et al., 2010; Tappeiner et al., 2012). The spatial resolution of our arena will not bias the experimental results unless the stimulus is displayed in the most extreme positions.

Stimulus design

We designed visual stimuli, transformed them to geographical coordinates, and mapped them onto the physical positions of each individual LED with custom MATLAB software. We have made this code available for free under a Creative Commons NC-BY-SA 4.0 license ( Source code 1A). The mapped stimulus was then uploaded to the hardware controllers using custom-built C code originally developed by Väinö Haikala. Each stimulus type was shown twice to each individual. First, all stimulus types were shown in random order, with a different random sequence for each larva; then, the same stimuli were shown a second time, in the opposite order as before.

To investigate OKR gain dependence on stimulus location, we chose to present stimuli centred on 36 different locations distributed nearly equidistantly across the spherical arena, as well as symmetrically distributed between the left and right, upper and lower, front and rear hemispheres (Figure 3a). These positions were determined numerically: First, we populated one eighth of the sphere surface by placing one stimulus centre at a fixed location at the intersection of the equator and the most lateral meridian (90 degrees azimuth, 0 degrees elevation), constraining two more stimulus centres to move along this lateral meridian (90 degrees azimuth, initially random positive elevation), constraining yet another stimulus centre to move along the equator (initially random positive azimuth, 0 degrees elevation), and allowing three more stimulus centre to move freely across the surface of this eighth of the sphere (initially random positive azimuth and elevation), for a total of seven positions. Second, we placed additional stimulus centres onto all 29 positions that were mirror-symmetric to the initial 7, with mirror planes placed between the six hemispheres listed above. We then simulated interactions between all 38 stimulus centres akin to electromagnetic repulsion, until a stable pattern emerged. Resulting coordinate values were rounded for convenience (Source code 1B, Video 1).

Embedding procedure

To immobilise fish on the glass tip inside the sphere, we developed a novel embedding method. A cast of the glass triangle (and of the glass rod on which it is mounted) was made by placing it inside a Petri dish, which was then filled with a heated 2% agarose solution. After agarose cooled down and polymerised, agarose within a few millimetres of the tip of the glass triangle was manually removed, before removing the triangle itself. The resulting cast was stored in a refrigerator and then used to hold the glass triangle during all subsequent embedding procedures, limiting the freedom of movement of the larva to be embedded. The triangle was stored separately at room temperature. Before each embedding, we coated the glass triangle with polylysine and dried it overnight in an incubator at 29°C to increase the subsequent adhesion of agarose. We then returned the glass triangle into its cast, and constructed a tight, 2 mm high circular barrier around its tip using pieces of congealed agarose. A larva was picked up with as little water as possible using a glass pipette and very briefly placed inside 1 ml of 1.6% low-melting agarose solution at 37°C. Using the same pipette, the larvae was then transferred onto the glass triangle along with the entire agarose. After the larva had been placed a few millimetres away from the tip of the glass triangle, the orientation of the animal could be manipulated with custom-made platinum wire tools without touching its body, as previously described (Arrenberg, 2016). Before the agarose congeals, swimming motions of the animal were exploited to guide it towards the tip and ensure an upright posture. The final position of the fish was chosen as such that its eyes are aligned with the axis of the glass rod, its body is upright without any rotation, and its head protrudes forward from the tip of the glass triangle, maximising the fraction of its field of view unobstructed by glass elements. The agarose was left to congeal, and the Petri dish was filled with in E3 solution. The freshly congealed agarose surrounding the glass triangle was then removed using additional, flattened platinum wire tools, once again separating the glass triangle from the cast. Using the same tools, we finally cut triangular holes into the remaining agarose to completely free both eyes. To ensure free movement of both eyes, we confirmed the presence of large and even optokinetic eye movements using a striped paper drum before the experiment.

We then pick up the glass triangle by the glass rod attached to it, cut off any remaining agarose detritus, and place it inside the E3-filled glass bulb. No air remained in the bulb, and no pieces of detritus were introduced into the bulb, as these would accumulate near the top and bottom of the bulb, respectively, interfering with the optical pathway and thus reduce image quality.

Vertical eye position

A separate group of 5 dpf and 6 dpf fish were individually embedded in agarose, their eyes freed, and spontaneous eye position filmed at 60 Hz for about 3 min both along the dorsoventral axis and along the rostrocaudal axis simultaneously. Because changes in horizontal eye position affect the apparent shape of the eye when seen from the front, we sampled the recording with 1 Hz to evaluate the horizontal eye trace. The baseline eye position in the dorsoventral view was determined by mean ±0.5 std. Next, vertical eye position was evaluated if and only if the eyes in the dorsoventral view were at their baseline position. The vertical eye position was further corrected by embedding angle relative to the environmental horizon.

Data analysis

Video images of behaving zebrafish larvae were processed in real time using a precursor of the ZebEyeTrack software (Dehmelt et al., 2018), available from http://www.zebeyetrack.com/. The resulting traces of angular eye position were combined with analogue output signals from the hardware controllers of the spherical arena to match eye movement to the various stimulus phases. This was achieved using custom-built MATLAB software, which is freely available under a Creative Commons NC-BY-SA 4.0 license ( Source code 1C).

Data was then analysed further by detecting and removing saccades and fitting a piece-wise sinusoidal function to the eye position traces. The parameters of the fit were then compared to the parameters of the equally sinusoidally changing angular positions of the stimulus. For each fish, eye, and stimulus phase, the ratio between the amplitude of the fit to eye position and the amplitude of stimulus position represents one value of the gain of the optokinetic response.

For each interval between two subsequent saccades, or inter-saccade-interval (ISI), the fit function to the eye position data is defined by

f(tISIk)=-c1cosc2t+c3+ck+3

Here, t are the time stamps of data points falling within the k -th ISI, c1, c2, and c3 are the amplitude, frequency and phase shift of oscillation across all ISIs, and ck+3 is a different constant offset within each ISI, which corrects for the eye position offsets brought about by each saccade. The best fit value c1 was taken as an approximation of the amplitude aE of eye movement, aEc1. The process of cropping saccades from the raw data and fitting a sinusoid to the remaining raw data is demonstrated in Figure 1—figure supplement 1.

The OKR gain g is a common measure of visuomotor function. It is defined as the ratio between the amplitude aE of eye movement and the amplitude aS of the visual stimulus evoking eye movement,

g=aEaS=c1aS

In other words, OKR gain indicates the degree to which zebrafish larvae track a given visual stimulus. For each eye, a single gain value per stimulus phase is computed. While a value of 1 would indicate a ‘perfect’ match between eye movement and stimulus motion, zebrafish larvae at 5 dpf often exhibit much lower OKR gains (Rinner et al., 2005). While highest gains are obtained for very slowly moving stimuli, in our experiments, we chose higher stimulus velocities. Although these velocities are only tracked with small gains, the absolute velocities of the eyes are high, which allowed us to collect data with high signal-to-noise levels and reduce the needed recording time.

To rule out asymmetries induced by the arena itself or by its surroundings, we recorded two sets of stimulus-position-dependence data, one with the arena in its original configuration, and another with the arena rotated by 180 degrees (Figure 4a–b). Each set contained data from multiple larvae, and with at least two separate presentations of each stimulus position. For each stimulus position, and separately for both sets of data, we computed the median OKR gain across fish and stimulus repetitions. We then averaged between the two datasets, yielding a single OKR gain value per stimulus position. As asymmetries are less crucial when studying stimulus frequency and size (Figure 5), we did not repeat those with a rotated arena, and could thus omit the final step of the analysis.

Von Mises-Fisher fits to data

Based on the assumption that OKR position tuning could be normally distributed with respect to each angle, OKR gain would be approximated by a two-dimensional, circular von Mises-Fisher function centred on the preferred stimulus location. Because the eyes are yoked, the OKR gain of one eye will be high around its own preferred position, as well as around the preferred position of the contralateral eye. To account for this, we fit the sum of two independent von Mises-Fisher functions to our OKR gain data:

Fα,β=C1κ1expκ1μ1Tξ2πexpκ1-exp-κ1+C2κ2expκ2μ2Tξ2πexpκ2-exp-κ2+C3

Here, ξ is the Cartesian coordinate vector of a point on the sphere surface, and corresponds to the geographic coordinates azimuth α and elevation β. μ1 and μ2 are Cartesian coordinate vectors pointing to the centre of the two distributions, and κ1 and κ2 express their respective concentrations, or narrowness. The parameters μj, κj, the amplitudes C1, C2 and the offset C3 are fit numerically.

Figure 3 and Figure 4e–h show the best von-Mises-Fisher fits to data as coloured sphere surfaces, whereas the individual data points are shown as small coloured circles in Figure 3. All results shown in Figure 3 and Figure 4e–h pertain to a single, constant size of cropping disks, and the colour of any point across the sphere surface represents the expected value of OKR gain if another stimulus were shown with a cropping disk centered on this location.

Permutation test

We tested whether the preferred stimulus positions are indeed ‘above’ in world-centred coordinates, instead of ‘dorsal’ in body-centred coordinates. Two groups of fish with different embedding direction were compared (Figure 4a vs. Figure 4c). Within each group, we pooled data from all individuals, then computed the best von Mises-Fisher fit. Within each group, we then averaged between the two elevations obtained for the left and right hemispheres, to obtain a single measure of elevation per group. Our test statistic was the difference in (body-centred) elevation between the groups. We then permuted the group-identity labels of each fish, thus assigning fish to two new groups. We then computed the test statistic again for each permuted dataset. This was repeated for all 120 possible permutations that maintain the same relative group size as the two experimental groups. The reported p-value is the fraction of permutations resulting in a test statistic greater or equal to that of the original, unpermuted groups (Figure 4—figure supplement 3).

Yoking index, asymmetry, and mathematical modelling

To quantify asymmetries in the gain between left and right, stimulated and unstimulated eyes, we introduce the yoking index

Y=gL-gRgL+gR

Here, gL and gR are the OKR gains of the left eye and right eye, measured during the same stimulus phase. Depending on stimulus phase, only the left eye, only the right eye or both eyes may have been stimulated. If the yoking index is positive, the left eye responded more strongly than the right eye; if it is negative, the amplitude of right eye movement was larger. An index of zero indicates 'perfect yoking', that is identical amplitudes for both eyes.

In addition, we define a ‘bias’ index to capture innate or induced asymmetries between responses to stimuli presented within the left or right hemisphere of the visual field,

B=mL-mRmL+mR

Here, mL and mR are the medians of OKR gains after pooling across either all left-side or all right-side stimulus types (D1-D19 and D20-D38, respectively). Several sources of asymmetry contribute to B: (1) arena- or environment-related differences in stimulus perception, constant across individuals; (2) a biologically encoded preference for one of the two eyes, constant across individuals; (3) inter-individual differences between the eyes, constant across stimulus phases for each individual; (4) other sources of variability unaccounted for, and approximated as a noise term η. We hypothesise that the overall asymmetry observed for each larva k is given by a simple linear combination of these contributions,

Bk=φb1+b2+b3,k+η

The parameter φ is 1 for the default arena setup, and –1 during control experiments with a horizontally flipped arena setup. To determine b1, b2 and b3, we fit this system of equations by multivariate linear regression to experimentally observed bias indices. The system is initially underdetermined, as it contains n+2 coefficients for every n fish observed. However, if we assume that individual biases average out across the population, we can determine the population-wide coefficients b1 and b2 by setting aside the individual b3,k for a first regression. To determine how far each individual deviates from the rest of the population, we then substitute their best regression values of b1 and b2 into the full equation, and perform a second regression for the remaining b3,k.

Data repository

The raw recordings as well as pre-processed datasets and programming scripts to generate the figures are available at a GIN repository at https://gin.g-node.org/Arrenberg_Lab/spherical_arena/ (Dehmelt et al., 2020).

Acknowledgements

We thank Väinö Haikala and Dierk F Reiff (University of Freiburg) for sharing their code and design for hardware controllers, Alexander Borst (MPI Neurobiology, Martinsried) for providing the LED panel board design based on original precursory versions designed by Reiser et al. (Reiser and Dickinson, 2008), Thomas Nieß (glassblowing workshop, University of Tübingen) and Klaus Vollmer (precision mechanics workshop, University of Tübingen) for technical support, Prudenter-Agas (Hamburg, Germany) for generating glass bulb illustrations, Andre Maia Chagas for help with 3D printing procedures, and Jan Benda (University of Tübingen) for discussions on visual acuity. Funding This work was funded by the Deutsche Forschungsgemeinschaft (DFG) grants EXC307 (CIN – Werner Reichardt Centre for Integrative Neuroscience) and a Human Frontier Science Program (HFSP) Young Investigator Grant RGY0079.

Funding Statement

The funders had no role in study design, data collection and interpretation, or the decision to submit the work for publication.

Contributor Information

Aristides B Arrenberg, Email: aristides.arrenberg@uni-tuebingen.de.

Kristin Tessmar-Raible, University of Vienna, Austria.

Tirin Moore, Stanford University, United States.

Funding Information

This paper was supported by the following grants:

  • Deutsche Forschungsgemeinschaft EXC307 to Aristides B Arrenberg.

  • Human Frontier Science Program RGY0079 to Aristides B Arrenberg.

Additional information

Competing interests

No competing interests declared.

Author contributions

Software, Formal analysis, Supervision, Validation, Visualization, Methodology, Writing - original draft, Writing - review and editing.

Software, Formal analysis, Validation, Investigation, Methodology, Writing - original draft.

Methodology, Writing - original draft, Writing - review and editing.

Resources, Writing - review and editing.

Validation, Investigation.

Validation, Investigation.

Methodology, Writing - review and editing.

Resources, Writing - review and editing.

Conceptualization, Software, Funding acquisition, Methodology, Writing - original draft, Project administration, Writing - review and editing.

Ethics

Animal experimentation: Animal experiments were performed in accordance with licenses granted by local government authorities (Regierungspräsidium Tübingen) in accordance with German federal law and Baden-Württemberg state law. Approval of this license followed consultation of both in-house animal welfare officers and an external ethics board appointed by the local government.

Additional files

Source code 1. Five separate sets of MATLAB code.

(A) Code to design visual stimuli, convert them to geographic coordinates, and map them onto the actual position of individual LEDs. (B) Code to numerically identify a distribution of nearly equidistant stimulus centres that is symmetric between the left and right, upper and lower, as well as front and rear hemispheres. (C) Code to read raw eye traces, identify individual stimulus phases, detect and remove saccades, compute piece-wise fits to cropped and pre-processed raw data, and return OKR gains. (D) Code to recreate the results figures from data, especially Figure 3, Figure 5, Figure 6. (E) Code to assess the significance of differences between the best fits to data for fish embedded upright or upside-down. This last set of code requires access to the raw data repository.

elife-63355-code1.zip (148KB, zip)
Supplementary file 1. Seven supplementary tables.

(A) The arena cross-section, (B) stimulus parameters for position dependence experiments, (C) absolute positions of fish and setup elements in different experiments, (D) stimulus parameters for control experiments, (E) parameters for whole-field and hemisphere stimuli, (F) stimulus parameters for frequency dependence experiments, and (G) stimulus parameters for size dependence experiments.

elife-63355-supp1.docx (66.2KB, docx)
Transparent reporting form

Data availability

Analysis code, pre-processed data and examples of raw data have been deposited in GIN by G-Node and published (https://doi.org/10.12751/g-node.qergnn).

The following dataset was generated:

Dehmelt FA, Meier R, Hinz J, Yoshimatsu T, Simacek CA, Wang K, Baden T, Arrenberg AB. 2020. Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish. GIN (G-Node Infrastructure)

The following previously published dataset was used:

Zimmermann MJY, Nevala NE, Yoshimatsu T, Osorio D, Nilsson DE, Berens P, Baden T. 2019. Data from: Zebrafish differentially process colour across visual space to match natural scenes. Dryad Digital Repository.

References

  1. Allen AE, Martial FP, Lucas RJ. Form vision from melanopsin in humans. Nature Communications. 2019;10:2274. doi: 10.1038/s41467-019-10113-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Andrew RJ, Dharmaretnam M, Gyori B, Miklósi A, Watkins JA, Sovrano VA. Precise endogenous control of involvement of right and left visual structures in assessment by zebrafish. Behavioural Brain Research. 2009;196:99–105. doi: 10.1016/j.bbr.2008.07.034. [DOI] [PubMed] [Google Scholar]
  3. Arrenberg AB. Fiber Optic-Based photostimulation of larval zebrafish. Methods in Molecular Biology. 2016;1451:343–354. doi: 10.1007/978-1-4939-3771-4_24. [DOI] [PubMed] [Google Scholar]
  4. Arunachalam M, Raja M, Vijayakumar C, Malaiammal P, Mayden RL. Natural history of zebrafish (Danio rerio) in India. Zebrafish. 2013;10:1–14. doi: 10.1089/zeb.2012.0803. [DOI] [PubMed] [Google Scholar]
  5. Avanesov A, Malicki J. Analysis of the retina in the zebrafish model. Methods in Cell Biology. 2010;100:153–204. doi: 10.1016/B978-0-12-384892-5.00006-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Baden T, Schubert T, Chang L, Wei T, Zaichuk M, Wissinger B, Euler T. A tale of two retinal domains: near-optimal sampling of achromatic contrasts in natural scenes through asymmetric photoreceptor distribution. Neuron. 2013;80:1206–1217. doi: 10.1016/j.neuron.2013.09.030. [DOI] [PubMed] [Google Scholar]
  7. Baier H, Scott EK. Genetic and optical targeting of neural circuits and behavior--zebrafish in the spotlight. Current Opinion in Neurobiology. 2009;19:553–560. doi: 10.1016/j.conb.2009.08.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Beck JC, Gilland E, Tank DW, Baker R. Quantifying the ontogeny of optokinetic and vestibuloocular behaviors in zebrafish, medaka, and goldfish. Journal of Neurophysiology. 2004;92:3546–3561. doi: 10.1152/jn.00311.2004. [DOI] [PubMed] [Google Scholar]
  9. Bianco IH, Kampff AR, Engert F. Prey capture behavior evoked by simple visual stimuli in larval zebrafish. Frontiers in Systems Neuroscience. 2011;5:101. doi: 10.3389/fnsys.2011.00101. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Bisazza A, Rogers LJ, Vallortigara G. The origins of cerebral asymmetry: a review of evidence of behavioural and brain lateralization in fishes, reptiles and amphibians. Neuroscience & Biobehavioral Reviews. 1998;22:411–426. doi: 10.1016/S0149-7634(97)00050-X. [DOI] [PubMed] [Google Scholar]
  11. Brandt T, Dichgans J, Koenig E. Differential effects of central verses peripheral vision on egocentric and exocentric motion perception. Experimental Brain Research. 1973;16:476–491. doi: 10.1007/BF00234474. [DOI] [PubMed] [Google Scholar]
  12. Brockerhoff SE, Hurley JB, Janssen-Bienhold U, Neuhauss SC, Driever W, Dowling JE. A behavioral screen for isolating zebrafish mutants with visual system defects. PNAS. 1995;92:10545–10549. doi: 10.1073/pnas.92.23.10545. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Chouinard-Thuly L, Gierszewski S, Rosenthal GG, Reader SM, Rieucau G, Woo KL, Gerlai R, Tedore C, Ingley SJ, Stowers JR, Frommen JG, Dolins FL, Witte K. Technical and conceptual considerations for using animated stimuli in studies of animal behavior. Current Zoology. 2017;63:5–19. doi: 10.1093/cz/zow104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Clark D. Visual responses in the developing zebrafish . Brachydanio Rerio.1981. [Google Scholar]
  15. Cohen B, Matsuo V, Raphan T. Quantitative analysis of the velocity characteristics of optokinetic nystagmus and optokinetic after-nystagmus. The Journal of Physiology. 1977;270:321–344. doi: 10.1113/jphysiol.1977.sp011955. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Cronin TJ, Marshal NJ, Warrant EJ. Visual Ecology. Princeton University Press; 2014. [Google Scholar]
  17. Davies WI, Tamai TK, Zheng L, Fu JK, Rihel J, Foster RG, Whitmore D, Hankins MW. An extended family of novel vertebrate photopigments is widely expressed and displays a diversity of function. Genome Research. 2015;25:1666–1679. doi: 10.1101/gr.189886.115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Dehmelt FA, von Daranyi A, Leyden C, Arrenberg AB. Evoking and tracking zebrafish eye movement in multiple larvae with ZebEyeTrack. Nature Protocols. 2018;13:1539–1568. doi: 10.1038/s41596-018-0002-0. [DOI] [PubMed] [Google Scholar]
  19. Dehmelt FA, Meier R, Hinz J, Yoshimatsu T, Simacek CA, Huang R, Wang K, Baden T, Arrenberg AB. 2020. Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish. G-Node. [DOI]
  20. Easter SS, Nicola GN. The development of vision in the zebrafish (Danio rerio) Developmental Biology. 1996;180:646–663. doi: 10.1006/dbio.1996.0335. [DOI] [PubMed] [Google Scholar]
  21. Engeszer RE, Patterson LB, Rao AA, Parichy DM. Zebrafish in the wild: a review of natural history and new notes from the field. Zebrafish. 2007;4:21–40. doi: 10.1089/zeb.2006.9997. [DOI] [PubMed] [Google Scholar]
  22. Fontinha BM, Zekoll T, Al-Rawi M, Gallach M, Reithofer F, Barker AJ, Hofbauer M, Fischer RM, von Haeseler A, Baier H, Tessmar-Raible K. TMT-Opsins differentially modulate medaka brain function in a context-dependent manner. PLOS Biology. 2021;19:e3001012. doi: 10.1371/journal.pbio.3001012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Güntürkün O, Ströckens F, Ocklenburg S. Brain lateralization: a comparative perspective. Physiological Reviews. 2020;100:1019–1063. doi: 10.1152/physrev.00006.2019. [DOI] [PubMed] [Google Scholar]
  24. Hafed ZM, Chen CY. Sharper, stronger, faster upper visual field representation in primate superior colliculus. Current Biology. 2016;26:1647–1658. doi: 10.1016/j.cub.2016.04.059. [DOI] [PubMed] [Google Scholar]
  25. Haug MF, Biehlmaier O, Mueller KP, Neuhauss SC. Visual acuity in larval zebrafish: behavior and histology. Frontiers in Zoology. 2010;7:8. doi: 10.1186/1742-9994-7-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Heap LAL, Vanwalleghem G, Thompson AW, Favre-Bulle IA, Scott EK. Luminance changes drive directional startle through a thalamic pathway. Neuron. 2018;99:293–301. doi: 10.1016/j.neuron.2018.06.013. [DOI] [PubMed] [Google Scholar]
  27. Howard IP, Ohmi M. The efficiency of the central and peripheral retina in driving human optokinetic nystagmus. Vision Research. 1984;24:969–976. doi: 10.1016/0042-6989(84)90072-5. [DOI] [PubMed] [Google Scholar]
  28. Hoy JL, Yavorska I, Wehr M, Niell CM. Vision drives accurate approach behavior during prey capture in laboratory mice. Current Biology. 2016;26:3046–3052. doi: 10.1016/j.cub.2016.09.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Huang YY, Neuhauss SC. The optokinetic response in zebrafish and its applications. Frontiers in Bioscience. 2008;13:1899–1916. doi: 10.2741/2810. [DOI] [PubMed] [Google Scholar]
  30. Joesch M, Plett J, Borst A, Reiff DF. Response properties of motion-sensitive visual interneurons in the lobula plate of Drosophila melanogaster. Current Biology. 2008;18:368–374. doi: 10.1016/j.cub.2008.02.022. [DOI] [PubMed] [Google Scholar]
  31. Johnson RE, Linderman S, Panier T, Wee CL, Song E, Herrera KJ, Miller A, Engert F. Probabilistic models of larval zebrafish behavior reveal structure on many scales. Current Biology. 2020;30:70–82. doi: 10.1016/j.cub.2019.11.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kim SS, Rouault H, Druckmann S, Jayaraman V. Ring attractor dynamics in the Drosophila central brain. Science. 2017;356:849–853. doi: 10.1126/science.aal4835. [DOI] [PubMed] [Google Scholar]
  33. Kist AM, Portugues R. Optomotor swimming in larval zebrafish is driven by global Whole-Field visual motion and local Light-Dark transitions. Cell Reports. 2019;29:659–670. doi: 10.1016/j.celrep.2019.09.024. [DOI] [PubMed] [Google Scholar]
  34. Kubo F, Hablitzel B, Dal Maschio M, Driever W, Baier H, Arrenberg AB. Functional architecture of an optic flow-responsive area that drives horizontal eye movements in zebrafish. Neuron. 2014;81:1344–1359. doi: 10.1016/j.neuron.2014.02.043. [DOI] [PubMed] [Google Scholar]
  35. Lafortune SH, Ireland DJ, Jell RM. Suppression of optokinetic velocity storage in humans by static tilt in roll. Journal of Vestibular Research : Equilibrium & Orientation. 1990;1:347–355. [PubMed] [Google Scholar]
  36. Lagogiannis K, Diana G, Meyer MP. Learning steers the ontogeny of an efficient hunting sequence in zebrafish larvae. eLife. 2020;9:e55119. doi: 10.7554/eLife.55119. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Lindsey BW, Smith FM, Croll RP. From inflation to flotation: contribution of the swimbladder to whole-body density and swimming depth during development of the zebrafish (Danio rerio) Zebrafish. 2010;7:85–96. doi: 10.1089/zeb.2009.0616. [DOI] [PubMed] [Google Scholar]
  38. Lucas RJ, Allen AE, Milosavljevic N, Storchi R, Woelders T. Can we see with melanopsin? Annual Review of Vision Science. 2020;6:453–468. doi: 10.1146/annurev-vision-030320-041239. [DOI] [PubMed] [Google Scholar]
  39. Maisak MS, Haag J, Ammer G, Serbe E, Meier M, Leonhardt A, Schilling T, Bahl A, Rubin GM, Nern A, Dickson BJ, Reiff DF, Hopp E, Borst A. A directional tuning map of Drosophila elementary motion detectors. Nature. 2013;500:212–216. doi: 10.1038/nature12320. [DOI] [PubMed] [Google Scholar]
  40. McLean DL, Fetcho JR. Movement, technology and discovery in the zebrafish. Current Opinion in Neurobiology. 2011;21:110–115. doi: 10.1016/j.conb.2010.09.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Mearns DS, Donovan JC, Fernandes AM, Semmelhack JL, Baier H. Deconstructing hunting behavior reveals a tightly coupled Stimulus-Response loop. Current Biology. 2020;30:54–69. doi: 10.1016/j.cub.2019.11.022. [DOI] [PubMed] [Google Scholar]
  42. Merigan WH, Katz LM. Spatial resolution across the macaque retina. Vision Research. 1990;30:985–991. doi: 10.1016/0042-6989(90)90107-V. [DOI] [PubMed] [Google Scholar]
  43. Miklósi A, Andrew RJ. Right eye use associated with decision to bite in zebrafish. Behavioural Brain Research. 1999;105:199–205. doi: 10.1016/S0166-4328(99)00071-6. [DOI] [PubMed] [Google Scholar]
  44. Murasugi CM, Howard IP. Human horizontal optokinetic nystagmus elicited by the upper versus the lower visual fields. Visual Neuroscience. 1989;2:73–79. doi: 10.1017/S095252380000434X. [DOI] [PubMed] [Google Scholar]
  45. Muto A, Orger MB, Wehman AM, Smear MC, Kay JN, Page-McCaw PS, Gahtan E, Xiao T, Nevin LM, Gosse NJ, Staub W, Finger-Baier K, Baier H. Forward genetic analysis of visual behavior in zebrafish. PLOS Genetics. 2005;1:e66. doi: 10.1371/journal.pgen.0010066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Naumann EA, Fitzgerald JE, Dunn TW, Rihel J, Sompolinsky H, Engert F. From Whole-Brain data to functional circuit models: the zebrafish optomotor response. Cell. 2016;167:947–960. doi: 10.1016/j.cell.2016.10.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Nawrocki L, BreMiller R, Streisinger G, Kaplan M. Larval and adult visual pigments of the zebrafish, Brachydanio rerio. Vision Research. 1985;25:1569–1576. doi: 10.1016/0042-6989(85)90127-0. [DOI] [PubMed] [Google Scholar]
  48. Neuhauss SC, Biehlmaier O, Seeliger MW, Das T, Kohler K, Harris WA, Baier H. Genetic disorders of vision revealed by a behavioral screen of 400 essential loci in zebrafish. The Journal of Neuroscience. 1999;19:8603–8615. doi: 10.1523/JNEUROSCI.19-19-08603.1999. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Orger MB, Kampff AR, Severi KE, Bollmann JH, Engert F. Control of visually guided behavior by distinct populations of spinal projection neurons. Nature Neuroscience. 2008;11:327–333. doi: 10.1038/nn2048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Parichy DM. Advancing biology through a deeper understanding of zebrafish ecology and evolution. eLife. 2015;4:e05635. doi: 10.7554/eLife.05635. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Pettorossi VE, Ferraresi A, Botti FM, Panichi R, Barmack NH. Head position modulates optokinetic nystagmus. Experimental Brain Research. 2011;213:141–152. doi: 10.1007/s00221-011-2785-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Portugues R, Feierstein CE, Engert F, Orger MB. Whole-brain activity maps reveal stereotyped, distributed networks for visuomotor behavior. Neuron. 2014;81:1328–1343. doi: 10.1016/j.neuron.2014.01.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Reiser MB, Dickinson MH. A modular display system for insect behavioral neuroscience. Journal of Neuroscience Methods. 2008;167:127–139. doi: 10.1016/j.jneumeth.2007.07.019. [DOI] [PubMed] [Google Scholar]
  54. Rinner O, Rick JM, Neuhauss SCF. Contrast sensitivity, spatial and temporal tuning of the larval zebrafish optokinetic response. Investigative Opthalmology & Visual Science. 2005;46:137. doi: 10.1167/iovs.04-0682. [DOI] [PubMed] [Google Scholar]
  55. Schaerer S, Kirschfeld K. The role of background movement in goldfish vision. Journal of Comparative Physiology A: Sensory, Neural, and Behavioral Physiology. 2000;186:583–593. doi: 10.1007/s003590050010. [DOI] [PubMed] [Google Scholar]
  56. Scott EK, Mason L, Arrenberg AB, Ziv L, Gosse NJ, Xiao T, Chi NC, Asakawa K, Kawakami K, Baier H. Targeting neural circuitry in zebrafish using GAL4 enhancer trapping. Nature Methods. 2007;4:323–326. doi: 10.1038/nmeth1033. [DOI] [PubMed] [Google Scholar]
  57. Shcherbakov D, Knörzer A, Espenhahn S, Hilbig R, Haas U, Blum M. Sensitivity differences in fish offer near-infrared vision as an adaptable evolutionary trait. PLOS ONE. 2013;8:e64429. doi: 10.1371/journal.pone.0064429. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Shimizu N, Tabata H, Wada Y, Sugita Y, Yamanaka T, Hosoi H, Kawano K. Distribution of optokinetic sensitivity across the retina of mice in relation to eye orientation. Neuroscience. 2010;168:200–208. doi: 10.1016/j.neuroscience.2010.03.025. [DOI] [PubMed] [Google Scholar]
  59. Smolka J, Zeil J, Hemmi JM. Natural visual cues eliciting predator avoidance in fiddler crabs. Proceedings. Biological Sciences. 2011;278:3584–3592. doi: 10.1098/rspb.2010.2746. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Sovrano VA, Andrew RJ. Eye use during viewing a reflection: behavioural lateralisation in zebrafish larvae. Behavioural Brain Research. 2006;167:226–231. doi: 10.1016/j.bbr.2005.09.021. [DOI] [PubMed] [Google Scholar]
  61. Spence R, Gerlach G, Lawrence C, Smith C. The behaviour and ecology of the zebrafish, Danio rerio. Biological Reviews. 2008;83:13–34. doi: 10.1111/j.1469-185X.2007.00030.x. [DOI] [PubMed] [Google Scholar]
  62. Stowers JR, Hofbauer M, Bastien R, Griessner J, Higgins P, Farooqui S, Fischer RM, Nowikovsky K, Haubensak W, Couzin ID, Tessmar-Raible K, Straw AD. Virtual reality for freely moving animals. Nature Methods. 2017;14:995–1002. doi: 10.1038/nmeth.4399. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Suver MP, Huda A, Iwasaki N, Safarik S, Dickinson MH. An array of descending visual interneurons encoding Self-Motion in Drosophila. The Journal of Neuroscience. 2016;36:11768–11780. doi: 10.1523/JNEUROSCI.2277-16.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Tappeiner C, Gerber S, Enzmann V, Balmer J, Jazwinska A, Tschopp M. Visual acuity and contrast sensitivity of adult zebrafish. Frontiers in Zoology. 2012;9:10. doi: 10.1186/1742-9994-9-10. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Trivedi CA, Bollmann JH. Visually driven chaining of elementary swim patterns into a goal-directed motor sequence: a virtual reality study of zebrafish prey capture. Frontiers in Neural Circuits. 2013;7:86. doi: 10.3389/fncir.2013.00086. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Wang K, Hinz J, Haikala V, Reiff DF, Arrenberg AB. Selective processing of all rotational and translational optic flow directions in the zebrafish pretectum and tectum. BMC Biology. 2019;17:29. doi: 10.1186/s12915-019-0648-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Wang K, Hinz J, Zhang Y, Thiele TR, Arrenberg AB. Parallel channels for motion feature extraction in the pretectum and tectum of larval zebrafish. Cell Reports. 2020;30:442–453. doi: 10.1016/j.celrep.2019.12.031. [DOI] [PubMed] [Google Scholar]
  68. Wang K, Arrenberg B, Hinz J, Arrenberg AB. Reduction of visual stimulus artifacts using a spherical tank for small, aquatic animals. Scientific Reports. 2021;11:3204. doi: 10.1038/s41598-021-81904-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Watkins J, Miklósi A, Andrew RJ. Early asymmetries in the behaviour of zebrafish larvae. Behavioural Brain Research. 2004;151:177–183. doi: 10.1016/j.bbr.2003.08.012. [DOI] [PubMed] [Google Scholar]
  70. Yang X, Guo A. Distinct acute zones for visual stimuli in different visual tasks in Drosophila. PLOS ONE. 2013;8:e61313. doi: 10.1371/journal.pone.0061313. [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Yoshimatsu T, Schröder C, Nevala NE, Berens P, Baden T. Fovea-like photoreceptor specializations underlie single UV cone driven Prey-Capture behavior in zebrafish. Neuron. 2020;107:320–337. doi: 10.1016/j.neuron.2020.04.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Zhang Y, Kim IJ, Sanes JR, Meister M. The most numerous ganglion cell type of the mouse retina is a selective feature detector. PNAS. 2012;109:E2391–E2398. doi: 10.1073/pnas.1211547109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Zimmermann MJY, Nevala NE, Yoshimatsu T, Osorio D, Nilsson DE, Berens P, Baden T. Zebrafish differentially process color across visual space to match natural scenes. Current Biology. 2018;28:2018–2032. doi: 10.1016/j.cub.2018.04.075. [DOI] [PubMed] [Google Scholar]
  74. Zolotilina EG, Eremina SV, Orlov IV. Horizontal optokinetic nystagmus in the pigeon during static tilts in the frontal plane. Neuroscience and Behavioral Physiology. 1995;25:300–306. doi: 10.1007/BF02360041. [DOI] [PubMed] [Google Scholar]

Decision letter

Editor: Kristin Tessmar-Raible1
Reviewed by: Andrew D Straw2

In the interests of transparency, eLife publishes the most substantive revision requests and the accompanying author responses.

Acceptance summary:

The paper is of interest to people studying visual systems neuroscience. The authors used an impressive new experimental set-up to determine the dependency of optokinetic response tuning to stimulus location, size, and frequency.

Decision letter after peer review:

Thank you for submitting your article "Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish" for consideration by eLife. Your article has been reviewed by 2 peer reviewers, and the evaluation has been overseen by a Reviewing Editor and Tirin Moore as the Senior Editor. The following individual involved in review of your submission has agreed to reveal their identity: Andrew D Straw (Reviewer #1).

The reviewers have discussed the reviews with one another and the Reviewing Editor has drafted this decision to help you prepare a revised submission.

We would like to draw your attention to changes in our revision policy that we have made in response to COVID-19 (https://elifesciences.org/articles/57162). Thus the revisions requested below will likely be addressable without any additional lab experiments.

The reviewers and the Reviewing Editor agreed that the design of the set-up presents a technical innovation that will be highly useful for future studies. There is clear disagreement on the level of conceptual insight.

Based on the technical advancement we agreed on a major revision of the manuscript. However, the resubmission should not be handed in as a research article, but as a Tools and Resources article.

Please address our major concerns listed below.

1) Statistical analysis. For several results only means are shown, and often there does not appear to be an indication of the variance between individuals. For instance:

– In Figures3 and 4 it would be interesting to know how this varies across animals. One possibility would be to also plot a variance map. Another would be to include the maps for each animal individually as supplementary data.

– In Figure 5 error bars should be shown in d,e, and also perhaps the plots for individual animals as supplementary data.

– Line 123: how variable is the visual field size? The current writing suggests it's exactly 163 degrees in every fish.

– Line 248: What is the variance over animals of 80.3 degrees?

– Line 289-292: What is the variance in these numbers?

– Line 299: please show us what the distribution for the permutation test looks like.

2) The authors argue that "extra-retinal processing" is responsible for the minor shift in the center of the highest gain OKR spot. However, while the effect indeed seems likely to require extra-retinal input, the authors provide no evidence that the processing mediating this should occur outside of the retina. Given other examples of non-visual information being sent to the retina, it is at least plausible that retinal processing is involved.

– Line 79: The evidence for it being "extra-retinal processing" is lacking. The source of the effect is likely extra-retinal but whether the processing happens in the retina or elsewhere is not shown.

– 182: The "determinants" are indeed likely extra-retinal, but this does not mean the processing is.

– Line 428: You have not demonstrated that the processing is extra-retinal, only that it is not occurring in retinal coordinates. There could be a projection into the retina, for example.

3) The authors should consider three coordinate systems explicitly, particularly in the lines 273-303. These would be retinal coordinates, head coordinates and environmental coordinates.

– Line 277: It seems there's also another representation which is missing from your consideration here – a body centric representation. This is distinct from retinal coordinates and distinct from environmental coordinates but is known from primate studies, for example.

– Line 282. Now retina-centered is missing. (The retina can move independently from the head/body).

4) Especially as we consider this manuscript rather an advancement for the "Tools and Resources" section, please consider that the Stowers J et al. 2017 "Virtual Reality for Freely Moving Animals" was a full-surround setup – at least as defined by the standard for the other citations – used successfully on Drosophila and, importantly here, fish. You need to be more specific in your wording (e.g. to limit statements to systems based on LEDs capable of producing the desired spectral content) or cite that paper.

5) The statement that the gain of the OKR broadly follow the density of the longer-wavelength sensitive photoreceptors in the retina is entirely ignoring the fact that the fish retina is full of Opsins in all kind of different cells types. This includes different melanopsins, but also VA-opsins (which are likely sensitive in the green), opn5s (UVA), tmt-opsins etc. Evidence from mouse and humans starts to suggest that even in these mammals- which have a highly reduced number of non-rod, non-cone opsins- melanopsin+ RGCs are involved in vision (see Annu Rev Vis Sci. 2020 Sep 15;6:453-468. doi: 10.1146/annurev-vision-030320-041239 for a recent review). This aspect needs to be addressed in their optokinetic response gain covariant analyses.

6) Spectral information on the LEDs needs to be provided (in photons per area and time).

eLife. 2021 Jun 8;10:e63355. doi: 10.7554/eLife.63355.sa2

Author response


The reviewers and the Reviewing Editor agreed that the design of the set-up presents a technical innovation that will be highly useful for future studies. There is clear disagreement on the level of conceptual insight.

Based on the technical advancement we agreed on a major revision of the manuscript. However, the resubmission should not be handed in as a research article, but as a Tools and Resources article.

Please address our major concerns listed below.

1) Statistical analysis. For several results only means are shown, and often there does not appear to be an indication of the variance between individuals. For instance:

– In Figures3 and 4 it would be interesting to know how this varies across animals. One possibility would be to also plot a variance map. Another would be to include the maps for each animal individually as supplementary data.

We chose the latter option suggested by the reviewer and have now included the OKR gain maps for each individual animal as our new Figure 4—figure supplement 2. Inspection of single-fish data led us to conclude that data from one individual from one of the control groups (fish no. 1 from the rotated arena group) had been corrupted. The seemingly arbitrary spatial arrangement of OKR gains likely indicates a mistake by the experimenter, such as misidentifying the stimulus presentation order, and cannot be corrected post-hoc. For maximum transparency, we still show the results from this one fish in the first panel of Figure 4—figure supplement 2b, but we no longer include them in further analyses.

In the process of revisiting our OKR gain maps, we realised that while our von-MisesFisher (vMF) fits accurately identify the centres of mass of the spatial distribution of OKR gains, their appearance tends to mislead readers about the details of the OKR gain distribution away from these points. Most importantly, frontal stimuli (azimuth and elevation 0 degrees) still evoke intermediate-magnitude OKR, but were previously shown in similar colour as polar stimuli (elevation near +/- 90 degrees), which evoke much weaker OKR. To avoid misleading viewers, we now replaced our previous heat maps (showing only the best vMF fit to data) with more informative heat maps (showing the raw OKR gain data filtered with a vMF kernel and using a logarithmic colour map). Figure 3, Figure 4 and Figure 4—figure supplement 2 have been changed accordingly.

– In Figure 5 error bars should be shown in d,e, and also perhaps the plots for individual animals as supplementary data.

Agreed.

Action taken: We have now included error bars in Figure 5 D-E. We have also created the new Figure 5—figure supplement 2 (stimulus frequency) and Figure 5—figure supplement 3 (stimulus area) showing the data from individual animals.

– Line 123: how variable is the visual field size? The current writing suggests it's exactly 163 degrees in every fish.

While there is some evidence that visual field size increases early development, little is known about variability between individuals. The value of 163 degrees cited refers to that reported by Easter and Nicola, 1996 Dev Biol, for 4 dpf larvae. The same authors report a smaller visual field for younger fish, namely 139 degrees at 3 dpf. As all our fish were older than 4 dpf, we assume their visual fields to cover at least 163 degrees.

Action taken: To avoid giving the false impression of an exact known value across fish, we have now rephrased the sentence. It reads:

“Zebrafish are lateral-eyed animals and have a large visual field, which increases during early development and in individual 4day old larvae has been reported to cover about 163° per eye (17), though little is known about interindividual variability.”

– Line 248: What is the variance over animals of 80.3 degrees?

– Line 289-292: What is the variance in these numbers?

Because each individual fish can only be tested for a few hours at a time, and because their rapid development precludes repeated testing on subsequent days, each stimulus type could only be shown twice to each animal (that is, for 2x100 seconds). We thus obtain only two gain values per stimulus per fish, so individual-fish data is inherently variable, and only part of this variability represents genuine biological interindividual variability. We thus made the conscious choice not to report “variance over animals” in this context, and instead pooled the data across fish before further analysis. After pooling and averaging across fish, results from the main experimental group were further corrected with those from one of our control groups (“rotated arena”). Only then did we apply our fitting procedure, which was designed for this specific purpose, and may not be reliable for the much noisier and uncorrected individual-fish data. To provide the reviewer with the requested numbers we nonetheless applied the procedure, and provide the resulting best-fit coordinates for the reviewers’ discretion, but caution against their reliability.

Action taken: No changes to the manuscript itself, but we computed best fits to individual-fish data to better inform the reviewers (see Author response table 1 and Author response table 2 listing angles in degrees, and in a body-centred reference frame):

Author response table 1.

regular arena azimuth 1 elevation 1 azimuth 2 elevation 2
fish no. 1 -77.6 7.5 70.7 9.2
fish no. 2 -78.7 6.4 52.9 13.2
fish no. 3 -76.5 8.2 64.7 8.0
fish no. 4 -78.7 6.4 52.9 13.2
fish no. 5 -77.9 5.3 92.6 0.0
fish no. 6 -90.7 3.3 84.8 -2.9
fish no. 7 -83.2 9.9 91.9 -3.3

Author response table 2.

rotated arena azimuth 1 elevation 1 azimuth 2 elevation 2
fish no. 1* -27.4 34.2 141.5 -39.6
fish no. 2 -64.9 10.3 71.7 -9.6
fish no. 3 -76.3 0.9 86.7 4.5
fish no. 4 -97.9 7.6 73.1 -6.6
fish no. 5 -93.3 -1.9 76.9 4.8

*fish excluded from all analyses as explained above and in manuscript

– Line 299: please show us what the distribution for the permutation test looks like.

Yes, we agree that this information is helpful. The reviewer’s request also prompted us to notice that a detailed verbal description of the permutation test was missing from the Methods section.

Action taken: We have now created a new Figure 4—figure supplement 3 to show the distribution, and reference it from the main text. We also now include a verbal description in the Methods section.

2) The authors argue that "extra-retinal processing" is responsible for the minor shift in the center of the highest gain OKR spot. However, while the effect indeed seems likely to require extra-retinal input, the authors provide no evidence that the processing mediating this should occur outside of the retina. Given other examples of non-visual information being sent to the retina, it is at least plausible that retinal processing is involved.

– Line 79: The evidence for it being "extra-retinal processing" is lacking. The source of the effect is likely extra-retinal but whether the processing happens in the retina or elsewhere is not shown.

– 182: The "determinants" are indeed likely extra-retinal, but this does not mean the processing is.

– Line 428: You have not demonstrated that the processing is extra-retinal, only that it is not occurring in retinal coordinates. There could be a projection into the retina, for example.

It is true that such additional retinal processing could explain the effect we observe, even though no evidence for such functional projections has so far been found. But this does indeed leave both hypotheses unfalsified.

Action taken: We have rephrased our discussion to suggest both extra-retinal processing and putative projections onto the retina as valid hypotheses. We now write:

“In addition, further processing appears to affect the preferred OKR stimulus location, as suggested by the experiments in upside-down embedded animals. This could include extra-retinal processing, or hitherto unknown projections to the retina itself.”

3) The authors should consider three coordinate systems explicitly, particularly in the lines 273-303. These would be retinal coordinates, head coordinates and environmental coordinates.

The reviewer is correct that this paragraph was not sufficiently explicit regarding the possible coordinate systems.

Action taken: We have revised the text accordingly to now include discussion of all three reference frames (see next point).

– Line 277: It seems there's also another representation which is missing from your consideration here – a body centric representation. This is distinct from retinal coordinates and distinct from environmental coordinates but is known from primate studies, for example.

Indeed, this is a crucial point we need to clarify. Our fish are immobilised, upright, and oriented towards 0 azimuth, 0 elevation. While the eyes themselves can move in their sockets, the head cannot move relative to the body. The “fish-centred coordinates” (body-centred coordinates) we discuss on line 278 are thus identical to head-centred coordinates in our experiments.

Action taken: We now present our results in the context of these four reference systems (body-, head-, retina-, and world-centred representations) in the Results section, and also added further detail in the Methods section. In the Results section, we now write:

“A priori, it is unclear whether the sampling preference originates from the peculiarities of the sensory periphery in the eye, or the behavioural relevance inferred by central brain processing. […] To start distinguishing these possible…” as well as:

“Eye-, head-, or body- centred reference frames are therefore not sufficient to fully explain the observed optokinetic stimulus location preferences. Instead, the stimulus location preference is additionally related to the environmental reference frame, suggesting that the behavioural relevance of our motion stimuli depends on their environmental elevation positions.”

In the Methods section, we added:

“Geographic coordinates can be used to identify stimulus position in at least four distinct reference frames: body-centred, head-centred, retina-centred, or world-centred coordinates. […] Therefore, our comparison of the OKR gain map to retinal photoreceptor densitiy distributions is based on the time-averaged eye position during optokinetic stimulation.”

– Line 282. Now retina-centered is missing. (the retina can move independently from the head/body).

The reviewer is correct that we didn’t discuss a possible retina-centered reference frame at this position. The eye position varies continuously during OKR, while the body – at least in our immobilised fish – remains stationary. To convincingly show OKR gain maps in a retina-centred reference frame, the stimulus position should ideally be updated in the experiment based on the current eye position, which we didn’t do. We did find, however, that the average eye position (the beating field) does not depend on stimulus position as described in the manuscript. Therefore, across different stimulus phases, the retina-centred reference frame is similar to the bodycentred reference frame.

Our discussion of fish- and world-centred coordinates on line 282 is in the context of identifying stimulus locations that evoke maximum OKR gain. We later identify the part of the retina that was stimulated by this “optimal” stimulus while the eye was at its average position. And we observe that this retinal location – as well as its surroundings that would be stimulated when the eye is outside its average position – correspond to high photoreceptor densities.

Action taken: We now discuss all possible reference systems explicitly in the Results section (also see previous point).

4) Especially as we consider this manuscript rather an advancement for the "Tools and Resources" section, please consider that the Stowers J et al. 2017 "Virtual Reality for Freely Moving Animals" was a full-surround setup – at least as defined by the standard for the other citations – used successfully on Drosophila and, importantly here, fish. You need to be more specific in your wording (e.g. to limit statements to systems based on LEDs capable of producing the desired spectral content) or cite that paper.

The published method suggested by the reviewer is indeed a valuable contribution and should be discussed explicitly in our manuscript. The FreemoVR platform is highly versatile, but differs from our solution in several aspects that are important to studies of zebrafish visual function (visual field coverage, light refraction, interoperability with laser-scanning microscopes). This merits a qualified discussion that was so far absent from our manuscript.

Action taken: We now included a reference to Stowers et al. 2017, and briefly compare and contrast it to our own approach. We write:

“At least one existing solution for stimulating and tracking freely moving zebrafish supports unusually large stimuli, though despite its versatility, it still only covers part of the visual field and does not address the remaining issues of total internal reflection and the interoperability with laser-scanning microscopes (43).”

5) The statement that the gain of the OKR broadly follow the density of the longer-wavelength sensitive photoreceptors in the retina is entirely ignoring the fact that the fish retina is full of Opsins in all kind of different cells types. This includes different melanopsins, but also VA-opsins (which are likely sensitive in the green), opn5s (UVA), tmt-opsins etc. Evidence from mouse and humans starts to suggest that even in these mammals- which have a highly reduced number of non-rod, non-cone opsins- melanopsin+ RGCs are involved in vision (see Annu Rev Vis Sci. 2020 Sep 15;6:453-468. doi: 10.1146/annurev-vision-030320-041239 for a recent review). This aspect needs to be addressed in their optokinetic response gain covariant analyses.

This is indeed an interesting aspect gaining increased recognition in the field. While the good match between a higher density of longer-wavelength photoreceptors and higher OKR gains could be explained without regard to other, previously overlooked opsins, we should still address their potential role in our manuscript.

Action taken: We have now expanded our discussion to acknowledge the putative role of other opsin-expressing retinal neurons. We write:

“Recent studies suggest that melanopsin-expressing ganglion cells also contribute to form vision (59, 60), and melanopsins as well as a plethora of other opsins are expressed in the zebrafish retina and brain (61, 62). It is therefore possible that photoreception via opsin-expressing neurons in the inner retina contributes to zebrafish OKR as well.”

6) Spectral information on the LEDs needs to be provided (in photons per area and time).

Yes, we agree that information on the brightness of the stimulus should be included. This will supplement the provided spectral information regarding peak wavelength and LED type in the manuscript and thereby facilitate comparison to the results of future studies.

Action taken: We have measured the luminance (10.94 ± 0.59 𝑐𝑑 𝑚−2) and now report it in the Methods section.

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Data Citations

    1. Dehmelt FA, Meier R, Hinz J, Yoshimatsu T, Simacek CA, Wang K, Baden T, Arrenberg AB. 2020. Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish. GIN (G-Node Infrastructure) [DOI]
    2. Zimmermann MJY, Nevala NE, Yoshimatsu T, Osorio D, Nilsson DE, Berens P, Baden T. 2019. Data from: Zebrafish differentially process colour across visual space to match natural scenes. Dryad Digital Repository. [DOI] [PubMed]
    3. Dehmelt FA, Meier R, Hinz J, Yoshimatsu T, Simacek CA, Huang R, Wang K, Baden T, Arrenberg AB. 2020. Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish. G-Node. [DOI]

    Supplementary Materials

    Figure 1—source data 1. SCAD files for 3D-printing the arena scaffold.
    Figure 3—source data 1. Numerical data and graphical elements of Figure 3b–d.
    Figure 3—source data 2. Numerical data and graphical elements of Figure 3e.
    Figure 4—source data 1. Numerical data and graphical elements of Figure 4e.
    Figure 4—source data 2. Numerical data and graphical elements of Figure 4f.
    Figure 4—source data 3. Numerical data and graphical elements of Figure 4g.
    Figure 4—source data 4. Numerical data and graphical elements of Figure 4h.
    Figure 5—source data 1. Numerical data and graphical elements of Figure 5.
    Figure 6—source data 1. Numerical data and graphical elements of Figure 6.
    Source code 1. Five separate sets of MATLAB code.

    (A) Code to design visual stimuli, convert them to geographic coordinates, and map them onto the actual position of individual LEDs. (B) Code to numerically identify a distribution of nearly equidistant stimulus centres that is symmetric between the left and right, upper and lower, as well as front and rear hemispheres. (C) Code to read raw eye traces, identify individual stimulus phases, detect and remove saccades, compute piece-wise fits to cropped and pre-processed raw data, and return OKR gains. (D) Code to recreate the results figures from data, especially Figure 3, Figure 5, Figure 6. (E) Code to assess the significance of differences between the best fits to data for fish embedded upright or upside-down. This last set of code requires access to the raw data repository.

    elife-63355-code1.zip (148KB, zip)
    Supplementary file 1. Seven supplementary tables.

    (A) The arena cross-section, (B) stimulus parameters for position dependence experiments, (C) absolute positions of fish and setup elements in different experiments, (D) stimulus parameters for control experiments, (E) parameters for whole-field and hemisphere stimuli, (F) stimulus parameters for frequency dependence experiments, and (G) stimulus parameters for size dependence experiments.

    elife-63355-supp1.docx (66.2KB, docx)
    Transparent reporting form

    Data Availability Statement

    Analysis code, pre-processed data and examples of raw data have been deposited in GIN by G-Node and published (https://doi.org/10.12751/g-node.qergnn).

    The following dataset was generated:

    Dehmelt FA, Meier R, Hinz J, Yoshimatsu T, Simacek CA, Wang K, Baden T, Arrenberg AB. 2020. Gaze stabilisation behaviour is anisotropic across visual field locations in zebrafish. GIN (G-Node Infrastructure)

    The following previously published dataset was used:

    Zimmermann MJY, Nevala NE, Yoshimatsu T, Osorio D, Nilsson DE, Berens P, Baden T. 2019. Data from: Zebrafish differentially process colour across visual space to match natural scenes. Dryad Digital Repository.


    Articles from eLife are provided here courtesy of eLife Sciences Publications, Ltd

    RESOURCES