Abstract
Bees, wasps and ants navigate successfully between feeding sites and their nest, despite the small size of their brains which contain less than a million neurons. A long history of studies examining the role of visual memories in homing behavior show that insects can localize a goal by finding a close match between a memorized view at the goal location and their current view (“snapshot matching”). However, the concept of static snapshot matching might not explain all aspects of homing behavior, as honeybees are able to use landmarks that are statically camouflaged. In this case the landmarks are only detectable by relative motion cues between the landmark and the background, which the bees generate when they perform characteristic flight maneuvers close to the landmarks. The bees' navigation performance can be explained by a matching scheme based on optic flow amplitudes (“dynamic snapshot matching”). In this article, I will discuss the concept of dynamic snapshot matching in the light of previous literature.
Key words: insect navigation, view-based homing, snapshot matching, vision
Goal-Seeking in Insects
Insects, with their miniature brains, have evolved a simple strategy to find both their nests and profitable food sources. They approach them by “snapshot matching”, i.e., finding a close match between the current view and a memorized view of the goal location (reviewed in ref. 1 and 2). How is the relevant information encoded in a snapshot? There are currently two main views: (1) Insects may extract the relevant visual features, such as the contours of prominent objects which define the goal location.3 In addition to retinal size and the position of these objects, honeybees also use color and distance cues as features to localize a food source.3–8 Currently, it is still unclear how many and what features are stored in the retinotopic snapshot. (2) A fairly new idea, which does not involve the extraction of certain features and object identification, is the “global image matching” method.9 This means that the insects memorize a raw panoramic image at the goal location. The image implicitly contains all important static visual features of the goal location. The insects then get back to their goal location by moving in a way that increases the similarity of the current view with the memorized image. Recent model implementations on robotic platforms show that this method works.9–14 Even for view-based homing in natural environments, Zeil and colleagues showed that panoramic image similarities can be used.9,15 More recently, the behavior of ants and crickets in goal finding tasks could be explained by “global image matching”.16,17
Dynamic Snapshots
Within the insect navigation literature, snapshot matching is presently often pictured as a static process, which involves three steps: stopping, comparing the current view with the memorized view at the goal, and then moving to increase the similarity between the views. However, our study shows that static snapshot matching might not explain all aspects of homing behavior as honeybees are able to use landmarks that are statically camouflaged.18 Honeybees, trained to locate an inconspicuous feeder surrounded by three high contrast landmarks (Fig. 1A), were tested either with one landmark removed or with three landmarks with the same contrast and texture as the background (Fig. 1B). Removing a landmark affected the search performance of the honeybees. Surprisingly, the navigational performance was not impaired when the landmarks were camouflaged.18 In this case the landmarks could not be detected by their contrast and texture, and if the bees would had memorized a static image of the scene they would not have been able to find the goal. These landmarks only become visible when the bee moves in a characteristic way (see Fig. 1C inset).
Figure 1.

In our study honeybees were trained to visit a perspex feeder surrounded by three landmarks (diameter of 5 cm, height of 25 cm) placed at different distances (10, 20, 40 cm) from the feeder in a circular flight arena (diameter of 195 cm; height of 50 cm).18 (A) During training the landmarks were covered with a homogeneous texture providing luminance contrast to the background. (B) After being trained with three homogenous textured landmarks, honeybees were tested individually with landmarks that had the same random dot texture as the background of the arena. (C) The top view shows an example flight trajectory of a honeybee. Approach flights to the feeder (black circle) surrounded by three landmarks (grey circles) were recorded with three high-speed cameras at 125 fps. The position of the bee is indicated by grey circles at each 32 ms interval; straight lines indicate the orientation of the long axis of the bee. The inset shows a sideways-directed flight manoeuvre of a honeybee close to the landmark (two times magnified; Reproduced from ref. 18). (D) shows an optic flow snapshot calculated at the feeder position. White indicates high optic flow amplitudes. The landmarks “pop out” of the background although they have the same texture as the background (B). From a 3D model of the arena, panoramic images were rendered at 1°/pixel resolution. Optic flow fields were generated by simulated translations of 2 mm in the flight arena, obtaining four images at (x ± 1 mm, y), (x, y ± 1 mm). The optic flow for these movements was computed using a modified version of the Lucas-Kanade algorithm. To generate flow amplitudes independent of the direction of motion, flow fields for the two steps (in x and y) were squared and summed.18 The optic flow snapshot implicitly contains information about the depth structure of the scene, as closer objects move faster and create a higher optic flow amplitude. This can be seen by comparing the near landmark in the centre of the picture to the middle far (left) and the far (right) landmark.
We found that honeybees perform scanning movements in the vicinity of the landmarks.18 Through these sideways maneuvers in front of the landmarks, they become visible by relative motion. Furthermore, they employ a flight style that facilitates depth perception from motion parallax. The bees' trajectories consist of straight flight segments combined with rapid turns.18,19 Between turns, gaze stabilization leads to a behavioral elimination of rotational components from the optical flow pattern,19,20 making it easier to use the remaining translational optic flow for homing, as it contains range information. This is because images of close objects move faster across the retina than those of more distant objects. In our experiments with landmarks that had the same texture as the background this helped detection and possibly also estimation of the distance to landmarks as they ‘pop out’ when the animals move in front of them. In other behavioral tasks, motion parallax is used as a cue, e.g., for shape discrimination or distance estimation.21,22
But what does a snapshot look like, and what information is stored? By means of model simulations, we made a first attempt to answer this question. Global image matching cannot explain the search behavior with statically camouflaged landmarks. Taking the honeybees movement into account, we simulated the motion pattern on the bees' eyes during pure translational movements at the goal location (Fig. 1D). By memorizing this optic flow snapshot and comparing it with the optic flow fields at other locations in the flight arena, the honeybees could navigate even with camouflaged landmarks.18 Our study emphasises that snapshot matching should be considered a dynamic rather than a purely static process. Taking into account how the insects interact with their environment helped us to identify possible mechanisms of navigation.
Contents and Extent of Retinotopic Snapshots
Neither a static nor a dynamic snapshot can explain all the results obtained from navigation experiments. When trained with different colored landmarks, honeybees use also color cues.4 Perhaps a snapshot consists of landmark boundaries, which can be detected by luminance, color or motion contrast. You might also think of a combination of optic flow and texture information in an extended snapshot scheme.
The selection and weighting of static or dynamic information might vary among different species and depend on the kind of locomotion the insect performs. Walking insects (e.g., ants) can more easily stay motionless relative to their surroundings (an advantage for static image matching) whereas flying insects (e.g., bees) can fly sideways (an advantage if using motion parallax information). There might also exist species-dependent differences in whether the insect memorizes a snapshot of the full field of view and whether all regions of the visual field are weighted equally. Ants store a snapshot while fixating a landmark with their frontal field of view, which extends at least 120° into the periphery.23 When trained with several landmarks surrounding the nest or food source, they focus on their frontal field of view and employ a sequential matching strategy of single landmarks.24,25 If the landmark array is enlarged without changing the size of the landmarks, the ants do not search where the retinal position of the landmarks fits but where the retinal size of a single landmark fits.25–27 In contrast, bees tested under similar conditions seem to match the retinal positions of all landmarks, if there is a conflict with the retinal size of the landmarks.3 But although the snapshot is centred on one landmark in ants, it was concluded that they identify landmarks with the aid of the background pattern, indicating that the panorama is somehow included in their representation and might help to enhance the reliability of the snapshot recall.27 An experiment, in which different species are tested with the same behavioral task, might provide insights into the relevance and weighting of different navigational mechanisms and the contents and size of snapshot memories.
A Single Snapshot? Where and When?
Despite a series of studies on snapshot matching in insects, there remain some crucial questions to be answered. Today, it is still unclear how many snapshots the insects memorize and where they take these snapshots. It might be that there is no universal answer to these questions. How snapshots are used for navigation probably depends on the species, and even on the task or the learning state of the insect. Nevertheless, differences can teach us something about the underlying navigational mechanisms and their constraints. Although agents can navigate successfully to a goal location with the aid of one snapshot close to the goal, e.g.12 there is evidence that ants, for example, memorize several snapshots and use them for returning to the goal location.28,29 The current idea is that snapshots are taken during learning walks in ants and during learning flights in bees and wasps.30–33 These learning phases have a common feature: when leaving the nest or a newly discovered food source, the insects turn back and perform distinct maneuvers. The learning walks of ants include stopping phases, in which they fixate nearby landmarks (wood ants24) or the goal itself (Namibian desert ants33) and approach the target for a short time. The acquisition of memories is thought to take place during these clear-cut events of the learning walks. Wasps and honeybees are thought to memorize snapshots while fixating the goal at the end of arcs (inspection points), when they turn back and look at the feeder during their learning flights.30 Assuming that honeybees would indeed memorize an optic flow snapshot, it could take place during the translatory phases of flight when they face their goal.19
These straight segments of the learning flights provide them with spatial information, e.g., the distance to the landmarks, which fits to the long standing interpretation that learning flights help the insects to identify close landmarks and learn their spatial relationship.31,32,34–36 If insects memorize snapshots during these learning maneuvers, they should follow a somewhat similar trajectory in their next return to the goal. But instead the similarity of viewing directions and flight paths between learning and return flights seems to differ between species and is less striking in honeybees.30,32,37,38 On the other hand, the insects might replicate some of the relevant movement sequences without following the same paths. The motion dynamics of learning and return flights are similar in bees19 and wasps (Boeddeker personal communication). The functional connection between learning walks or flights and returns still has to be understood. Analysing the detailed movements of the insects during navigation should give further insights. In a recent study, Lent and colleagues analyzed the movements of ants during the return to a goal in detail.39 These wood ants perform saccadic-like body turns to match the learned visual features. The turns depend on the difference between the desired and current retinal positions of a visual feature, and they are also used to correct errors after the ants have drifted off the route. Although this provides us with an idea how snapshots are used to guide the ants' path it is yet unclear, how this connects to their learning walks. Do insects memorize all possible snapshots during their learning maneuvers? How are these snapshots used to structure the return? Learning maneuvers only occur in the early phase of foraging and even the return paths change over time.6,40 It might well be that several snapshots are somehow integrated into one spatial representation of the goal environment over time, allowing the insect to return reliably but at the same time to navigate flexibly to the goal location. I suggest tackling these questions using comparative studies and by analysing the insects' behavior in detail (e.g. ref. 38, 39, 41, reviewed in ref. 42). Identifying differences in the navigational strategies depending on the species, the task, the complexity of the visual environment, and the learning state of the insect, will help us understand the relevance and the constraints of the underlying navigational mechanisms.
Acknowledgements
I thank Norbert Boeddeker, Martin Egelhaaf and Nicole Carey for their helpful comments on the manuscript and Wolfgang Stürzl for providing the “optic flow snapshot” image. Financial support came from the DFG.
References
- 1.Collett TS, Graham P, Harris RA, de Ibarra NH. Navigational memories in ants and bees: Memory retrieval when selecting and following routes. Adv Study Behav. 2006;36:123–172. [Google Scholar]
- 2.Zeil J, Boeddeker N, Stürzl W. Visual homing in insects and robots. In: Floreano D, et al., editors. Flying insects and robots. New York, NY: Springer; 2009. pp. 87–100. [Google Scholar]
- 3.Cartwright BA, Collett TS. Landmark learning in bees—experiments and models. J Comp Physiol. 1983;151:521–543. [Google Scholar]
- 4.Cheng K, Collett TS, Wehner R. Honeybees learn the colors of landmarks. J Comp Physiol A. 1986;159:69–73. [Google Scholar]
- 5.Cheng K, Collett TS, Pickhard A, Wehner R. The use of visual landmarks by honeybees—bees weight landmarks according to their distance from the goal. J Comp Physiol A. 1987;161:469–475. [Google Scholar]
- 6.Fry SN, Wehner R. Look and turn: landmark-based goal navigation in honey bees. J Exp Biol. 2005;208:3945–3955. doi: 10.1242/jeb.01833. [DOI] [PubMed] [Google Scholar]
- 7.Cartwright BA, Collett TS. How honeybees know their distance from a near-by visual landmark. J Exp Biol. 1979;82:367–372. [Google Scholar]
- 8.Lehrer M, Collett TS. Approaching and departing bees learn different cues to the distance of a landmark. J Comp Physiol A. 1994;175:171–177. [Google Scholar]
- 9.Zeil J, Hofmann MI, Chahl JS. Catchment areas of panoramic snapshots in outdoor scenes. J Opt Soc Am A. 2003;20:450–469. doi: 10.1364/josaa.20.000450. [DOI] [PubMed] [Google Scholar]
- 10.Franz MO, Schölkopf B, Mallot HA, Bülthoff H. Where did I take that snapshot? Scene-based homing by image matching. Biol Cybern. 1998;79:191–202. [Google Scholar]
- 11.Stürzl W, Mallot HA. Efficient visual homing based on Fourier transformed panoramic images. Robot Auton Syst. 2006;54:300–313. [Google Scholar]
- 12.Vardy A, Möller R. Biologically plausible visual homing methods based on optical flow techniques. Connect Sci. 2005;17:47–89. [Google Scholar]
- 13.Möller R, Vardy A. Local visual homing by matched-filter descent in image distances. Biol Cybern. 2006;95:413–430. doi: 10.1007/s00422-006-0095-3. [DOI] [PubMed] [Google Scholar]
- 14.Möller R. Local visual homing by warping of two-dimensional images. Robot Auton Syst. 2009;57:87–101. [Google Scholar]
- 15.Stürzl W, Zeil J. Depth, contrast and view-based homing in outdoor scenes. Biol Cybern. 2007;96:519–531. doi: 10.1007/s00422-007-0147-3. [DOI] [PubMed] [Google Scholar]
- 16.Wystrach A, Beugnon G. Ants learn geometry and features. Curr Biol. 2009;19:61–66. doi: 10.1016/j.cub.2008.11.054. [DOI] [PubMed] [Google Scholar]
- 17.Mangan M, Webb B. Modelling place memory in crickets. Biol Cybern. 2009;101:307–323. doi: 10.1007/s00422-009-0338-1. [DOI] [PubMed] [Google Scholar]
- 18.Dittmar L, Stürzl W, Baird E, Boeddeker N, Egelhaaf M. Goal seeking in honeybees: matching of optic flow snapshots? J Exp Biol. 2010;213:2913–2923. doi: 10.1242/jeb.043737. [DOI] [PubMed] [Google Scholar]
- 19.Boeddeker N, Dittmar L, Stürzl W, Egelhaaf M. The fine structure of honeybee head and body movements in a homing task. Proc Biol Sci. 2010;277:1899–1906. doi: 10.1098/rspb.2009.2326. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Boeddeker N, Hemmi JM. Visual gaze control during peering flight manoeuvres in honeybees. Proc Biol Sci. 2010;277:1209–1217. doi: 10.1098/rspb.2009.1928. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Lehrer M, Campan R. Generalization of convex shapes by bees: what are shapes made of? J Exp Biol. 2005;208:3233–3247. doi: 10.1242/jeb.01790. [DOI] [PubMed] [Google Scholar]
- 22.Srinivasan MV, Zhang SW. Visual motor computations in insects. Annu Rev Neurosci. 2004;27:679–696. doi: 10.1146/annurev.neuro.27.070203.144343. [DOI] [PubMed] [Google Scholar]
- 23.Durier V, Graham P, Collett TS. Snapshot memories and landmark guidance in wood ants. Curr Biol. 2003;13:1614–1618. doi: 10.1016/j.cub.2003.08.024. [DOI] [PubMed] [Google Scholar]
- 24.Nicholson DJ, Judd SPD, Cartwright BA, Collett TS. Learning walks and landmark guidance in wood ants (Formica rufa) J Exp Biol. 1999;202:1831–1838. doi: 10.1242/jeb.202.13.1831. [DOI] [PubMed] [Google Scholar]
- 25.Narendra A, Si A, Sulikowski D, Cheng K. Learning, retention and coding of nest-associated visual cues by the Australian desert ant, Melophorus bagoti. Behav Ecol Sociobiol. 2007;61:1543–1553. [Google Scholar]
- 26.Wehner R, Räber F. Visual spatial memory in desert ants, Cataglyphis bicolor (Hymenoptera, Formicidae) Experientia. 1979;35:1569–1571. [Google Scholar]
- 27.Graham P, Durier V, Collett TS. The binding and recall of snapshot memories in wood ants (Formica rufa L.) J Exp Biol. 2004;207:393–398. doi: 10.1242/jeb.00771. [DOI] [PubMed] [Google Scholar]
- 28.Judd SPD, Collett TS. Multiple stored views and landmark guidance in ants. Nature. 1998;392:710–714. [Google Scholar]
- 29.Harris RA, Graham P, Collett TS. Visual cues for the retrieval of landmark memories by navigating wood ants. Curr Biol. 2007;17:93–102. doi: 10.1016/j.cub.2006.10.068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Collett TS, Lehrer M. Looking and learning—a spatial pattern in the orientation flight of the wasp vespulavulgaris. P Roy Soc Lond B Biol. 1993;252:129–134. [Google Scholar]
- 31.Zeil J. Orientation flights of solitary wasps (Cerceris, Sphecidae, Hymenoptera): I. Description of flight. J Comp Physiol A. 1993;172:189–205. [Google Scholar]
- 32.Zeil J. Orientation flights of solitary wasps (Cerceris, Sphecidae, Hymenoptera): II. Similarities between orientation and return flights and the use of motion parallax. J Comp Physiol A. 1993;172:207–222. [Google Scholar]
- 33.Müller M, Wehner R. Path integration provides a scaffold for landmark learning in desert ants. Curr Biol. 2010;20:1368–1371. doi: 10.1016/j.cub.2010.06.035. [DOI] [PubMed] [Google Scholar]
- 34.Lehrer M. Why do bees turn back and look? J Comp Physiol A. 1993;172:549–563. [Google Scholar]
- 35.Zeil J, Kelber A, Voss R. Structure and function of learning flights in bees and wasps. J Exp Biol. 1996;199:245–252. doi: 10.1242/jeb.199.1.245. [DOI] [PubMed] [Google Scholar]
- 36.Zeil J. The control of optic flow during learning flights. J Comp Physiol A. 1997;180:25–37. [Google Scholar]
- 37.Collett TS. Making learning easy—the acquisition of visual information during the orientation flights of social wasps. J Comp Physiol A. 1995;177:737–747. [Google Scholar]
- 38.de Ibarra NH, Philippides A, Riabinina O, Collett TS. Preferred viewing directions of bumblebees (Bombus terrestris L.) when learning and approaching their nest site. J Exp Biol. 2009;212:3193–3204. doi: 10.1242/jeb.029751. [DOI] [PubMed] [Google Scholar]
- 39.Lent DD, Graham P, Collett TS. Image-matching during ant navigation occurs through saccade-like body turns controlled by learned visual features. Proc Natl Acad Sci USA. 2010;107:16348–16353. doi: 10.1073/pnas.1006021107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Lehrer M. Bees which turn back and look. Naturwissenschaften. 1991;78:274–276. [Google Scholar]
- 41.Collett TS, Rees JA. View-based navigation in Hymenoptera: Multiple strategies of landmark guidance in the approach to a feeder. J Comp Physiol A. 1997;181:47–58. [Google Scholar]
- 42.Zeil J, Boeddeker N, Hemmi JM. Vision and the organization of behaviour. Curr Biol. 2008;18:320–323. doi: 10.1016/j.cub.2008.02.017. [DOI] [PubMed] [Google Scholar]
