Abstract
As animals move through the world in search of resources, they change course in reaction to both external sensory cues and internally-generated programs. Elucidating the functional logic of complex search algorithms is challenging because the observable actions of the animal cannot be unambiguously assigned to externally- or internally-triggered events. We present a technique that addresses this challenge by assessing quantitatively the contribution of external stimuli and internal processes. We apply this technique to the analysis of rapid turns (“saccades”) of freely flying Drosophila melanogaster. We show that a single scalar feature computed from the visual stimulus experienced by the animal is sufficient to explain a majority (93%) of the turning decisions. We automatically estimate this scalar value from the observable trajectory, without any assumption regarding the sensory processing. A posteriori, we show that the estimated feature field is consistent with previous results measured in other experimental conditions. The remaining turning decisions, not explained by this feature of the visual input, may be attributed to a combination of deterministic processes based on unobservable internal states and purely stochastic behavior. We cannot distinguish these contributions using external observations alone, but we are able to provide a quantitative bound of their relative importance with respect to stimulus-triggered decisions. Our results suggest that comparatively few saccades in free-flying conditions are a result of an intrinsic spontaneous process, contrary to previous suggestions. We discuss how this technique could be generalized for use in other systems and employed as a tool for classifying effects into sensory, decision, and motor categories when used to analyze data from genetic behavioral screens.
Author Summary
Researchers have spent considerable effort studying how specific sensory stimuli elicit behavioral responses and how other behaviors may arise independent of external inputs in conditions of sensory deprivation. Yet an animal in its natural context, such as searching for food or mates, turns both in response to external stimuli and intrinsic, possibly stochastic, decisions. We show how to estimate the contribution of vision and internal causes on the observable behavior of freely flying Drosophila. We developed a dimensionality reduction scheme that finds a one-dimensional feature of the visual stimulus that best predicts turning decisions. This visual feature extraction is consistent with previous literature on visually elicited fly turning and predicts a large majority of turns in the tested environment. The rarity of stimulus-independent events suggests that fly behavior is more deterministic than previously suggested and that, more generally, animal search strategies may be dominated by responses to stimuli with only modest contributions from internal causes.
Introduction
Active movement is one of the defining features of animals, and the use of locomotion to search for resources within the environment is likely among the most ancient of behaviors. Observations on motile organisms, ranging in scale from bacteria to whales, indicate that search patterns are structured by a combination of internal processes and external cues [1], [2]. Sensory systems enable organisms to detect favorable objects at a great distance [3]–[5] and they use this ability to localize resources by either directed motion (taxis) or changes in locomotor statistics (kinesis). Prior research suggests that, in the absence of external cues, the animal behavior is generated by internal processes, and that the overall animal fitness is sensitive to the exact characteristics of this internal process (e.g., Levy statistics) [6]–[18]; it has also been questioned whether observed large-scale statistics can give any insight on an internal process that generated the behavior, and whether the internal processes can dominate over stimuli-elicited behavior [19]–[22]. As for the internal processes, these can be divided into truly stochastic sources, and deterministic results of a deliberate, but unobservable, internal mechanism based on internal metabolic/neural states. When observing an intact motile organism, it is not easy to determine which components of its locomotion behavior are triggered by internal processes versus external cues, yet such classification is essential for deciphering the underlying logic of its movement and search behavior. The task is further complicated by the fact that an external observer might not be able to distinguish between truly stochastic processes and the deterministic results of a deliberate, but unobservable, internal mechanism. For example, software pseudo-random number generators produce strictly deterministic sequences, which appear to be random to an external observer who does not have access to the internal state of the system [23]. A major goal of both cell biology and neuroscience is explaining the molecular and cellular bases of these three qualitatively different processes (sensory-driven, purely stochastic, and deterministically based on internal states).
If the salient features of the external world are known, it is possible to gain insight into sensory-driven behaviors through the use of sensory-response correlation [24]. The analysis of the internally-driven processes is much more challenging. Given uncertainty in measurement and the inability to perfectly reproduce experimental conditions from trial to trial, variability in the results of behavioral experiments has often been treated as a limit on our ability to measure stimulus-driven behavior. In this view, variability in responses from trial to trial reflects irrelevant components of behavior, which are averaged until the mean—interpreted as the response the animal ideally would have produced—becomes clear [25]. From the opposite perspective, many researchers have attempted to artificially remove all relevant sensory input to an animal and measure behaviors in conditions of sensory deprivation to reveal intrinsic properties, especially the statistical distributions of behaviors [26]–[28]. Although focusing in isolation on either the stochastic [9], [29], [26] or the sensory components [30] of search behavior have provided key insights, neither of these extremes is sufficient to capture the full range of processes at play as an animal moves under natural conditions. Attempts to investigate the interaction of internal and external processes include studies of bacteria [31] and nematode worms [32], [33], organisms for whom chemicals provide the most salient cues for food search. For larger animals with image-forming eyes, vision may provide another essential cue in search algorithms, because vision is the only sense which allows to perceive remote parts of the environment. Often vision cannot be considered separately from the mechanics of locomotion [34].
Flies are a model of computational efficiency and robustness, to date not equaled by artificial systems, which often seek to imitate nature [35], [36]. Much is known about fly vision [37], [38]. Since the pioneering work of Kennedy [39] and Mittelstadt [40], the behavioral responses of flies to experimenter-defined visual stimuli have been extensively investigated. Electrophysiological recordings have complemented and extended our knowledge of phenomena such as the neural basis of motion detection [41]–[44] and other key aspects of sensory processing, such as receptive field tuning [45]. However, there are many challenges in the identification of neural processing and how it produces complex behavior, especially as regards the characterization of “discrete” behaviors, such as the rapid turns (“saccades”) of Drosophila, which are the object of this study. In fact, many studies which offered complete characterization of the animal response are limited to “continuous” behavior, for which they provide linear (or “linearized”) models [46]–[49]; this allows using techniques such as linear system identification.
Identifying the neural causes for “discrete” behavior involves solving a different set of problems. Firstly, there are the problems of segmentation and classification of behaviors (including the definition of what “behavior” and “a behavior” are), for which it is often necessary the use of nonlinear machine learning methods [50]. Then, there is the problem of building models that can correlate the stimulus with the behavior(s). While it is possible to postulate models that also integrate well with our understanding of lower-level behavior [51], [52], it is not clear how such methods can be identified from the data. On the practical side, it is evident that discrete decisions, such as turning decisions, are meant to guide exploration and therefore should be investigated in naturalistic situations. This poses practical problems of tracking the animal position in a large environment, and it also precludes (at the current level of technology) the uses of direct neural recording. In fact, comparatively few attempts have been made to correlate parameters of visual stimulus with behavioral responses in unrestrained conditions [53]–[56].
In this work, we present an analysis that can quantitatively discriminate the effect of visual stimulus as opposed to internal processes in the generation of saccades in the fruit fly. Our conclusions are that visual stimulus has a dominant role. One important message of our work is that it is very difficult to identify models of complex behavior that can explain everything, often because insufficient data can be collected. Therefore, it is important to “search for simplicity” [57], for example by framing the problem as dimensionality reduction, and to use models that a posteriori can justify their assumptions. While we describe this analysis for visual processing in Drosophila, our goal is to construct a general method that can be used for other sensory systems, other animal species, or in the context of genetic screens.
Methods
Fly care and experimental treatment
Flies from the laboratory stock derived from 200 wild-caught females were reared on a 16 h:8 h light dark cycle under standard laboratory conditions. Three day old adult female flies were anesthetized with cold and individually housed within centrifuge tubes containing a moist tissue paper. Flies were starved (but provided with water) in the tubes for four to six hours before being released into the flight arena. Most flies would immediately begin flying, and we terminated tracking after the fly landed. We then removed each fly with a wand attachment of a vacuum cleaner before introducing another fly. Thus, each recorded trajectory is derived from a fly's initial experience exploring the novel environment.
The flight arena was a 2 meter diameter, 80 cm high cylinder (see Figure 1A). 10 cm×10 cm red and green gel filters (Roscolux) were attached to the arena in a regular checkerboard arrangement and provided a high contrast visual stimulus to flies near the wall. One meter from the wall (i.e., at the center of the arena), the angular wavelength of this pattern was ∼11°, and consequently would be twice the inter-ommatidial spacing of a ∼5.5° in Drosophila [58]. The particular red and green filters were chosen to have similar infrared transmission to facilitate tracking using cameras outfitted with long (IR) pass filters. The arena was illuminated from outside with a circular array of eight 750W Fresnel stage lights pointing towards the arena center. These lights provided both visible and infrared light for fly visual responses and machine vision tracking, respectively.
Fly tracking
A detailed description of our tracking system may be found in [59]. Briefly, we used 11 cameras (6 monochrome Pt. Grey Firefly MV USB cameras and 5 monochrome Basler A602f cameras) with wide-angle lenses and infrared pass, visible cut filters (R72, Hoya Filters) to view the interior volume of the flight chamber. The cameras were positioned so that a fly within the tracking volume was viewed by 2 or more cameras at any given time, enabling a 3D estimate of its position (Figure 1Bi). The cameras were first calibrated to compensate for image warping non-linearities (deviations from the pinhole model) and then the extrinsic and intrinsic parameters describing the pinhole model were found. Flies were tracked with an extended Kalman filter (EKF), in which the motion model was a linear constant velocity model, and fly maneuvering is captured by the stochastic component of the Kalman filter. Because tracking updates occurred at a high rate (60 fps) relative to fly maneuvering, we found this simplification to work well in practice. The 3D estimate of the fly position is recovered by triangulation from the 2D tracking data of each camera, and taking into account the relative uncertainty of each observations.
Saccade detection
Many species of flies, including Drosophila, exhibit rapid changes in heading as they fly, termed “saccades” [53]. Between saccades, flies tend to maintain an approximately straight course, and saccades account for at least 80% of the total net change in heading during flight [60]. There is little doubt that saccades can be triggered by visual stimuli, but the degree to which visual feedback plays a role in determining the velocity, duration, and amplitude of the resulting turn is unclear. Experiments using a magnetic tether, which permits free rotation about the yaw axis, suggest that flies do not respond to visual feedback during a saccade [61]. On the other hand, Stewart et al. [56] have observed a rebound effect after saccades in free flight, which they suggest is consistent with active optomotor feedback during the maneuver. This discrepancy is not of direct interest here, however, as we deal exclusively with the decision of initiating a saccade.
To analyze saccades within a flight trajectory, one should choose a detection algorithm that, given the trajectory data, returns a series of saccade events, possibly with other attributes such as direction, amplitude, velocity, etc. In the past, several detection algorithms have been proposed, each one implicitly using a slightly different definition of saccade, and each one able to compensate for different sources of noise. In practice, large saccades are such distinct events that all algorithms agree with respect to most classifications, but different algorithms may disagree on detection of small saccades. We make sure that our results are robust to the choice of the algorithm, by using two distinct algorithms based on different principles. The two algorithms are described in detail in Text S1 and their source code is available on line. Briefly, the Geometric Saccade Detector (GSD) detects saccades from the x-y planar trajectory. The Angular-Velocity based Saccade Detector (AVSD) works primarily by considering the smoothed angular heading rather than the planar position. Unless otherwise noted, the statistics shown through the paper are derived using GSD, which is a posteriori shown to be better suited for these particular experimental conditions and equipment. Alternative figures showing the same statistics obtained from the AVSD algorithm are available as part of Text S1.
Capturing behavior determinism and randomness using rate-variant Poisson processes
Figure 2A illustrates the conceptual approach of our analysis. We denote by the animal's physical spatial configuration (its position and velocity in a fixed reference frame). The stimulus is the set of all sensory cues perceived by the animal, and it is a function of both the spatial configuration and the appearance of the world . Whereas is a concrete variable that we can possibly measure, the stimulus and the world are placeholders for things that, in general, are unknown. The actions (e.g. saccades in our case) are the external manifestations of the internal neural processing, which depend both on the instantaneous stimulus as well as on , another placeholder variable that represents the animal's internal state (metabolic states, neural states, etc.), and which has dynamics of its own. We assume that it is possible to observe the spatial configuration as well as infer the actions from the observations, but that the internal state is not observable.
We make a distinction between obtaining a functional model of an animal's behavior and identifying the underlying neural processes. Obtaining a functional description of behavior means obtaining a model that can predict the actions given the spatial configuration and a description of the world . In principle, we can do this by observing an animal's behavior with enough samples of , and . In general, however, there are a variety of neural models that could produce the same functional model. For example, many behaviors appear to be well-localized in time, suggesting an “action potential” neural model, but the underlying neural model can have very different properties [62] (in other words, the microscopic explanation might be quite different than what the macroscopic observations suggest). The model that we now describe and that we will identify should be interpreted as a purely functional model, which can inform the search for neural models, to make sure that they are compatible with the externally observable free flight behavior.
Figure 2B shows the particular model that we use in this paper. It is a particular form of the general model discussed above (Figure 2A). In this model, we propose that the animal's actions can be summarized by the saccade events. We divide the saccade events in two classes: left and right saccades. In principle, one would want to consider additional attributes of the saccades, such as speed, duration, and amplitude. The analysis might also be expanded to consider other easily identifiable events [63]. However, limiting ourselves to a binary characterization of saccades allows us to model the behavior generation as Poisson processes, which offers relatively easy inference. We model saccade generation using rate-variant Poisson processes, i.e., we assume that, for each class of events, internal and external factors influence a time-varying event rate according to a quantitative relation that we will attempt to identify.
The most important assumption of our method (which can and will be verified a posteriori) is that, for the purpose of generating the behavior, the high-dimensional output can be compressed down to a low dimensional “feature” . This assumption is implicit in many other previous studies, and it is informed by the knowledge of the underlying neurobiology: the first level of sensory processing in flies and other animals consists in taking a very high-dimensional sensory stream and computing the few behaviorally-relevant features from it. Our only assumption is that this low-dimensional feature exists - we do not assume that we know this feature. However, we can attempt to automatically identify this feature from the observable data. It is important to note that we do not assume to know how this feature is computed from the stimulus. Indeed, the advantage of our method is that it allows identifying this feature based only on the observable behavior, without postulating anything on the sensory processing.
Figure 2B also shows explicitly that, in addition to the feature-dependent pathway in our model, other unmodeled processing influences the behavior. The effect of this unmodeled processing will be quantitatively estimated as well. The saccade events are assumed to be generated by a set of interacting Poisson process with variable rate . The index stands for either one of the two classes of events (L: left, R: right). The variable rate is assumed to depend both on the stimulus and the internal state , thus incorporating both random and deterministic effects. We write as the sum of three factors:
(1) |
where the term is the contribution of the external stimulus through the feature ; the term is the contribution of the internal state ; and the term represents the contribution of a purely random stochastic process that does not depend either on an internal state or the stimulus. By omitting some of the terms in the equation above, one can recover many other simpler models. For example, purely random behavior is obtained by setting .
The Poisson processes interact by inhibition. If any process generates an event, then any event generated from that process or any other process for a period of length Δ is ignored. This is meant to model a feature of many fixed action patterns that, once initiated, must run to completion before a different motor program can be initiated.
Finally, Figure 2B shows another variable , which we call “reduced configuration”. We define as the subset of the spatial configuration variables that actually influence the stimulus, for a particular class of environments . In general, for a freely flying animal, is at least a 12 dimensional quantity, including the 6 degrees of freedom for position/orientation and the corresponding 6 for velocities (additional degrees of freedom in the animal spatial configuration would be derived from the positions of body joints, such as the neck and wing positions). For particular environments, however, the stimulus is only dependent on a subset of . For example, if the environment is distant enough, then the visual stimulus does not depend on the forward velocity. Therefore, even though the spatial configuration is at least 12-dimensional, actually the stimulus depends on a smaller variable , i.e., the reduced configuration.
We will show that it is possible to identify all unknowns in this model. In particular, we will identify how the feature depends on the reduced configuration and how the rates depend on the feature. Remarkably, it is possible to do this without assumptions on how the feature is computed from the stimulus or how the stimulus depends on the reduced configuration . We only assume to be able to observe the reduced configuration and the generated saccade events. Before describing the method, we first discuss how this model based on rate-variant Poisson processes allows us to represent different functional models.
Predictions of different functional models
In Figure 3 we illustrate the predictions of four qualitatively different functional models in terms of the observed statistics. On the left side we show the functional model, and on the right we show the expected observed event rates as a function of the feature . This exercise assumes that we know how to estimate the feature, which we will show later. Here we describe what we would expect to find, before embarking on the actual computation of .
Figure 3A shows a “hard threshold” model, based on the computation of a single feature , which is then thresholded to obtain the event rate. A Poisson process then generates the events based on this time variant rate. The “stochastic trigger” in the figure masks the fact that there are two processes generating two classes of events, and that these processes are interacting (see discussion above), which is not relevant to the present discussion. If the absolute value of the feature is below a threshold, no event is generated; otherwise, saccades to the left and right are generated at a fixed rate. A large fixed rate would mean that the model is practically deterministic, with a large stimulus feature resulting in a behavioral event with only rare failures. On the right side of the figure, we show the observed event rates as a function of the feature . For this simple model, the observed rates as a function of the feature are straight steps. We remark that our analysis does not assume necessarily that the feature exhibits a hard threshold as in this simple model. We choose this shape merely because it allows visualizing the effect of different sources of noise.
In particular, we are interested in understanding the implications of a noise source that acts on the computation of the feature (sensory noise) compared to noise that generates behavior in a parallel process independent of the stimulus-computed feature (decision-making or motor noise).
Figure 3B shows the effect of measurement noise on the hard threshold model. Random fluctuations in the feature turn the hard threshold into a soft threshold.
Figure 3C shows the effect of adding a spontaneous generation process in parallel to the feature pathway. This has the effect of raising the predicted event rate by a constant value, as the parallel process is independent of the feature. A parallel generation process that depended on an unobservable internal state would have the same expected statistics if the internal state is uncorrelated with the feature. This means that a constant baseline event rate that is independent of the feature must be interpreted as the joint contribution of a purely stochastic spontaneous event generation together with a deterministic response based on internal states.
It is also important to consider the effect of another unmodeled feature on the event rate statistics, if we only model the dependence of one feature . This stems primarily from practical concerns, because the dimensionality of the feature that it is possible to identify depends primarily on the amount of data available. Therefore, once the dimensionality of the feature is fixed, we need a way to judge whether that dimension is sufficient to describe the behavior.
Figure 3D augments the model of Figure 3A with an additional pathway that uses a different feature . In such a case, if we plot the rates versus the feature, we will not find a clear functional dependency, indicating that the feature is no longer sufficient to explain the event rates. Conversely, if we find a clear functional dependency, then we can say that the feature is sufficient to capture the influence of the sensory stimulus on the behavior. This does not imply that is the only behaviorally relevant feature of the stimulus, because there could be other features that are relevant for other behaviors not considered in the analysis.
Our identification algorithm, described in the next section, recovers the best one-dimensional feature that explains the event rates. This permits constructing a function in which the experimental event rate is plotted against the feature curve. However, we anticipate that the experimental results, being dependent on experimental data, will have error bars both for dependent and independent variables. Strictly speaking, even if one finds a one-dimensional feature that uncovers a deterministic dependency between feature and rates compatible with the error bars, it is not possible to conclude that there is only one feature, because the effect of a second feature might be masked by the measurement noise. In this sense, our claims that one feature is sufficient is an application of parsimony.
In summary, we can identify the contributions of several qualitative factors by plotting the event generation rates as a function of . Measurement noise will soften the curve (e.g., a hard threshold is turned into a soft threshold). A parallel purely stochastic event generation process has the same effect of a deterministic process based on an internal state uncorrelated with the feature, namely it raises the curve by a fixed baseline rate independent of . If another unmodeled feature influences the behavior, there is not a strict functional dependence between the rates and the feature .
Identification of the feature
We devised a procedure that obtains an estimate of the best one-dimensional feature of the input that predicts the observed event rates. We explain here the basic idea, and provide details in Text S1. Intuitively, the feature and event rates can be obtained from the spatial statistics of the observed behavioral output. With respect to the discussion so far, the main conceptual step consists in translating the problem from the time to the space domain. So far, we have written the feature as a time-varying quantity. We have also assumed that depends on the stimulus , and that the stimulus depends on the animal spatial configuration , or more precisely, on the reduced configuration . Therefore, we rewrite our model writing instead of . The quantity is a spatial field that we interpret as the feature computed from the typical stimulus experienced at the reduced configuration . We will fit a model of the kind:
(2) |
where is the average event rate for the -th class (: left, R: right) observed at the reduced configuration ; denotes the event generation rates for left and right saccades as a function of the feature, and is constant term that we call baseline event rate.
Note the differences with respect to the previous model (Eq. 1). First, we have written the rates as a function of the reduced spatial configuration instead of time. Moreover, we do not model explicitly the contribution of the internal state. As argued above, given that we cannot measure the unobservable internal states , we cannot distinguish between a purely stochastic contribution and the contribution of an internal state
Therefore, the constant term will be an estimate of the joint contribution of the two terms that we cannot distinguish:
(3) |
where indicates the expected value taken over the whole trajectory.
We summarize here the three main phases for estimating from the behavioral data, while leaving the details to Text S1. First, the reduced configuration space is discretized into spatial cells with a resolution that depends on the amount of data available. For each of these cells, basic statistics are computed, such as the average time spent in each cell, as well as the observed event rates in the cell. One advantage of the algorithm is that these spatial statistics, averaged over the whole trajectory, are intrinsically robust to measurement noise and uncertainty in the event detection algorithm. Next, the event generation rates are computed from the observed rates. Because we assume that the Poisson processes interact with each other, and therefore the statistics of each process cannot be processed separately, and appropriate steps are required to take into account the interaction.
Once the average event generation rates are estimated, then we find the feature field that explains both event generation rates, in the sense that there exist two functions and such that the constraint described by equation (2) holds. Writing the constraint explicitly for each cell , and letting the value of the feature to estimate, we can see that we have a system of constraints of the kind:
The generated event rates and on the left side have already been estimated, while both the feature value and and have to be estimated. The constants and can be incorporated as part of and . Note that this can be interpreted as a dimensionality reduction problem, because we have to find one cause (the feature ) that explains two effects (left and right event rates) at the same time.
In our case, we solve a relatively simple instance of the problem in which is assumed to be a scalar function. Therefore, the constraints can be algebraically manipulated to obtain a closed form solution, which also takes into account the uncertainty in all the data and provide error bars for the estimated feature. The details are given in Text S1. Our approach is very generic, and can be extended to scenarios with more than 2 behaviors and more than 1 feature.
The feature should be considered a dimensionless quantity of arbitrary scale. In fact, the equations that define it have multiple solutions. For example, suppose that is one solution of the system of equations given by (Eq. 2). If is any invertible function, then one can verify that is a solution as well. Therefore, once we have obtained a solution for , we can rescale it using any function that we find convenient. In the following, we choose the rescaling function such that is uniformly distributed in the interval .
Results
Event statistics and estimated feature
We tracked 88 flies for a total of 5130 seconds or approximately 1.4 hours. Of the total recorded time, we considered only the 4814 seconds of data in which the flight speed exceeded 5 centimeters per second. This threshold on the linear velocity allowed working on tracks for which saccades were easier to detect. We detected a total of 6613 saccades with this criterion, giving an average saccade rate of 1.37 saccades per second.
We chose a reduced configuration that is two-dimensional. This follows from considering only planar motion (which reduces the effective degrees of freedom to 3), and using the symmetry of the circular arena (which reduces the degrees of freedom to 2). An implied assumption (which can be verified a posteriori) is that the fly's response is not dependent on the variables not considered in the analysis; for example, even though it is known that flies [64] and other insects [65] use gaze to stabilize vision, there is no gaze variable in our model. This is because the resolution of our measurements is not enough to observe directly the relative pose of head and body, in terms of pitch, roll, or yaw. All components of the spatial configuration that are theoretically relevant for the stimulus, but cannot be measured, are “hidden” states whose contribution is lumped into the constant term in (2).
The two-dimensional reduced configuration can be parameterized in different ways, the results being independent of the particular parameterization. The primary parameterization that we use for computation uses for coordinates: is the distance to the wall and is the angle that the fly heading forms with respect to the axis of the arena (Figure 1Ci ). We chose this parameterization because it corresponds to two behaviorally relevant variables. We preferred the axis angle over other potentially valid representations for the heading (e.g., approach angle) because the representation is not singular, as for any value of .
We compute all statistics in the space, but we also use another choice of coordinates to visualize the same data. We rotate the original configuration of the fly around the center of the arena, such that the new coordinates are . These “fly-centric” coordinates are displayed using a top-down view of the arena, in which the fly always points up (Figure 1Cii ).
The reduced configuration was discretized in a grid with sides of 36 cells (for ) and 20 cells (for d) (Figure 1Di ). The angle was discretized in 36 cells of equal size 10 deg. The distance was discretized in 20 unequal intervals (note the unequal axis in Figure 1Di ). Intervals for are smaller at the center of the arena and larger near the border, in such a way that each annulus of radius and width had the same area. To compensate for the sparseness of the data, each cell extends 50% into the neighbor's area. Although these choices were somewhat arbitrary, we obtain qualitatively similar results if we vary the number of the cells.
Figure 1D shows the distribution of time spent at each point of the arena, and Figure 1E shows the distribution of the detected saccades using the GSD algorithm (see Text S1 for figures using the saccades detected by the alternative AVSD algorithm). As clearly evident in Figure 1Eii , most of the detected saccades correspond to the fly avoiding the walls on the left or on the right. However, those are the configurations where the flies spent more time (Figure 1D). Therefore, we need to normalize this data to see the behavioral patterns.
Figure 4A shows the estimated saccade generation function across the reduced configuration space. These rates are obtained by first computing the observed generation rates by averaging the number of saccades (Figure 1E) by the time spent in each cell (Figure 1D). Then the rates are obtained from by correcting for an estimated inhibition interval s. Panels B and C show the data separately for left and right saccades ( and ). The most evident phenomenon is that the fly tends to turn left when the wall is on the right (and vice versa), however, there are many saccades of the opposite direction initiated, even when the turning would orient the fly towards the wall rather than away from it. This is the phenomenon that we want the feature to explain: we want to find the best spatial scalar value such that both and can be written as a function of . Figure 5Ai-ii shows the estimated feature as a function of the reduced configuration c. This is the unidimensional feature that best explains both the left and saccade rates. The estimated feature using the alternative saccade detector is qualitatively similar (Figure S1). We now have the spatial feature as well as the rates as a function of the reduced configuration and can now plot as a function of (using as an implicit variable). This is shown in Figure 5B, which shows, for each cell , the value of as a function of . Figure 5Bi shows the data as a scatter plot, while Figure 5Bii shows the error bars on the estimated rates at the 95% significance level.
Predictive power of the estimated feature
The data in Figure 5B indicate the predictive power of the feature. If the feature was perfectly predictive of the event rates, then would be a function of . In this case, taking into account the error bounds on the rates, it is possible to find two functions that predict the event rates in approximately 93% of the environment. More specifically, given a generic cell corresponding to the spatial configuration we find that the predicted event rates and are compatible with the observed rates and at the 95% level of significance. In practice, this means that the data in Figure 5B can be explained by two smooth curves ( and ) that intersect 93% of the confidence intervals corresponding to each spatial configuration. In the remaining cells (51 of 720 cells), the rates cannot be predicted by this feature alone. Further inspection (data not shown) reveals that such points correspond to configurations with the fly pointing directly against the wall at a small distance (<0.3 m). Note that being able to predict the rates from the feature does not mean that one is able to predict the direction of each single saccade event. For example, in the middle of the arena the probability of left and right saccade is 50%, and this percentage is perfectly predicted by the feature; however, it is impossible to predict the direction of the single saccade better than chance.
Bounds on the contributions of random process and internal states
As explained before, using only external observations of the animal spatial configuration, we cannot distinguish among the contribution of a purely random endogenous saccade generation process, a deterministic process based on an internal state, any unmodeled features computed from the stimulus, and any unobservable spatial configuration that we cannot observe due to the limited resolution of our instruments. These contributions are lumped together in a baseline saccade rate. By examining the curves in Figure 5Bi we can estimate a baseline event rate of about saccades/sec. By comparing with a maximum estimated event rate of saccades/sec, we can estimate that roughly 90% of the saccades are stimulus-driven in the regions of maximum stimulus. This value depends on the geometry and texture of this particular arena (e.g., it would be different if the arena was larger or smaller). However, we predict that the baseline rate of saccades/sec that we measure at the center of the arena should be independent of the geometry, as the size and textures of this arena were chosen such that the fly cannot perceive significant visual contrast from the center.
We can make some informed guesses for the contribution of the various possible processes by considering circumstantial evidence from other experiments. In tethered flight experiments, deliberately performed in the absence of salient visual stimuli, spontaneous saccade rates are on the order of saccades/sec [66]. If we assume that these values obtained in tethered experiments are a good approximation of an assumed spontaneous generation process in free-flight, then we can account for approximately 75% of the unexplained 0.4 saccades/sec as the joint contribution of a random process and unobservable internal states. This leaves roughly 25% of unexplained data, which could possibly be explained by estimating an additional feature , perhaps dependent on components of the spatial configuration that we cannot observe, such as the gaze direction. The contribution of a hypothetical feature is therefore very small with respect to the contribution of the estimated , as could possibly explain about 0.1 saccades/sec versus the 4 saccades/sec explained by .
We conclude that the saccade behavior of Drosophila that depends on external visual stimulus appears to depend for the most part on only a one-dimensional feature of the stimulus. These conclusions must be limited to the particular experimental condition, as we cannot exclude that more complex environments would elicit more complex responses that require a higher dimensional feature to be explained. However, even in our relatively simple flight environment, our analysis implies that the vast majority of saccades we observed are stimulus-driven and are not due to an internal, stimulus-independent search algorithm (e.g. Levy flights), as has been suggested for Drosophila and many other species [6]–[18].
Approximating the feature field using known parametric structures for visual processing
We have been able to compute the feature field from the observable fly trajectory, without any assumptions on the fly visual processing. Nevertheless, it is interesting to test whether this independently identified feature is compatible with existing models of the first stages of visual processing in flies. In particular, we test the hypothesis of whether the identified feature can be expressed as a linear function of the perceived optic flow.
We assume the following generative model for :
(4) |
where is the optic flow, or angular velocity, at the retinal angle at time , and is a retinal input kernel. The value corresponds to the animal's center front visual field.
The function is an arbitrary nonlinear function that we include in the model, because the identification procedure allows us to know only up to a monotone transformation (i.e., if is a solution of the constraints system, then also is a valid solution). We can characterize the optimal as the solution of an optimization problem:
(5) |
where the error function is given by:
(6) |
In this last expression, is the typical optic flow that the animal experiences at the reduced configuration . By solving this optimization problem, we try to best approximate the estimated feature over the whole environment, assuming it can be expressed as a linear function of the optic flow.
Unfortunately, we found that this optimization problem is ill posed given our data. In particular, is known only at a discrete set of values (720 cells — the density of these is constrained by the finite amount of data that we have), and it is quite noisy, whereas the unknowns are of high dimension. Given that the resolution of the fly's visual system is around deg, it makes sense to use at least 70 numbers (∼330/5) for representing . Furthermore, can be any monotonic nonlinear function.
We tried to improve the results by penalizing large values and large spatial variations of (measured either by the spatial derivatives or ). The modified error function is:
(7) |
for and different values of . In general, by varying and , we found a multitude of solutions, all very different from each other, having approximately the same predictive power (Figure 6). We noticed that for increasing regularization values the estimated linear kernel tended to be shaped as an harmonic function, as illustrated by the kernel obtained by regularizing the second derivative () and using a large value of (), shown in Figure 6D. This kernel is still asymmetric. If we impose that the kernel must be symmetric, we find that the best approximation using one harmonic is:
(8) |
This kernel and relative feature field is shown in Figure 6E, and it is a good approximation of the feature estimated from the data.
We conclude that the identified feature can be expressed as a simple function of the optic flow. However, while obtaining the behaviorally relevant feature from the external observations alone is a well-posed mathematical problem, finding the function that maps the stimulus to the feature is an ill-posed problem, because the set of possible models is of very high dimension compared with the data that we have. Note that these issues are already evident when considering only linear functions of the optic flow, and would be even more pressing if we were to add other nonlinear components to the model that are known to exist in the neural circuits of the fly.
Most of the estimated kernels obtained using some form of regularization share a particular feature: is never 0 for corresponding to the back of the animal, but has opposite sign in the front of the animal (Figure 6B,C,D,E). Further investigation shows that these non-zero values in the caudal region are responsible for the two small side lobes that appear in the feature field when plotted in fly-centric coordinates. If the kernel is set to zero in the back, these side lobes disappear. This is apparent by comparing the feature field in Figure 6F (corresponding to the kernel ) with that in Figure 6E. These results suggest that the optic flow in the back of the animal influences the fly's turning decisions. This response cannot be interpreted as pure obstacle avoidance, given that flies tend to fly forward and obstacles in the back are not expected to represent a threat for collision. For convex environments, the saccades initiated from this response would tend to align the fly's course in parallel to the environment boundaries and the overall result is to follow walls rather than completely avoid them (similar behavior has been observed in bees [67]). Such a visuo-motor system might provide a functional advantage with respect to the balance of collision avoidance and object search. An animal that balances attraction and obstacle avoidance would tend to remain relatively close to interesting visual features, whereas an animal whose primary reflex is to fly away from visual features would tend to find itself in large open areas, far from potential landmarks or food sources. The only way of quantitatively verifying this attraction-deflection hypothesis would be to obtain data from experiments within larger environments with more varied visual features. These results are also compatible with the observation in previous experiments on tethered flies that the optomotor response can be written as the function of a kernel in which the rear and front visual fields give opposite contributions [68], suggesting that a similar visual feature might be used for both behaviors.
Discussion
In this paper, we introduced a novel method to obtain an estimate of a low-dimensional feature of the stimulus that best predicts the observable behavioral event generation rates. The feature can be obtained from observable quantities, such as the recording of the trajectory of the animal, without any assumption on the nature of the stimulus and its underlying neural processing. Using this method, we have concluded that most of the saccade events generated by fruit flies exploring a structured laboratory environment are induced by visual stimuli, and that the instantaneous stimulus can be compressed down to a one-dimensional feature, while still being predictive of the event rates in % of the environment. Using this method, it is not possible to distinguish between the contribution of an endogenous random process and a deterministic contribution dependent on an unobservable state. However, we can bound the contributions of these two terms in a baseline saccade rate that we estimate at 0.4 saccades/sec, roughly a tenth of the maximum rate. The strength of this method is that the feature can be estimated working backwards from an animal's actions, rather than forward by postulating a model for the stimulus and guessing what is the relevant feature. Once we know , as a second step, it is possible to attempt to fit a parametric representation of neural processing to find the forward function from to , based on other assumptions about sensory processing, though this is not guaranteed to be a well posed problem, as one must optimize over all plausible models compatible with the animal's biology. In this particular case, we have shown that the feature responsible for turning decisions in Drosophila can be written as a linear function of the optic flow, and that the particular linear kernel we obtain is compatible with that identified in tethered conditions, for a particular choice of regularization penalty to make the problem well posed. Conversely, finding from the behavioral data is a well posed and intuitive problem, because it can be understood as a dimensionality reduction problem (find the one feature that explains multiple behaviors).
The main advantage of this approach, compared with previous methods, is that it can be applied to freely moving animals, and thus permits asking about responses to naturally important stimuli. Moreover, it does not need any assumption of linearity between some aspect of the stimulus and response, a precondition strongly needed in techniques such as reverse correlation [48]. Even advanced reverse correlation techniques in single sensory neurons [69] are not easy to generalize into models of network functionality that could be used to predict behavior.
In the future, this method could be applied to different behaviors of the fruit fly and other animals [70]. The formalization is quite generic, though some generalizations are possible. The algorithm documented in Text S1 assumes that the feature is one-dimensional in order to obtain a closed-form solution. To identify a feature of higher dimension, this must be generalized, for example by using one of the various more computationally expensive dimensionality reduction algorithms in machine learning (e.g., [71]). In any case, the rate-variant interacting Poisson process model seems apt for modeling many other behaviors (e.g., landing, taking off) that can be reliably localized in time (i.e., they have a clear beginning and end), and that can be caused by both external and internal causes.
Thinking in terms of the feature as a proxy of the stimulus can potentially be useful in understanding how different sensory modalities contribute to the same behavior. The feature is independent of the sensory modality because it is just a function of the animal configuration, and it is a proxy of the typical stimulus perceived at the location, so it could be used to study, for example, the influence of olfaction instead of visual processing on turning behavior, or their interaction, which has been the object of much research [72]–[75] [56]. Note, however, that we do have the strong assumption that the stimulus is a constant function of the configuration, so the framework cannot be easily extended to time-varying stimuli.
This approach might also be useful to study different behaviors at the same time. Drosophila has a large repertoire of behaviors/reflexes which are stimulus-triggered, such as landing, take-off, chasing mates, and escaping from small targets. In this case, we focused on saccades, and we found the feature encoding the relevant function of the stimulus for saccade decisions. If one repeated the analysis for a different behavior (e.g., landing), there would likely be another feature , that would be different from . However, if this was repeated for all fly behaviors, one would find that at some point the new identified features would be redundant; for example, in the case of vision, the number of features is upper bounded by the number of upstream signals towards the lobula. Ultimately, this exercise might provide a prediction of whether two behaviors are likely to share the same neural pathways.
Potentially, this technique could help in quantifying the behavioral differences of different genotypes. This model makes a distinction between the feature and the event generation rate functions . Whereas z is assumed to be correlated with computed percepts, might be correlated more with the motor functions. This distinction could be used to obtain insight regarding the function of genetic manipulations such as a screen in which populations of neurons are “silenced” with a hyperpolarizing ion channel or synaptic release blockade. For example, if a modified animal gives the same feature but modified rate functions , it would be evidence that the silenced neurons are involved with motor generation rather than with stimulus processing. Consequently, with a large-scale screen [76], [77], it might be possible to obtain a classification of phenotypes into sensory, decision making, and motor deficits. Similarly, we could use this feature to quantitatively compare the properties of different species.
Another interesting but more substantial extension of this work would be to expand the mathematical formalism to incorporate measurements of neuronal activity into the internal processing structure. This is now done in freely moving worms [78], [79], [32] and zebrafish [80]; in adult flies, most neural recording during behavior is being done on fixed flies [81]–[83], [42].
Supporting Information
Acknowledgments
We thank Kit Longden for insightful comments on a preliminary version of this work.
Funding Statement
This study was funded by AFOSR (FA9550-06-1-0079 and FA9550-10-1-0086), ARO (DAAD 19-03-D-0004), and NSF (NSF-0931746), DARPA (MSEE). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1. Berg HC, Brown DA (1972) Chemotaxis in Escherichia coli analysed by Three-dimensional Tracking. Nature 239: 500–504. [DOI] [PubMed] [Google Scholar]
- 2. Tynan CT (1998) Ecological importance of the Southern Boundary of the Antarctic Circumpolar Current. Nature 392: 708–710 doi:10.1038/33675.. [Google Scholar]
- 3. Fuortes MG, Yeandle S (1964) Probability of occurrence of discrete potential waves in the eye of Limulus. The Journal of general physiology 47: 443–63. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Baylor DA, Lamb TD, Yau KW (1979) Responses of retinal rods to single photons. The Journal of physiology 288: 613–34. [PMC free article] [PubMed] [Google Scholar]
- 5. Menini A, Picco C, Firestein S (1995) Quantal-like current fluctuations induced by odorants in olfactory receptor cells. Nature 373: 435–7 doi:10.1038/373435a0.. [DOI] [PubMed] [Google Scholar]
- 6. Shimada I, Kawazoe Y, Hara H (1993) A temporal model of animal behavior based on a fractality in the feeding of Drosophila melanogaster. Biological Cybernetics 68: 477–481. [DOI] [PubMed] [Google Scholar]
- 7. Cole B, Cole J (1995) Fractal time in animal behaviour: the movement activity of Drosophila. Animal Behaviour 50: 1317–1324 doi:10.1016/0003-3472(95)80047-6.. [Google Scholar]
- 8. Bartumeus F, Levin SA (2008) Fractal reorientation clocks: Linking animal behavior to statistical patterns of search. Proceedings of the National Academy of Sciences 105: 19072. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Viswanathan GM, Afanasyev V, Buldyrev SV, Murphy EJ, Prince PA, et al. (1996) Lévy flight search patterns of wandering albatrosses. Nature 381: 413–415 doi:10.1038/381413a0.. [DOI] [PubMed] [Google Scholar]
- 10. Bartumeus F, Peters F, Pueyo S, Marrasé C, Catalan J (2003) Helical L{é}vy walks: adjusting searching statistics to resource availability in microzooplankton. Proceedings of the National Academy of Sciences 100: 12771. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Humphries NE, Queiroz N, Dyer JRM, Pade NG, Musyl MK, et al. (2010) Environmental context explains Lévy and Brownian movement patterns of marine predators. Nature 465: 1066–1069 doi:10.1038/nature09116.. [DOI] [PubMed] [Google Scholar]
- 12. Maye A, Hsieh C-H, Sugihara G, Brembs B (2007) Order in spontaneous behavior. PloS one 2: e443 doi:10.1371/journal.pone.0000443.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Reynolds AM, Frye Ma (2007) Free-flight odor tracking in Drosophila is consistent with an optimal intermittent scale-free search. PloS one 2: e354 doi:10.1371/journal.pone.0000354.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Reynolds AM, Smith AD, Reynolds DR, Carreck NL, Osborne JL (2007) Honeybees perform optimal scale-free searching flights when attempting to locate a food source. Journal of Experimental Biology 4: 3763–3770 doi:10.1242/jeb.009563.. [DOI] [PubMed] [Google Scholar]
- 15. Osborne JL, Andrew P, Carreck NL, Swain JL, Knight ME, et al. (2008) Bumblebee flight distances in relation to the forage landscape. Journal of Animal Ecology 77: 406–415 doi:10.1111/j.1365-2656.2007.01333.x.. [DOI] [PubMed] [Google Scholar]
- 16. Nevitt Ga, Losekoot M, Weimerskirch H (2008) Evidence for olfactory search in wandering albatross, Diomedea exulans. Proceedings of the National Academy of Sciences of the United States of America 105: 4576–4581 doi:10.1073/pnas.0709047105.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Sims DW, Southall EJ, Humphries NE, Hays GC, Bradshaw CJA, et al. (2008) Scaling laws of marine predator search behaviour. Nature 451: 1098–1102 doi:10.1038/nature06518.. [DOI] [PubMed] [Google Scholar]
- 18. Nathan R (2008) An emerging movement ecology paradigm. Proceedings of the National Academy of Sciences 105: 19050. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Reynolds A (2008) How many animals really do the Lévy walk? Comment. Ecology 89: 2347–2351. [DOI] [PubMed] [Google Scholar]
- 20. Stumpf MPH, Porter MA (2012) Critical Truths About Power Laws. Science 335: 665–666 doi:10.1126/science.1216142.. [DOI] [PubMed] [Google Scholar]
- 21. Edwards AM, Phillips Ra, Watkins NW, Freeman MP, Murphy EJ, et al. (2007) Revisiting Lévy flight search patterns of wandering albatrosses, bumblebees and deer. Nature 449: 1044–8 doi:10.1038/nature06199.. [DOI] [PubMed] [Google Scholar]
- 22. Bartumeus F (2007) Lévy processes in animal movement: an evolutionary hypothesis. Fractals 15: 151. [Google Scholar]
- 23.Goldreich O (2001) Foundations of Cryptography: Basic Tools. Cambridge University Press.
- 24. Hassenstein B, Reichardt W (1956) Systemtheoretische Analyse der Zeit-, Reihenfolgen- und Vorzeichenauswertung bei der Bewegungsperzeption des Rüsselkafers \emph{Chlorophanus}. Zeitschrift Für Naturforschung 11b: 513–524. [Google Scholar]
- 25. Srinivasan MV, Zhang SW, Lehrer M, Collett TS (1996) Honeybee navigation en route to the goal: Visual flight control and odometry. Journal of Experimental Biology 199: 237–244. [DOI] [PubMed] [Google Scholar]
- 26. Reynolds AM, Frye MA (2007) Free-flight odor tracking in Drosophila is consistent with an optimal intermittent scale-free search. PloS one 2: e354 doi:10.1371/journal.pone.0000354.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Maye A, Hsieh C-H, Sugihara G, Brembs B (2007) Order in spontaneous behavior. PloS one 2: e443 doi:10.1371/journal.pone.0000443. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Srivastava N, Clark DA, Samuel ADT (2009) Temporal Analysis of Stochastic Turning Behavior of Swimming C. elegans . Journal of Neurophysiology 1172–1179 doi:10.1152/jn.90952.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Edwards AM, Phillips RA, Watkins NW, Freeman MP, Murphy EJ, et al. (2007) Revisiting Lévy flight search patterns of wandering albatrosses, bumblebees and deer. Nature 449: 1044–8 doi:10.1038/nature06199.. [DOI] [PubMed] [Google Scholar]
- 30. Nevitt Ga, Losekoot M, Weimerskirch H (2008) Evidence for olfactory search in wandering albatross, Diomedea exulans. Proceedings of the National Academy of Sciences of the United States of America 105: 4576–81 doi:10.1073/pnas.0709047105.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Berg HC, Brown DA (1972) Chemotaxis in Escherichia coli analysed by Three-dimensional Tracking. Nature 239: 500–504. [DOI] [PubMed] [Google Scholar]
- 32. Piggott BJ, Liu J, Feng Z, Wescott SA, Xu XZS (2011) The Neural Circuits and Synaptic Mechanisms Underlying Motor Initiation in C. elegans. Cell 147: 922–933 doi:10.1016/j.cell.2011.08.053.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Kawano T, Po MD, Gao S, Leung G, Ryu WS, et al. (2011) An Imbalancing Act: Gap Junctions Reduce the Backward Motor Circuit Activity to Bias C. elegans for Forward Locomotion. Neuron 72: 572–86 doi:10.1016/j.neuron.2011.09.005.. [DOI] [PubMed] [Google Scholar]
- 34. Cowan NJ, Fortune ES (2007) The critical role of locomotion mechanics in decoding sensory systems. The Journal of neuroscience : the official journal of the Society for Neuroscience 27: 1123–8 doi:10.1523/JNEUROSCI.4198-06.2007.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Beyeler A, Zufferey J-C, Floreano D (2009) Vision-based control of near-obstacle flight. Autonomous Robots 27: 201–219 doi:10.1007/s10514-009-9139-6.. [Google Scholar]
- 36. Conroy J, Gremillion G, Ranganathan B, Humbert JS (2009) Implementation of wide-field integration of optic flow for autonomous quadrotor navigation. Autonomous Robots 1–10. [Google Scholar]
- 37. Borst A, Euler T (2011) Seeing things in motion: models, circuits, and mechanisms. Neuron 71: 974–94 doi:10.1016/j.neuron.2011.08.031. [DOI] [PubMed] [Google Scholar]
- 38. Borst A (2009) Drosophila's view on insect vision. Current biology 19: R36–47 doi:10.1016/j.cub.2008.11.001.. [DOI] [PubMed] [Google Scholar]
- 39. Kennedy JS (1940) The visual responses of flying mosquitoes. Proceedings of the Zoological Society of London Series A 109: 221–242 doi:10.1111/j.1096-3642.1940.tb00831.x.. [Google Scholar]
- 40. Mittelstaedt H (1949) Telotaxis und Optomotorik von Eristalis bei Augeninversion. Naturwissenschaften 36: 90–91. [Google Scholar]
- 41. Eichner H, Joesch M, Schnell B, Reiff DF, Borst A (2011) Internal structure of the fly elementary motion detector. Neuron 70: 1155–64 doi:10.1016/j.neuron.2011.03.028.. [DOI] [PubMed] [Google Scholar]
- 42. Clark DA, Bursztyn L, Horowitz MA, Schnitzer MJ, Clandinin TR (2011) Defining the computational structure of the motion detector in Drosophila. Neuron 70: 1165–77 doi:10.1016/j.neuron.2011.05.023.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Reiff DF, Plett J, Mank M, Griesbeck O, Borst A (2010) Visualizing retinotopic half-wave rectified input to the motion detection circuitry of Drosophila. Nature neuroscience 13: 973–978 doi:10.1038/nn.2595.. [DOI] [PubMed] [Google Scholar]
- 44. Joesch M, Schnell B, Raghu SV, Reiff DF, Borst A (2010) ON and OFF pathways in Drosophila motion vision. Nature 468: 300–304 doi:10.1038/nature09545.. [DOI] [PubMed] [Google Scholar]
- 45. Krapp HG, Hengstenberg R (1996) Estimation of self-motion by optic flow processing in single visual interneurons. Nature 384: 463–6 doi:10.1038/384463a0.. [DOI] [PubMed] [Google Scholar]
- 46.Roth E, Reiser MB, Dickinson MH, Cowan NJ (2012) A Task-level Model for Optomotor Yaw Regulation in Drosophila Melanogaster: A Frequency-Domain System Identification Approach. In: Proceedings of the Conference on Decisions and Controls (CDC), Maui, Hawaii.
- 47. Aptekar JW, Shoemaker PA, Frye MA (2012) Figure Tracking by Flies Is Supported by Parallel Visual Streams. Current biology 22: 482–487 doi:10.1016/j.cub.2012.01.044.. [DOI] [PubMed] [Google Scholar]
- 48. Theobald JC, Ringach DL, Frye MA (2010) Dynamics of optomotor responses in Drosophila to perturbations in optic flow. The Journal of experimental biology 213: 1366–75 doi:10.1242/jeb.037945.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Mronz M, Lehmann F-O (2008) The free-flight response of Drosophila to motion of the visual environment. The Journal of experimental biology 211: 2026–45 doi:10.1242/jeb.008268.. [DOI] [PubMed] [Google Scholar]
- 50. Braun E, Geurten B, Egelhaaf M (2010) Identifying prototypical components in behaviour using clustering algorithms. PloS one 5: e9361 doi:10.1371/journal.pone.0009361.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Lindemann JP, Weiss H, Möller R, Egelhaaf M (2008) Saccadic flight strategy facilitates collision avoidance: closed-loop performance of a cyberfly. Biological Cybernetics 98: 213–227 doi:10.1007/s00422-007-0205-x.. [DOI] [PubMed] [Google Scholar]
- 52. Boeddeker N, Egelhaaf M (2005) A single control system for smooth and saccade-like pursuit in blowflies. J Exp Biol 208: 1563–1572 doi:10.1242/jeb.01558.. [DOI] [PubMed] [Google Scholar]
- 53. Land MF, Collett TS (1974) Chasing behavior of houseflies (Fannia canicularis): a description and analysis. J Comp Physiol 89: 331–357. [Google Scholar]
- 54. Wagner H (1986) Flight performance and visual control of the flight of the free-flying housefly (Musca domestica). II. Pursuit of targets. Philos Trans R Soc Long B Biol Sci 312: 553–579. [Google Scholar]
- 55. Tammero LF, Dickinson MH (2002) The influence of visual landscape on the free flight behavior of the fruit fly Drosophila melanogaster. Journal of Experimental Biology 205: 327–343. [DOI] [PubMed] [Google Scholar]
- 56. Stewart FJ, Baker DA, Webb B (2010) A model of visual-olfactory integration for odour localisation in free-flying fruit flies. Journal of Experimental Biology 213: 1886–1900 doi:10.1242/jeb.026526.. [DOI] [PubMed] [Google Scholar]
- 57. Stephens GJ, Osborne LC, Bialek W (2010) Searching for simplicity: Approaches to the analysis of neurons and behavior. Proc Natl Acad Sci U S A 108: 15565–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Heisenberg M, Wolf R (1984) Vision in Drosophila: genetics of Microbehavior. Springer-Verlag.
- 59. Straw AD, Branson K, Neumann TR, Dickinson MH (2011) Multi-camera Realtime 3D Tracking of Multiple Flying Animals. J R Soc Interface 8: 395–409 doi:10.1098/rsif.2010.0230.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Breugel F, van Dickinson M (2012) The visual control of landing and obstacle avoidance in the fruit fly, Drosophila. Journal of Experimental Biology 215: 1783–98. [DOI] [PubMed] [Google Scholar]
- 61. Bender JA, Dickinson MH (2006) A comparison of visual and haltere-mediated feedback in the control of body saccades in Drosophila melanogaster. Journal of Experimental Biology 209: 4597–4606 doi:10.1242/jeb.02583.. [DOI] [PubMed] [Google Scholar]
- 62. Stephens GJ, Bueno de Mesquita M, Ryu WS, Bialek W (2011) Emergence of long timescales and stereotyped behaviors in Caenorhabditis elegans. Proceedings of the National Academy of Sciences of the United States of America 108: 7286–9 doi:10.1073/pnas.1007868108.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63. Geurten BRH, Kern R, Braun E, Egelhaaf M (2010) A syntax of hoverfly flight prototypes. Journal of Experimental Biology 213: 2461–2475 doi:10.1242/jeb.036079.. [DOI] [PubMed] [Google Scholar]
- 64. Rosner R, Egelhaaf M, Grewe J, Warzecha AK (2009) Variability of blowfly head optomotor responses. Journal of Experimental Biology 212: 1170–1184 doi:10.1242/jeb.027060.. [DOI] [PubMed] [Google Scholar]
- 65. Boeddeker N, Dittmar L, Stürzl W, Egelhaaf M (2010) The fine structure of honeybee head and body yaw movements in a homing task. Proceedings Biological sciences/The Royal Society 277: 1899–906 doi:10.1098/rspb.2009.2326.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66. Bender JA, Dickinson MH (2006) Visual stimulation of saccades in magnetically tethered Drosophila. The Journal of experimental biology 209: 3170–82 doi:10.1242/jeb.02369.. [DOI] [PubMed] [Google Scholar]
- 67. Serres JR, Masson GP, Ruffier F, Franceschini N (2008) A bee in the corridor: centering and wall-following. Naturwissenschaften 95: 1181–1187 doi:10.1007/s00114-008-0440-6.. [DOI] [PubMed] [Google Scholar]
- 68. Tammero LF, Frye MA, Dickinson MH (2004) Spatial organization of visuomotor reflexes in Drosophila. J Exp Biol 207: 113–122 doi:10.1242/jeb.00724.. [DOI] [PubMed] [Google Scholar]
- 69. Pillow JW, Simoncelli EP (2006) Dimensionality reduction in neural models: an information-theoretic generalization of spike-triggered average and covariance analysis. Journal of vision 6: 414–28 doi:10.1167/6.4.9.. [DOI] [PubMed] [Google Scholar]
- 70. Gomez-Marin A, Stephens GJ, Louis M (2011) Active sampling and decision making in Drosophila chemotaxis. Nature communications 2: 441 doi:10.1038/ncomms1455.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71. Tenenbaum JB, Silva VD, Langford JC (2000) A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science 290: 2319–2323. [DOI] [PubMed] [Google Scholar]
- 72. Chow DM, Frye MA (2009) The neuro-ecology of resource localization in Drosophila: Behavioral components of perception and search. Fly 3: 50–61 doi:10.4161/fly.3.1.7775.. [DOI] [PubMed] [Google Scholar]
- 73. Krishnan P, Duistermars BJ, Frye MA (2011) Odor identity influences tracking of temporally patterned plumes in Drosophila. BMC Neuroscience 12: 62 doi:10.1186/1471-2202-12-62.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74. Frye M, Duistermars B (2009) Visually Mediated Odor Tracking During Flight in Drosophila. Journal of Visualized Experiments Jan 26 1110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. Budick SA, Dickinson MH (2006) Free-flight responses of Drosophila melanogaster to attractive odors. Journal of Experimental Biology 209: 3001–3017 doi:10.1242/jeb.02305.. [DOI] [PubMed] [Google Scholar]
- 76. Iyengar BG, Chou CJ, Vandamme KM, Klose MK, Zhao X, et al. (2011) Silencing synaptic communication between random interneurons during Drosophila larval locomotion. Genes, brain, and behavior 10: 883–900 doi:10.1111/j.1601-183X.2011.00729.x. [DOI] [PubMed] [Google Scholar]
- 77. Katsov AY, Clandinin TR (2008) Motion processing streams in Drosophila are behaviorally specialized. Neuron 59: 322–35 doi:10.1016/j.neuron.2008.05.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 78. Stirman JN, Crane MM, Husson SJ, Wabnig S, Schultheis C, et al. (2011) Real-time multimodal optical control of neurons and muscles in freely behaving Caenorhabditis elegans. Nature methods 8: 153–8 doi:10.1038/nmeth.1555. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79. Leifer AM, Fang-Yen C, Gershow M, Alkema MJ, Samuel ADT (2011) Optogenetic manipulation of neural activity in freely moving Caenorhabditis elegans. Nature Methods 8: 147–52 doi:10.1038/nmeth.1554.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80. Naumann EA, Kampff AR, Prober DA, Schier AF, Engert F (2010) Monitoring neural activity with bioluminescence during natural behavior. Nature neuroscience 13: 513–20 doi:10.1038/nn.2518.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81. Maimon G, Straw AD, Dickinson MH (2010) Active flight increases the gain of visual motion processing in Drosophila. Nat Neurosci 13: 393–399 doi:10.1038/nn.2492.. [DOI] [PubMed] [Google Scholar]
- 82. Seelig JD, Chiappe ME, Lott GK, Dutta A, Osborne JE, et al. (2010) Two-photon calcium imaging from head-fixed Drosophila during optomotor walking behavior. Nature methods 7: 535–40 doi:10.1038/nmeth.1468.. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83. Chiappe ME, Seelig JD, Reiser MB, Jayaraman V (2010) Walking Modulates Speed Sensitivity in Drosophila Motion Vision. Current Biology 20: 1470–1475 doi:10.1016/j.cub.2010.06.072.. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.