Abstract
An innovative insect-based visual sensor is designed to perform active marker tracking. Without any optics and a field-of-view of about 60°, a novel miniature visual sensor is able to locate flickering markers (LEDs) with an accuracy much greater than the one dictated by the pixel pitch. With a size of only 1 cm3 and a mass of only 0.33 g, the lensless sensor, called HyperCube, is dedicated to 3D motion tracking and fits perfectly with the drastic constraints imposed by micro-aerial vehicles. Only three photosensors are placed on each side of the cubic configuration of the sensing device, making this sensor very inexpensive and light. HyperCube provides the azimuth and elevation of infrared LEDs flickering at a high frequency (>1 kHz) with a precision of 0.5°. The minimalistic design in terms of small size, low mass and low power consumption of this visual sensor makes it suitable for many applications in the field of the cooperative flight of unmanned aerial vehicles and, more generally, robotic applications requiring active beacons. Experimental results show that HyperCube provides useful angular measurements that can be used to estimate the relative position between the sensor and the flickering infrared markers.
Keywords: infrared sensing, visual motion, photodetector, motion tracking, mobile robots
1. Introduction
Micro-aerial vehicles (MAVs) have required the development of very light-weight and miniature sensing devices. In [1], passive markers or light-emitting diodes (LEDs) are used in the visible spectrum. Therefore, cluttered environments and low-light conditions make their performance decrease. The design of advanced miniature sensors that can substitute for classical devices, such as bulky stereoscopic cameras [2] or ultrasonic sensors [3], is challenging. In [4], ultrasonic systems have been reviewed in detail. Their use in location estimation is discussed in terms of performance, accuracy and limitations. In [5], an acoustic location system for indoor applications is presented. It estimates the target position based on a fingerprinting technique. The result is similar to the Wi-Fi radio-based localization system. Recently, a novel class of optical sensors has been introduced for aerial applications. For instance, new flexible compound-eyes have been proposed for motion extraction and proximity estimation based on the optic flow principle [6,7]. Censi et al. [8] present an approach to low-latency pose tracking using an event-based camera in [9] with active LED markers, which are infrared (IR) LEDs blinking at a high frequency (>1 kHz). The advantage of this camera is that it can track frequencies up to several kilohertz. Monocular vision-based pose estimation is proposed, while accuracy, versatility and performance are improved. Pose estimation is performed with very low latencies. Nevertheless, the precision is constrained by the low sensor resolution, which is 128 × 128 pixels. It is not a commercial technology, and it requires specific skills in programming event-based sensors. In [10], an accurate, efficient and robust pose estimation system based on IR LEDs is also proposed. The system is used to stabilize a quadrotor, both indoors and outdoors. Furthermore, a proximity sensor has also been presented in [11,12]. These studies depict an embedded IR-based 3D relative positioning sensor dedicated to inter-robot spatial-coordination and cooperative flight. Dealing with the ability to determine the spatial orientation and placement of the unmanned aerial platform in real time, the cooperative flight tends to incorporate leader-follower UAVs using vision processing, radio-frequency data transmission and more sensors. Then, the cooperative flight requires the control of only one UAV and should allow the deployment of multiple UAVs. Etter et al. [13] have designed and evaluated a leader-follower autonomous aerial quadrotor system for search and rescue operations using an IR camera and multiple IR beacons. The system incorporates vision processing, radio-frequency data transmission and additional sensors to achieve flocking behavior. Thanks to numerical simulations, IR beacons installed on the platform provides accurate data with respect the spatial orientation and the placement of the system units. Nevertheless, in sunny conditions or with other infrared sources, the IR camera used will have noisy signals. Masselli et al. [14] have proposed a novel method for pose estimation for MAVs using passive visual markers and a monocular color camera mounted onto the MAV as the only sensor. For pose estimation, the P3Pproblem is solved in real time onboard the MAV. The perspective aims at filtering the pose estimate with a Kalman filter in order to accelerate the overall method.
In this paper, we present a novel miniature optical sensor without any optics, called HyperCube, endowed with hyperacuity [15] for active IR LED marker tracking applications. The principle on which HyperCube is based was developed by Kerhuel et al. and detailed in [16,17]. In [16], hyperacuity was obtained at a very low cost. The Vibrating Optical Device for the Kontrol of Autonomous robots (VODKA) sensor was mounted onto a miniature aerial robot that was able to track a moving target accurately by exploiting the robot's uncontrolled random vibration as the source of its microscanning movement. In [17], an improved bio-inspired optical position sensing device was designed. Insect-based retinal microscanning movements were used to detect and locate contrasting objects, such as edges or bars. The active micro-vibrations imposed upon the retina endowed the sensor with hyperacuity. In this new paper, we designed a novel miniature bioinspired sensor where the mechanical vibration is replaced by a flickering LED. It is demonstrated that using only three photosensors with a sensitive area equal to 0.23 mm2, we obtained hyperacuity by replacing the micro-movements with a modulation of infrared LEDs. The markers in use are off-the-shelf IR LEDs flickering at a high frequency (>1 kHz). We envision that this sensor could be used for cooperative flight, as presented in Figure 1.
The purpose of this paper is to present an alternative and a minimalistic solution involving three pixels for 3D localization with respect to IR markers. HyperCube assesses azimuth and elevation measurements of a pattern (planar object) composed of flickering IR LEDs. We also show the abilities of HyperCube to measure its relative position with respect to the markers. We propose a pose estimation system that consists of multiple flickering IR LEDs and a lightweight optical sensor. The LEDs are attached to a moving target, and the sensor provides angular measurements. We also show that the pose estimation is computationally inexpensive.
The paper is organized as follows: Section 2 details the principle of the sensor, the visual processing algorithm and its advantages. The principle of the sensor is discussed in Section 3. The experimental setup is presented in Section 4. Section 5 shows the experimental results.
2. Modeling the HyperCube Sensor
The prototype was built using off-the-shelf elements. Its volume is 1 cm3 and it weights 0.33 g. These characteristics make it inexpensive, light, small and power-lean.
2.1. Angular Sensitivity of the Photosensors
The HyperCube sensor is composed of three surface-mounted device (SMD) photodiodes (VISHAY TEMD7000). Each photodiode (or photosensor) is located on each side of the cube, as shown in Figure 2A. Thanks to this cubic assembly, the optical axes of each pair of photosensors are separated by an angle Δφ = 90°. On each side of the sensor, an analog circuit (Figure 2B) converts the output current of the photodiode into a proportional voltage thanks to a photoconductor electronic circuit (Figure 2C). The linear amplifier in Figure 2C is an analog low-pass filter with a cutoff frequency equal to 106 kHz. Considering the high gain given by the resistance of 150 kΩ, the linear amplifier aims at preventing the signal from oscillations.
HyperCube can extract two angular positions (azimuth and elevation) from the position of an object composed of several IR LEDs. The IR LEDs used in the experimental setup are the OSRAM SFH4232. Their spectrum has a maximum emissive power at a wavelength of 900 nm, which corresponds to the maximum of absorption of the photosensors. Thereafter, the IR LEDs will be considered as a point source. The signal processing used to estimate the elevation and azimuthal angles from the three HyperCube output signals runs in real time onboard a custom-made electronic board described in Section 2.2.
2.2. Model of the Photosensor Output Signal
The cosine-like angular sensitivity function for each photosensor depends on the azimuth and elevation, as shown in Figure 3.
Each IR LED signal is modulated at a specific frequency fi. The modulation is provided by a custom-made electronic board (Figure 4). The latter was designed to provide one specific modulation frequency fi to each LED, such that f1 = 1 kHz, f2 = 3.5 kHz, f3 = 11.5 kHz for the three separate IR LED emitters of the pattern. Then, an additional custom-made processing board (Figure 5) achieved the analog signal demodulation for each frequency fi. To summarize, two electronic boards have been designed: the modulation of the LEDs is performed by the first board in Figure 4, and the second electronic board in Figure 5 performs the analog demodulation and the digital visual processing.
Once the three HyperCube output signals are demodulated, amplified and filtered by a low-pass filter at 100 Hz, one can approximate for each photosensor Phi and frequency fi, the output signal Sphi,fi of HyperCube.
A Gaussian-like directivity function mimics the Gaussian angular sensitivity function of flies' photoreceptors [18,19]. The Gaussian-like angular sensitivity of each photosensor can be defined by the angle of acceptance, denoted Δρ. The angle Δρ is defined as the full width at half maximum, as depicted later in Figure 9. The Gaussian-like angular sensitivity of a photosensor is given by:
(1) |
where θ is the angle between the photosensor's optical axis and that of a point light source. We also introduce the cosine-like angular sensitivity function σ, such that:
(2) |
From Figure 6, the angular sensitivity of a photosensor as a blue dotted dashed line is compared to the Gaussian angular sensitivity given in Equation (1) as a green dashed line. One can note that the angular sensitivity of a photosensor does not strictly shape the Gaussian-like one and could also be approximated by a cosine-like angular sensitivity, which is plotted in red.
To characterize the dynamic response of the HyperCube sensor, we measured the demodulated output signal of a photosensor at 1 kHz in response to a step input. We turned on a modulated IR LED placed in front of the photosensor Ph1, and we measured the demodulated output signal by means of the board, as shown in Figure 5. To precisely know the step time, the power input of the modulation electronic board (+3.3 V) is connected to a digital input of the microcontroller. The sampling frequency is 10 kHz. As depicted in Figure 7, the settling time t95% is equal to 7.25 ms, and the rise time tm is equal to 1.6 ms. Therefore, given these dynamic properties, the use of the HyperCube sensor could be relevant for MAV applications.
From Figure 8, the azimuth angle φ and the elevation angle ψ are defined.
Therefore, SPhi,fi the output signal of the photodiode i is related to a specific flickering frequency fi as follows:
(3) |
(4) |
(5) |
With A1,i, A2,i, A3,i, the gains of the photosensors Ph1, Ph2 or Ph3 on each side of the HyperCube sensor. The index i refers to the frequencies 1 kHz, 3.5 kHz, 11.5 kHz. φ and ψ are the azimuth and elevation with respect the IR LED i.
The three photosensor (photodiode) output signals are processed by a low-cost microcontroller from Microchip (dsPIC33FJ128GP802). As shown in Figure 9A,B, the digital processing operates at 250 Hz and computes the relative difference over the sum of two adjacent photosensor signals. For each frequency 1 kHz, 3.5 kHz and 11.5 kHz, respectively, a custom-made demodulation board was designed as depicted in Figure 5. Therefore, for each frequency fi, a Sallen Key peak filter and a low pass filter with a cut off frequency equal to 100 Hz were chosen. The demodulated output signals , SPh2 and SPh3 were processed at 1kHz, 3.5 kHz and 11.5 kHz, respectively.
2.3. Advantages of HyperCube
HyperCube exhibits several advantages over other conventional sensors, like CMOS cameras, according to these following points:
low cost: there is no optics; the angular position measurement relies only on the angular sensitivity of photodiodes, which is, in essence, non-uniform (cosine law) and only requires very inexpensive off-the-shelf electronic components.
low computational resources: the measurements are the angular information; this does not require any image processing; a high refresh rate for the the same computational resource can easily be reached.
small size: because of a low number of components, the sensor is light weight (0.33 g) and very compact (1cm3).
3. Principle of the HyperCube Sensor
3.1. Angle Reconstruction
Let us consider the azimuth angle φ. It is defined as the angle between the reference plane Πφ (see Figure 8A) and a plane including the IR LED (see the red dotted line in Figure 8C). Πφ is the mid-plane between the photosensors Ph1 and Ph2. The second plane includes the IR LED and the intersection between the optical axes of Ph1 and Ph2. In Figure 8A, the azimuth φ is presented in front view and in top view in Figure 8C.
The digital processing operated in the microcontroller with respect to the azimuth φ returns an output signal whose theoretical expression can be written as:
(6) |
Equation (6) gives the output signal, which results from two photosensors (Ph1 and Ph2, respectively), to obtain the ratio between the difference and the sum of the differentiated and demodulated photosensor signals. This expression reduces the common mode noise signal introduced by the artificial light. Moreover, the expression of Equation (6) results from previous studies carried out at the laboratory on hyperacute visual sensors based on active micro-movements applied to an artificial eye in [16,17]. It is worth noting that we obtained here hyperacuity by replacing the micro-movements with a modulation of IR LEDs. From Equation (6) using Equations (3) and (4), one can write:
(7) |
A1 and A2 are the amplitudes of the signals delivered by the photosensors Ph1 and Ph2. If we assume A1 = A2, Equation (7) gives:
(8) |
Therefore:
(9) |
The elevation angle ψ is defined as the angle between the reference plane Πψ (see Figure 8B) and a plane including the IR LED (see the red dotted line in Figure 8D). The second plane includes the IR LED and the intersection of the optical axes of the virtual photodiode (Ph1 + Ph2) and Ph3. In Figure 8B, the elevation angle is presented in front view and in side view in Figure 8D.
The digital processing with respect to the elevation angle ψ involves a virtual photosensor (Ph1 + Ph2), which is the combination of Ph1 and Ph2. In the following, we introduce the signal SPhvirtual delivered by the virtual photosensor:
Similarly to Equation (9), if we assume the amplitudes are such that A = A1 = A2 = A3, the digital signal output for the elevation angle ψ gives: . Using the definition of SPhvirtual, one can write if :
(10) |
Considering the approximation of small angles for φ, σ (φ) ≃ 1, one can simplify:
Therefore:
(11) |
3.2. Position Estimation
Given the signal outputs Sψ and Sφ and one IR LED, the first step was to assess the relative position of the IR LED in the plane (XLED,YLED) with respect to HyperCube. As described in the experimental setup presented in Section 4, if we assume ẐLED is a priori known or assessed, using only angular measurements provided by HyperCube, the relative position of the IR LED with respect the sensor can be estimated such as:
(12) |
(13) |
where φ < 0 and ψ < 0 as defined in Figure 9A,B.
The experimental setup detailed in Section 4 justifies the approximation of the small angles. Therefore, one can assume that:
(14) |
(15) |
4. Experimental Setup
In this work, we aimed at demonstrating that the estimation of the relative position of an IR LED with respect to HyperCube can be achieved by means of angular measurements. The experimental setup is composed of a motor-driven XY table, as shown in Figure 10. We oriented HyperCube so that it was pointing upward. HyperCube can move along the X and Y directions thanks to two DC motors.
Two incremental encoders measure the HyperCube position along the X and Y directions. In the initial conditions, which define the origin of the inertial frame, an IR LED is located just above HyperCube at a height ZLED = 300 mm. As an absolute sensor's position of reference, a motion capture system provided by the company VICON was used for measuring the ground truth, i.e., the precise absolute position of HyperCube. The tracking system is equipped with 17 IR cameras and IR active markers that are pointed upward, as depicted in Figure 10. The HyperCube position of reference was therefore measured in real time with a sub-millimetric precision.
4.1. Calibration
In this work, we consider that the HyperCube sensor can move in a range of [−50 mm; +50 mm] in both X and Y directions. Given ZLED = 300 mm, one can assume that the angles are small enough to write from Equation (9) Sφ = −tan(φ) ⋍ −φ for azimuth and from Equation (11) Sψ = −tan(ψ) ⋍ −ψ for elevation. As shown in the experimental setup in Figure 10, we theoretically have and , where XLED, YLED and ZLED are the relative position of the IR LED to the HyperCube sensor. The positions of reference XLED, YLED and ZLED are provided by the tracking system VICON.
The calibration consists of adjusting the HyperCube outputs Sφ and Sψ to the ratios and . To this aim, the calibration curves shown in Figure 11 are plotted. We note that the HyperCube outputs are linear in the observed range [−50 mm; +50 mm]. Linear regression computes the slope (aφ, aψ) and the offset (bφ, bψ) by fitting each curve with a linear model. These coefficients calibrate the HyperCube outputs and , such that: and . Figure 11 presents the calibration results for each IR LED in both the X and Y directions. The sensor's outputs are linear in the range [−50 mm; +50 mm] with respect to the origin.
The calibration validation protocol takes ZLED = 300 mm. The estimated positions X̂LED and ŶLED of the IR LED were compared to the references XLED and YLED provided by the VICON system.
4.2. Height Estimation
Form the experimental setup described in Figure 10 and given the signal outputs Sψ and Sφ and two IR LEDs, one can assess the height ZLED between the sensor and two IR LEDs A and B.
The assessed height ẐLED is given by the relation: where AB is the distance between the LEDs A and B. φA and φB are the azimuthal angles given by HyperCube for the LEDs A and B. In the configuration described in Figure 12, YLED = 0 (i.e., ABPis orthogonal to the horizontal plane that intersects HyperCube).
5. Experimental Position Estimation in 2D
In this section, we used a single IR LED and HyperCube for each experiment.
We compared the estimation to HyperCube positions assessed by the tracking system for three different modulation frequencies. For the IR LED that flickers at 1 kHz, Figure 13A shows the plots of the reference trajectory as a dashed line and the trajectory depicted by HyperCube as a solid line. The trajectory is a circular path with a radius of 50 mm in the XY plane with ZLED = 300 mm. The distance between reference and estimated trajectories was also plotted. The histogram of the error gives a mean value μ = 1.6 mm and a standard deviation σ = 2.8 mm. We noted that the noise at 1 kHz for the demodulation was lower than the noise at 3.5 kHz and 11.5 kHz, as shown in Figure 13B,C. The higher levels of noise at 3.5 kHz and 11.5 kHz than the one at 1 kHz are due to the use of different operational amplifiers (TI OPA4322 and TIOPA4342) for the analog demodulation. Figure 13A-C highlight that the HyperCube sensor provides useful measurements to accurately estimate the dynamic relative position of an IR LED in the XY plane.
To reduce the measurement noise, the digital signal outputs given by the HyperCube sensor were filtered by a Kalman filter. The system is described in a state space form as follows:
(16) |
(17) |
where the state xk = [Xk Ẋk]T describes the position Xk and the velocity Ẋk at the time instant tk. wk and vk are the state and measurement white noise with known and constant covariance matrices W and V, respectively. They are supposed to be mutually uncorrelated. is the state transition matrix, C = [1 0] the observation matrix. In practice, we choose the covariance matrices as follows and where σmeas and σproc are the standard deviations of the measurement noise and the state noise, respectively.
A linear Kalman filter was also implemented for the state estimation along the Y direction.
5.1. Experimental Height Estimation
As described in Section 4.2, the height between the HyperCube sensor and a pair of IR LEDs was estimated. Thanks to the angular measurements using three flickering IR LEDs, one can reduce the measurement noise. The height estimation precision is therefore increased. The mean of the height estimate Ẑ̅ is written as:
(18) |
where ẐLED1,2, ẐLED2,3 and ẐLED1,3 are the height estimated by the pairs (1,2), (2,3) and (1,3), respectively. The height of the IR LEDS was modified, while the position of the HyperCube sensor was fixed in the XY plane. Using each pair of IR LEDs, the height estimation was computed on-line. From Figure 14, the height estimation as a solid line is precise around 300 mm with respect to the reference height as a dashed line.
In Figure 15, for different pairs of IR LEDs, we plotted the evolution of the height estimation while the sensor was moving relatively to the center of the pair of IR LEDs. We note from Figure 15 that the height estimation depends on the position (XLED, YLED) of the sensor. The estimation error can vary up to 50 mm.
The estimation remains better with the pair of IR LEDs (1 kHz and 3.5 kHz) than with the others.
5.2. IR LED Position Tracking in 2D
In this section, HyperCube measurements, which consist of azimuth φ and elevation ψ, feed the controller that aims at tracking a moving IR LED in the plane. A Smith predictor with proportional and integral actions was designed and implemented to ensure the good stability of the linear displacement of HyperCube despite the time lag inherent to the visual processing algorithm. Figure 16A shows the block diagram of the closed loop system with the assessed angles φ and ψ. In Figure 16B,C, estimated positions versus time are plotted. The tracking system gives the reference trajectory.
As shown in Figure 16B,C, the position tracking was accurate. The tracking error plotted in the Figure 16D shows that the error ranges within ±5 cm. Nevertheless, one can note in Figure 16B that there is an error between the estimated and the reference values due to a saturation of the motor of the X axis. Thanks to a fast motor for controlling the Y axis, HyperCube was able to follow faithfully the moving pattern. In Figure 16C, the HyperCube sensor perfectly tracks the moving IR LED along the Y axis.
6. Conclusions
We designed and presented the HyperCube sensor. It is a cubic, miniature and low-cost sensing device able to track active markers without any optics. The system discussed has the following key features:
It provides azimuth and elevation and is able to assess the angular position of three blinking IR LEDs at specific frequencies: 1 kHz, 3.5 kHz and 11.5 kHz.
The estimation is achieved by a computationally-inexpensive signal processing of the low-amplitude and high-frequency signal transmitted by the three photodiodes placed on each side of the cube.
Experimental results showed that for the position estimation and the position tracking in 2D, this miniature device is very efficient. It also features promising results in terms of distance estimation, as well.
Nevertheless, the prototype developed reveals some limitations with respect to the range of pose estimation, but it is well adapted for proximal localization (up to 30 cm). In addition, the size of the demodulation electronic board could be easily reduced. As a conclusion, the HyperCube sensor is a promising sensing device for proximal localization applications and could equip UAVs or MAVs. The short-term outlook consists of improving the range of sensitivity of HyperCube using silicon photodiodes, which are specifically designed for precision photometry and feature a high sensitivity, a high-speed response and low noise.
Acknowledgements
We thank M. Boyron for his assistance with the electrical design, B. Boisseau for his involvement in the experimentation and J. Diperi for his involvement in the mechanical design of the sensor. We also thank N. Franceschini for the motor-driven XY table.
Author Contributions
Thibaut Raharijaona, Raphaël Juston and Stéphane Viollet designed the research. Paul Mignon, Lubin Kerhuel, Raphaël Juston and Thibaut Raharijaona performed the research. Thibaut Raharijaona and Stéphane Viollet wrote the paper.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Breitenmoser A., Kneip L., Siegwart R. A monocular vision-based system for 6D relative robot localization. Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); San Francisco, CA, USA. 25–30 September 2011; pp. 79–85. [Google Scholar]
- 2.Audette R., Balthazaar J., Dunk C., Zelek J. A Stereo-Vision System for the Visually Impaired. University of Guelph; Guelph, ON, Canada: 2000. [Google Scholar]
- 3.Cardin S., Vexo F., Thalmann D. Wearable system for mobility improvement of visually impaired people. Vis. Comp. J. 2006;23:109–118. [Google Scholar]
- 4.Ijaz F., Yang H.K., Ahmad A., Lee C. Indoor positioning: A review of indoor ultrasonic positioning systems. Proceedings of the 15th International Conference on Advanced Communication Technology (ICACT); PyeongChang, Korea. 27–30 Janunary 2013; pp. 1146–1150. [Google Scholar]
- 5.Aloui N., Raoof K., Bouallegue A., Letourneur S., Zaibi S. Performance evaluation of an acoustic indoor localization system based on a fingerprinting technique. EURASIP J. Adv. Signal Process. 2014;2014 doi: 10.1186/1687-6180-2014-13. [DOI] [Google Scholar]
- 6.Dobrzynski M., Pericet-Camara R., Floreano D. Vision tape—A flexible compound camera for motion detection and proximity estimation. IEEE Sens. J. 2012;12:1131–1139. [Google Scholar]
- 7.Floreano D., Pericet-Camara R., Viollet S., Ruffier F., Brückner A., Leitel R., Buss W., Menouni M., Expert F., Juston R., et al. Miniature curved artificial compound eyes. Proc. Natl. Acad. Sci. USA. 2013;110:9267–9272. doi: 10.1073/pnas.1219068110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Censi A., Strubel J., Brandli C., Delbruck T., Scaramuzza D. Low-latency localization by active LED markers tracking using a dynamic vision sensor. Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Tokyo, Japan. 3–7 November 2013; pp. 891–898. [Google Scholar]
- 9.Lichtsteiner P., Posch C., Delbruck T. A 128× 128 120 dB 15 μ s latency asynchronous temporal contrast vision sensor. IEEE J. Sol. State Circuits. 2008;43:566–576. [Google Scholar]
- 10.Faessler M., Mueggler E., Schwabe K., Scaramuzza D. A monocular pose estimation system based on Infrared LEDs. Proceedinds of the IEEE International Conference on Robotics and Automation (ICRA); Hong Kong, China. 31 May–7 June 2014. [Google Scholar]
- 11.Roberts J., Stirling T., Zufferey J.C., Floreano D. 3-D relative positioning sensor for indoor flying robots. Auton. Robots. 2012;33:5–20. [Google Scholar]
- 12.Wenzel K., Masselli A., Zell A. Visual tracking and following of a quadrocopter by another quadrocopter. Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Vilamoura, Portugal. 7–12 October 2012; pp. 4993–4998. [Google Scholar]
- 13.Etter W., Martin P., Mangharam R. Cooperative flight guidance of autonomous unmanned aerial vehicles. Proceedings of the 2nd Internation Workshop on Networks of Cooperating Objects; Chicago, IL, USA. 11–12 April 2011. [Google Scholar]
- 14.Masselli A., Zell A. A novel marker based tracking method for position and attitude control of MAVs. Proceedings of International Micro Air Vehicle Conference and Flight Competition (IMAV); Braunschweig, Germany. 3–6 July 2012. [Google Scholar]
- 15.Westheimer G. Visual hyperacuity. Progress in Sensory Physiology. Springer; Berlin, Germany: 1981. pp. 1–30. [Google Scholar]
- 16.Kerhuel L., Viollet S., Franceschini N. The VODKA sensor: A bio-inspired hyperacute optical position sensing device. IEEE Sens. J. 2012;12:315–324. [Google Scholar]
- 17.Juston R., Kerhuel L., Franceschini N., Viollet S. Hyperacute edge and bar detection in a bioinspired optical position sensing device. IEEE/ASME Trans. Mechatron. 2013;19:1025–1034. [Google Scholar]
- 18.Götz K. Optomotorische Untersuchung des visuellen Systems einiger Augenmutanten der Fruchtfliege. Drosophila. Biol. Cybern. 1964;2:77–92. doi: 10.1007/BF00288561. [DOI] [PubMed] [Google Scholar]
- 19.Stavenga D. Angular and spectral sensitivity of fly photoreceptors. II. dependence on facet lens F-number and rhabdomere type in Drosophila. J. Comp. Phys. A. 2003;189:189–202. doi: 10.1007/s00359-003-0390-6. [DOI] [PubMed] [Google Scholar]