Next Article in Journal
Application of 3D Printing Technology in Increasing the Diagnostic Performance of Enzyme-Linked Immunosorbent Assay (ELISA) for Infectious Diseases
Next Article in Special Issue
Infrared Thermography to Evaluate Heat Tolerance in Different Genetic Groups of Lambs
Previous Article in Journal
Design and Realization of an Electromagnetic Guiding System for Blind Running Athletes
Previous Article in Special Issue
An Evaluation of the Pedestrian Classification in a Multi-Domain Multi-Modality Setup
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HyperCube: A Small Lensless Position Sensing Device for the Tracking of Flickering Infrared LEDs

Aix-Marseille Université, ISM UMR 7287, 13288, Marseille Cedex 09, France
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(7), 16484-16502; https://doi.org/10.3390/s150716484
Submission received: 30 March 2015 / Revised: 28 May 2015 / Accepted: 2 July 2015 / Published: 8 July 2015
(This article belongs to the Special Issue Frontiers in Infrared Photodetection)

Abstract

: An innovative insect-based visual sensor is designed to perform active marker tracking. Without any optics and a field-of-view of about 60°, a novel miniature visual sensor is able to locate flickering markers (LEDs) with an accuracy much greater than the one dictated by the pixel pitch. With a size of only 1 cm3 and a mass of only 0.33 g, the lensless sensor, called HyperCube, is dedicated to 3D motion tracking and fits perfectly with the drastic constraints imposed by micro-aerial vehicles. Only three photosensors are placed on each side of the cubic configuration of the sensing device, making this sensor very inexpensive and light. HyperCube provides the azimuth and elevation of infrared LEDs flickering at a high frequency (>1 kHz) with a precision of 0.5°. The minimalistic design in terms of small size, low mass and low power consumption of this visual sensor makes it suitable for many applications in the field of the cooperative flight of unmanned aerial vehicles and, more generally, robotic applications requiring active beacons. Experimental results show that HyperCube provides useful angular measurements that can be used to estimate the relative position between the sensor and the flickering infrared markers.

1. Introduction

Micro-aerial vehicles (MAVs) have required the development of very light-weight and miniature sensing devices. In [1], passive markers or light-emitting diodes (LEDs) are used in the visible spectrum. Therefore, cluttered environments and low-light conditions make their performance decrease. The design of advanced miniature sensors that can substitute for classical devices, such as bulky stereoscopic cameras [2] or ultrasonic sensors [3], is challenging. In [4], ultrasonic systems have been reviewed in detail. Their use in location estimation is discussed in terms of performance, accuracy and limitations. In [5], an acoustic location system for indoor applications is presented. It estimates the target position based on a fingerprinting technique. The result is similar to the Wi-Fi radio-based localization system. Recently, a novel class of optical sensors has been introduced for aerial applications. For instance, new flexible compound-eyes have been proposed for motion extraction and proximity estimation based on the optic flow principle [6,7]. Censi et al. [8] present an approach to low-latency pose tracking using an event-based camera in [9] with active LED markers, which are infrared (IR) LEDs blinking at a high frequency (>1 kHz). The advantage of this camera is that it can track frequencies up to several kilohertz. Monocular vision-based pose estimation is proposed, while accuracy, versatility and performance are improved. Pose estimation is performed with very low latencies. Nevertheless, the precision is constrained by the low sensor resolution, which is 128 × 128 pixels. It is not a commercial technology, and it requires specific skills in programming event-based sensors. In [10], an accurate, efficient and robust pose estimation system based on IR LEDs is also proposed. The system is used to stabilize a quadrotor, both indoors and outdoors. Furthermore, a proximity sensor has also been presented in [11,12]. These studies depict an embedded IR-based 3D relative positioning sensor dedicated to inter-robot spatial-coordination and cooperative flight. Dealing with the ability to determine the spatial orientation and placement of the unmanned aerial platform in real time, the cooperative flight tends to incorporate leader-follower UAVs using vision processing, radio-frequency data transmission and more sensors. Then, the cooperative flight requires the control of only one UAV and should allow the deployment of multiple UAVs. Etter et al. [13] have designed and evaluated a leader-follower autonomous aerial quadrotor system for search and rescue operations using an IR camera and multiple IR beacons. The system incorporates vision processing, radio-frequency data transmission and additional sensors to achieve flocking behavior. Thanks to numerical simulations, IR beacons installed on the platform provides accurate data with respect the spatial orientation and the placement of the system units. Nevertheless, in sunny conditions or with other infrared sources, the IR camera used will have noisy signals. Masselli et al. [14] have proposed a novel method for pose estimation for MAVs using passive visual markers and a monocular color camera mounted onto the MAV as the only sensor. For pose estimation, the P3Pproblem is solved in real time onboard the MAV. The perspective aims at filtering the pose estimate with a Kalman filter in order to accelerate the overall method.

In this paper, we present a novel miniature optical sensor without any optics, called HyperCube, endowed with hyperacuity [15] for active IR LED marker tracking applications. The principle on which HyperCube is based was developed by Kerhuel et al. and detailed in [16,17]. In [16], hyperacuity was obtained at a very low cost. The Vibrating Optical Device for the Kontrol of Autonomous robots (VODKA) sensor was mounted onto a miniature aerial robot that was able to track a moving target accurately by exploiting the robot's uncontrolled random vibration as the source of its microscanning movement. In [17], an improved bio-inspired optical position sensing device was designed. Insect-based retinal microscanning movements were used to detect and locate contrasting objects, such as edges or bars. The active micro-vibrations imposed upon the retina endowed the sensor with hyperacuity. In this new paper, we designed a novel miniature bioinspired sensor where the mechanical vibration is replaced by a flickering LED. It is demonstrated that using only three photosensors with a sensitive area equal to 0.23 mm2, we obtained hyperacuity by replacing the micro-movements with a modulation of infrared LEDs. The markers in use are off-the-shelf IR LEDs flickering at a high frequency (>1 kHz). We envision that this sensor could be used for cooperative flight, as presented in Figure 1.

The purpose of this paper is to present an alternative and a minimalistic solution involving three pixels for 3D localization with respect to IR markers. HyperCube assesses azimuth and elevation measurements of a pattern (planar object) composed of flickering IR LEDs. We also show the abilities of HyperCube to measure its relative position with respect to the markers. We propose a pose estimation system that consists of multiple flickering IR LEDs and a lightweight optical sensor. The LEDs are attached to a moving target, and the sensor provides angular measurements. We also show that the pose estimation is computationally inexpensive.

The paper is organized as follows: Section 2 details the principle of the sensor, the visual processing algorithm and its advantages. The principle of the sensor is discussed in Section 3. The experimental setup is presented in Section 4. Section 5 shows the experimental results.

2. Modeling the HyperCube Sensor

The prototype was built using off-the-shelf elements. Its volume is 1 cm3 and it weights 0.33 g. These characteristics make it inexpensive, light, small and power-lean.

2.1. Angular Sensitivity of the Photosensors

The HyperCube sensor is composed of three surface-mounted device (SMD) photodiodes (VISHAY TEMD7000). Each photodiode (or photosensor) is located on each side of the cube, as shown in Figure 2A. Thanks to this cubic assembly, the optical axes of each pair of photosensors are separated by an angle Δφ = 90°. On each side of the sensor, an analog circuit (Figure 2B) converts the output current of the photodiode into a proportional voltage thanks to a photoconductor electronic circuit (Figure 2C). The linear amplifier in Figure 2C is an analog low-pass filter with a cutoff frequency equal to 106 kHz. Considering the high gain given by the resistance of 150 kΩ, the linear amplifier aims at preventing the signal from oscillations.

HyperCube can extract two angular positions (azimuth and elevation) from the position of an object composed of several IR LEDs. The IR LEDs used in the experimental setup are the OSRAM SFH4232. Their spectrum has a maximum emissive power at a wavelength of 900 nm, which corresponds to the maximum of absorption of the photosensors. Thereafter, the IR LEDs will be considered as a point source. The signal processing used to estimate the elevation and azimuthal angles from the three HyperCube output signals runs in real time onboard a custom-made electronic board described in Section 2.2.

2.2. Model of the Photosensor Output Signal

The cosine-like angular sensitivity function for each photosensor depends on the azimuth and elevation, as shown in Figure 3.

Each IR LED signal is modulated at a specific frequency fi. The modulation is provided by a custom-made electronic board (Figure 4). The latter was designed to provide one specific modulation frequency fi to each LED, such that f1 = 1 kHz, f2 = 3.5 kHz, f3 = 11.5 kHz for the three separate IR LED emitters of the pattern. Then, an additional custom-made processing board (Figure 5) achieved the analog signal demodulation for each frequency fi. To summarize, two electronic boards have been designed: the modulation of the LEDs is performed by the first board in Figure 4, and the second electronic board in Figure 5 performs the analog demodulation and the digital visual processing.

Once the three HyperCube output signals are demodulated, amplified and filtered by a low-pass filter at 100 Hz, one can approximate for each photosensor Phi and frequency fi, the output signal Sphi,fi of HyperCube.

A Gaussian-like directivity function mimics the Gaussian angular sensitivity function of flies' photoreceptors [18,19]. The Gaussian-like angular sensitivity of each photosensor can be defined by the angle of acceptance, denoted Δρ. The angle Δρ is defined as the full width at half maximum, as depicted later in Figure 9. The Gaussian-like angular sensitivity of a photosensor is given by:

σ ( θ ) = 2 π ln ( 2 ) π Δ ρ e 4 ln ( 2 ) ( θ 2 Δ ρ 2 )
where θ is the angle between the photosensor's optical axis and that of a point light source. We also introduce the cosine-like angular sensitivity function σ, such that:
σ ( θ ) = { cos ( θ ) i f | θ | π 2 0 else

From Figure 6, the angular sensitivity of a photosensor as a blue dotted dashed line is compared to the Gaussian angular sensitivity given in Equation (1) as a green dashed line. One can note that the angular sensitivity of a photosensor does not strictly shape the Gaussian-like one and could also be approximated by a cosine-like angular sensitivity, which is plotted in red.

To characterize the dynamic response of the HyperCube sensor, we measured the demodulated output signal of a photosensor at 1 kHz in response to a step input. We turned on a modulated IR LED placed in front of the photosensor Ph1, and we measured the demodulated output signal by means of the board, as shown in Figure 5. To precisely know the step time, the power input of the modulation electronic board (+3.3 V) is connected to a digital input of the microcontroller. The sampling frequency is 10 kHz. As depicted in Figure 7, the settling time t95% is equal to 7.25 ms, and the rise time tm is equal to 1.6 ms. Therefore, given these dynamic properties, the use of the HyperCube sensor could be relevant for MAV applications.

From Figure 8, the azimuth angle φ and the elevation angle ψ are defined.

Therefore, SPhi,fi the output signal of the photodiode i is related to a specific flickering frequency fi as follows:

S P h 1 , f i ( φ , ψ ) = A 1 , i . σ ( φ π 4 ) . σ ( ψ + π 4 ) if | φ π 4 | < π 2 and | ψ + π 4 | < π 2 else S P h 1 , f i ( φ , ψ ) = 0
S P h 2 , f i ( φ , ψ ) = A 2 , i . σ ( φ + π 4 ) . σ ( ψ + π 4 )
if | φ + π 4 | < π 2 and | ψ + π 4 | < π 2 else S P h 2 , f i ( φ , ψ ) = 0 S P h 3 , f i ( φ , ψ ) = A 3 , i . σ ( φ ) . σ ( ψ + π 4 ) if | φ | < π 2 and | ψ π 4 | < π 2 else S P h 3 , f i ( φ , ψ ) = 0

With A1,i, A2,i, A3,i, the gains of the photosensors Ph1, Ph2 or Ph3 on each side of the HyperCube sensor. The index i refers to the frequencies 1 kHz, 3.5 kHz, 11.5 kHz. φ and ψ are the azimuth and elevation with respect the IR LED i.

The three photosensor (photodiode) output signals are processed by a low-cost microcontroller from Microchip (dsPIC33FJ128GP802). As shown in Figure 9A,B, the digital processing operates at 250 Hz and computes the relative difference over the sum of two adjacent photosensor signals. For each frequency 1 kHz, 3.5 kHz and 11.5 kHz, respectively, a custom-made demodulation board was designed as depicted in Figure 5. Therefore, for each frequency fi, a Sallen Key peak filter and a low pass filter with a cut off frequency equal to 100 Hz were chosen. The demodulated output signals , SPh2 and SPh3 were processed at 1kHz, 3.5 kHz and 11.5 kHz, respectively.

2.3. Advantages of HyperCube

HyperCube exhibits several advantages over other conventional sensors, like CMOS cameras, according to these following points:

  • low cost: there is no optics; the angular position measurement relies only on the angular sensitivity of photodiodes, which is, in essence, non-uniform (cosine law) and only requires very inexpensive off-the-shelf electronic components.

  • low computational resources: the measurements are the angular information; this does not require any image processing; a high refresh rate for the the same computational resource can easily be reached.

  • small size: because of a low number of components, the sensor is light weight (0.33 g) and very compact (1cm3).

3. Principle of the HyperCube Sensor

3.1. Angle Reconstruction

Let us consider the azimuth angle φ. It is defined as the angle between the reference plane Πφ (see Figure 8A) and a plane including the IR LED (see the red dotted line in Figure 8C). Πφ is the mid-plane between the photosensors Ph1 and Ph2. The second plane includes the IR LED and the intersection between the optical axes of Ph1 and Ph2. In Figure 8A, the azimuth φ is presented in front view and in top view in Figure 8C.

The digital processing operated in the microcontroller with respect to the azimuth φ returns an output signal whose theoretical expression can be written as:

S φ = S p h 1 S p h 2 S p h 1 + S p h 2

Equation (6) gives the output signal, which results from two photosensors (Ph1 and Ph2, respectively), to obtain the ratio between the difference and the sum of the differentiated and demodulated photosensor signals. This expression reduces the common mode noise signal introduced by the artificial light. Moreover, the expression of Equation (6) results from previous studies carried out at the laboratory on hyperacute visual sensors based on active micro-movements applied to an artificial eye in [16,17]. It is worth noting that we obtained here hyperacuity by replacing the micro-movements with a modulation of IR LEDs. From Equation (6) using Equations (3) and (4), one can write:

S φ = A 1 σ ( φ + π 4 ) · σ ( ψ + π 4 ) A 2 σ ( φ π 4 ) · σ ( ψ + π 4 ) A 1 σ ( φ + π 4 ) · σ ( ψ + π 4 ) + A 2 σ ( φ π 4 ) · σ ( ψ + π 4 )

A1 and A2 are the amplitudes of the signals delivered by the photosensors Ph1 and Ph2. If we assume A1 = A2, Equation (7) gives:

S φ = cos ( φ + π 4 ) cos ( φ π 4 ) cos ( φ + π 4 ) + cos ( φ π 4 )

Therefore:

S φ = tan ( φ )

The elevation angle ψ is defined as the angle between the reference plane Πψ (see Figure 8B) and a plane including the IR LED (see the red dotted line in Figure 8D). The second plane includes the IR LED and the intersection of the optical axes of the virtual photodiode (Ph1 + Ph2) and Ph3. In Figure 8B, the elevation angle is presented in front view and in side view in Figure 8D.

The digital processing with respect to the elevation angle ψ involves a virtual photosensor (Ph1 + Ph2), which is the combination of Ph1 and Ph2. In the following, we introduce the signal SPhvirtual delivered by the virtual photosensor:

S P h virtual = S P h 1 + S P h 2 = A σ ( ψ + π 4 ) . [ σ ( φ + π 4 ) + σ ( φ π 4 ) ] S P h virtual = { A σ ( ψ + π 4 ) i f | φ | < π 4 S P h 1 i f π 4 < φ < 3 π 4 S P h 2 i f 3 π 4 < φ < π 4 0 else

Similarly to Equation (9), if we assume the amplitudes are such that A = A1 = A2 = A3, the digital signal output for the elevation angle ψ gives: S ψ = S P h virtual S P h 3 S P h virtual + S P h 3. Using the definition of SPhvirtual, one can write if | φ | < π 4:

S ψ = σ ( ψ + π 4 ) σ ( φ ) . σ ( ψ π 4 ) σ ( ψ + π 4 ) σ ( φ ) . σ ( ψ π 4 )

Considering the approximation of small angles for φ, σ (φ) ≃ 1, one can simplify:

S ψ = σ ( ψ + π 4 ) σ ( ψ π 4 ) σ ( ψ + π 4 ) + σ ( ψ π 4 ) = cos ( ψ + π 4 ) cos ( ψ π 4 ) cos ( ψ + π 4 ) + cos ( ψ π 4 )

Therefore:

S ψ tan ( ψ )

3.2. Position Estimation

Given the signal outputs Sψ and Sφ and one IR LED, the first step was to assess the relative position of the IR LED in the plane (XLED,YLED) with respect to HyperCube. As described in the experimental setup presented in Section 4, if we assume LED is a priori known or assessed, using only angular measurements provided by HyperCube, the relative position of the IR LED with respect the sensor can be estimated such as:

X ^ LED = tan ( φ ) Z ^ LED
Y ^ LED = tan ( ψ ) Z ^ LED
where φ < 0 and ψ < 0 as defined in Figure 9A,B.

The experimental setup detailed in Section 4 justifies the approximation of the small angles. Therefore, one can assume that:

X ^ LED = φ · Z ^ LED
Y ^ LED = ψ · Z ^ LED

4. Experimental Setup

In this work, we aimed at demonstrating that the estimation of the relative position of an IR LED with respect to HyperCube can be achieved by means of angular measurements. The experimental setup is composed of a motor-driven XY table, as shown in Figure 10. We oriented HyperCube so that it was pointing upward. HyperCube can move along the X and Y directions thanks to two DC motors.

Two incremental encoders measure the HyperCube position along the X and Y directions. In the initial conditions, which define the origin of the inertial frame, an IR LED is located just above HyperCube at a height ZLED = 300 mm. As an absolute sensor's position of reference, a motion capture system provided by the company VICON was used for measuring the ground truth, i.e., the precise absolute position of HyperCube. The tracking system is equipped with 17 IR cameras and IR active markers that are pointed upward, as depicted in Figure 10. The HyperCube position of reference was therefore measured in real time with a sub-millimetric precision.

4.1. Calibration

In this work, we consider that the HyperCube sensor can move in a range of [−50 mm; +50 mm] in both X and Y directions. Given ZLED = 300 mm, one can assume that the angles are small enough to write from Equation (9) Sφ = −tan(φ) ⋍ −φ for azimuth and from Equation (11) Sψ = −tan(ψ) ⋍ −ψ for elevation. As shown in the experimental setup in Figure 10, we theoretically have S φ = tan φ = X LED Z LED and S ψ = tan ψ = Y LED Z LED, where XLED, YLED and ZLED are the relative position of the IR LED to the HyperCube sensor. The positions of reference XLED, YLED and ZLED are provided by the tracking system VICON.

The calibration consists of adjusting the HyperCube outputs Sφ and Sψ to the ratios X LED Z LED and Y LED Z LED. To this aim, the calibration curves shown in Figure 11 are plotted. We note that the HyperCube outputs are linear in the observed range [−50 mm; +50 mm]. Linear regression computes the slope (aφ, aψ) and the offset (bφ, bψ) by fitting each curve with a linear model. These coefficients calibrate the HyperCube outputs S φ c and S ψ c, such that: S φ c = a φ . S φ + b φ = X LED Z LED = φ and S ψ c = a ψ . S ψ + b ψ = X LED Z LED = ψ. Figure 11 presents the calibration results for each IR LED in both the X and Y directions. The sensor's outputs are linear in the range [−50 mm; +50 mm] with respect to the origin.

The calibration validation protocol takes ZLED = 300 mm. The estimated positions LED and LED of the IR LED were compared to the references XLED and YLED provided by the VICON system.

4.2. Height Estimation

Form the experimental setup described in Figure 10 and given the signal outputs Sψ and Sφ and two IR LEDs, one can assess the height ZLED between the sensor and two IR LEDs A and B.

The assessed height LED is given by the relation: Z ^ LED = A B tan ( φ A ) + tan ( φ B ) where AB is the distance between the LEDs A and B. φA and φB are the azimuthal angles given by HyperCube for the LEDs A and B. In the configuration described in Figure 12, YLED = 0 (i.e., ABPis orthogonal to the horizontal plane that intersects HyperCube).

5. Experimental Position Estimation in 2D

In this section, we used a single IR LED and HyperCube for each experiment.

We compared the estimation to HyperCube positions assessed by the tracking system for three different modulation frequencies. For the IR LED that flickers at 1 kHz, Figure 13A shows the plots of the reference trajectory as a dashed line and the trajectory depicted by HyperCube as a solid line. The trajectory is a circular path with a radius of 50 mm in the XY plane with ZLED = 300 mm. The distance between reference and estimated trajectories was also plotted. The histogram of the error gives a mean value μ = 1.6 mm and a standard deviation σ = 2.8 mm. We noted that the noise at 1 kHz for the demodulation was lower than the noise at 3.5 kHz and 11.5 kHz, as shown in Figure 13B,C. The higher levels of noise at 3.5 kHz and 11.5 kHz than the one at 1 kHz are due to the use of different operational amplifiers (TI OPA4322 and TIOPA4342) for the analog demodulation. Figure 13A-C highlight that the HyperCube sensor provides useful measurements to accurately estimate the dynamic relative position of an IR LED in the XY plane.

To reduce the measurement noise, the digital signal outputs given by the HyperCube sensor were filtered by a Kalman filter. The system is described in a state space form as follows:

x k + 1 = F x k + v k ,
y k = C x k + w k with k
where the state xk = [Xkk]T describes the position Xk and the velocity k at the time instant tk. wk and vk are the state and measurement white noise with known and constant covariance matrices W and V, respectively. They are supposed to be mutually uncorrelated. F = [ 1 T e 0 1 ] is the state transition matrix, C = [1 0] the observation matrix. In practice, we choose the covariance matrices as follows V = σ meas 2 [ 1 0 0 1 ] and W = σ proc 2 [ 1 0 0 1 ] where σmeas and σproc are the standard deviations of the measurement noise and the state noise, respectively.

A linear Kalman filter was also implemented for the state estimation along the Y direction.

5.1. Experimental Height Estimation

As described in Section 4.2, the height between the HyperCube sensor and a pair of IR LEDs was estimated. Thanks to the angular measurements using three flickering IR LEDs, one can reduce the measurement noise. The height estimation precision is therefore increased. The mean of the height estimate Ẑ̅ is written as:

Z ^ ¯ = Z ^ LED 1 , 2 + Z ^ LED 2 , 3 + Z ^ LED 1 , 3 3
where LED1,2, LED2,3 and LED1,3 are the height estimated by the pairs (1,2), (2,3) and (1,3), respectively. The height of the IR LEDS was modified, while the position of the HyperCube sensor was fixed in the XY plane. Using each pair of IR LEDs, the height estimation was computed on-line. From Figure 14, the height estimation as a solid line is precise around 300 mm with respect to the reference height as a dashed line.

In Figure 15, for different pairs of IR LEDs, we plotted the evolution of the height estimation while the sensor was moving relatively to the center of the pair of IR LEDs. We note from Figure 15 that the height estimation depends on the position (XLED, YLED) of the sensor. The estimation error can vary up to 50 mm.

The estimation remains better with the pair of IR LEDs (1 kHz and 3.5 kHz) than with the others.

5.2. IR LED Position Tracking in 2D

In this section, HyperCube measurements, which consist of azimuth φ and elevation ψ, feed the controller that aims at tracking a moving IR LED in the plane. A Smith predictor with proportional and integral actions was designed and implemented to ensure the good stability of the linear displacement of HyperCube despite the time lag inherent to the visual processing algorithm. Figure 16A shows the block diagram of the closed loop system with the assessed angles φ and ψ. In Figure 16B,C, estimated positions versus time are plotted. The tracking system gives the reference trajectory.

As shown in Figure 16B,C, the position tracking was accurate. The tracking error plotted in the Figure 16D shows that the error ranges within ±5 cm. Nevertheless, one can note in Figure 16B that there is an error between the estimated and the reference values due to a saturation of the motor of the X axis. Thanks to a fast motor for controlling the Y axis, HyperCube was able to follow faithfully the moving pattern. In Figure 16C, the HyperCube sensor perfectly tracks the moving IR LED along the Y axis.

6. Conclusions

We designed and presented the HyperCube sensor. It is a cubic, miniature and low-cost sensing device able to track active markers without any optics. The system discussed has the following key features:

  • It provides azimuth and elevation and is able to assess the angular position of three blinking IR LEDs at specific frequencies: 1 kHz, 3.5 kHz and 11.5 kHz.

  • The estimation is achieved by a computationally-inexpensive signal processing of the low-amplitude and high-frequency signal transmitted by the three photodiodes placed on each side of the cube.

  • Experimental results showed that for the position estimation and the position tracking in 2D, this miniature device is very efficient. It also features promising results in terms of distance estimation, as well.

Nevertheless, the prototype developed reveals some limitations with respect to the range of pose estimation, but it is well adapted for proximal localization (up to 30 cm). In addition, the size of the demodulation electronic board could be easily reduced. As a conclusion, the HyperCube sensor is a promising sensing device for proximal localization applications and could equip UAVs or MAVs. The short-term outlook consists of improving the range of sensitivity of HyperCube using silicon photodiodes, which are specifically designed for precision photometry and feature a high sensitivity, a high-speed response and low noise.

Acknowledgements

We thank M. Boyron for his assistance with the electrical design, B. Boisseau for his involvement in the experimentation and J. Diperi for his involvement in the mechanical design of the sensor. We also thank N. Franceschini for the motor-driven XY table.

Author Contributions

Thibaut Raharijaona, Raphaël Juston and Stéphane Viollet designed the research. Paul Mignon, Lubin Kerhuel, Raphaël Juston and Thibaut Raharijaona performed the research. Thibaut Raharijaona and Stéphane Viollet wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Breitenmoser, A.; Kneip, L.; Siegwart, R. A monocular vision-based system for 6D relative robot localization. Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), San Francisco, CA, USA, 25–30 September 2011; pp. 79–85.
  2. Audette, R.; Balthazaar, J.; Dunk, C.; Zelek, J. A Stereo-Vision System for the Visually Impaired; University of Guelph: Guelph, ON, Canada, 2000. [Google Scholar]
  3. Cardin, S.; Vexo, F.; Thalmann, D. Wearable system for mobility improvement of visually impaired people. Vis. Comp. J. 2006, 23, 109–118. [Google Scholar]
  4. Ijaz, F.; Yang, H.K.; Ahmad, A.; Lee, C. Indoor positioning: A review of indoor ultrasonic positioning systems. Proceedings of the 15th International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 27–30 Janunary 2013; pp. 1146–1150.
  5. Aloui, N.; Raoof, K.; Bouallegue, A.; Letourneur, S.; Zaibi, S. Performance evaluation of an acoustic indoor localization system based on a fingerprinting technique. EURASIP J. Adv. Signal Process. 2014, 2014. [Google Scholar] [CrossRef]
  6. Dobrzynski, M.; Pericet-Camara, R.; Floreano, D. Vision tape—A flexible compound camera for motion detection and proximity estimation. IEEE Sens. J. 2012, 12, 1131–1139. [Google Scholar]
  7. Floreano, D.; Pericet-Camara, R.; Viollet, S.; Ruffier, F.; Brückner, A.; Leitel, R.; Buss, W.; Menouni, M.; Expert, F.; Juston, R.; et al. Miniature curved artificial compound eyes. Proc. Natl. Acad. Sci. USA 2013, 110, 9267–9272. [Google Scholar]
  8. Censi, A.; Strubel, J.; Brandli, C.; Delbruck, T.; Scaramuzza, D. Low-latency localization by active LED markers tracking using a dynamic vision sensor. Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, 3–7 November 2013; pp. 891–898.
  9. Lichtsteiner, P.; Posch, C.; Delbruck, T. A 128× 128 120 dB 15 μ s latency asynchronous temporal contrast vision sensor. IEEE J. Sol. State Circuits 2008, 43, 566–576. [Google Scholar]
  10. Faessler, M.; Mueggler, E.; Schwabe, K.; Scaramuzza, D. A monocular pose estimation system based on Infrared LEDs. Proceedinds of the IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014.
  11. Roberts, J.; Stirling, T.; Zufferey, J.C.; Floreano, D. 3-D relative positioning sensor for indoor flying robots. Auton. Robots 2012, 33, 5–20. [Google Scholar]
  12. Wenzel, K.; Masselli, A.; Zell, A. Visual tracking and following of a quadrocopter by another quadrocopter. 4993–4998.
  13. Etter, W.; Martin, P.; Mangharam, R. Cooperative flight guidance of autonomous unmanned aerial vehicles. Proceedings of the 2nd Internation Workshop on Networks of Cooperating Objects, Chicago, IL, USA, 11–12 April 2011.
  14. Masselli, A.; Zell, A. A novel marker based tracking method for position and attitude control of MAVs. Proceedings of International Micro Air Vehicle Conference and Flight Competition (IMAV), Braunschweig, Germany, 3–6 July 2012.
  15. Westheimer, G. Visual hyperacuity, Progress in Sensory Physiology; Springer: Berlin, Germany, 1981; pp. 1–30. [Google Scholar]
  16. Kerhuel, L.; Viollet, S.; Franceschini, N. The VODKA sensor: A bio-inspired hyperacute optical position sensing device. IEEE Sens. J. 2012, 12, 315–324. [Google Scholar]
  17. Juston, R.; Kerhuel, L.; Franceschini, N.; Viollet, S. Hyperacute edge and bar detection in a bioinspired optical position sensing device. IEEE/ASME Trans. Mechatron. 2013, 19, 1025–1034. [Google Scholar]
  18. Götz, K. Optomotorische Untersuchung des visuellen Systems einiger Augenmutanten der Fruchtfliege. Drosophila. Biol. Cybern. 1964, 2, 77–92. [Google Scholar]
  19. Stavenga, D. Angular and spectral sensitivity of fly photoreceptors. II. dependence on facet lens F-number and rhabdomere type in Drosophila. J. Comp. Phys. A 2003, 189, 189–202. [Google Scholar]
Figure 1. Sketch of a collaborative flight using IR LEDs and the HyperCube sensor embedded onboard the follower.
Figure 1. Sketch of a collaborative flight using IR LEDs and the HyperCube sensor embedded onboard the follower.
Sensors 15 16484f1 1024
Figure 2. (A) Each side of the HyperCube sensor integrates one photosensor and an analog amplifier for the conversion of the photodiode current into an output voltage; (B) one side of the HyperCube sensor with one photosensor (tiny surface-mounted device (SMD) photodiode); (C) electrical schematic drawing of the photoconductor circuit.
Figure 2. (A) Each side of the HyperCube sensor integrates one photosensor and an analog amplifier for the conversion of the photodiode current into an output voltage; (B) one side of the HyperCube sensor with one photosensor (tiny surface-mounted device (SMD) photodiode); (C) electrical schematic drawing of the photoconductor circuit.
Sensors 15 16484f2 1024
Figure 3. Cosine-like angular sensitivity of a photosensor.
Figure 3. Cosine-like angular sensitivity of a photosensor.
Sensors 15 16484f3 1024
Figure 4. Custom-made electronic board for the frequency modulation of the IR LEDs. It produces three separate signals that flicker at 1 kHz, 3.5 kHz and 11.5 kHz, sent to the three IR LEDs of the object to be located.
Figure 4. Custom-made electronic board for the frequency modulation of the IR LEDs. It produces three separate signals that flicker at 1 kHz, 3.5 kHz and 11.5 kHz, sent to the three IR LEDs of the object to be located.
Sensors 15 16484f4 1024
Figure 5. Custom-made acquisition and demodulation board. It is composed of three analog demodulation circuits. A microcontroller processes the visual output signal yielded by HyperCube and provides the azimuth and the elevation, which are sent through a USB interface to a PC.
Figure 5. Custom-made acquisition and demodulation board. It is composed of three analog demodulation circuits. A microcontroller processes the visual output signal yielded by HyperCube and provides the azimuth and the elevation, which are sent through a USB interface to a PC.
Sensors 15 16484f5 1024
Figure 6. Comparison between the angular sensitivity of a photosensor embedded on HyperCube (red solid line) and the cosine-like angular sensitivity of the model (blue dashed dot line) given by Equation (2) and the Gaussian angular sensitivity (dashed green line) given by Equation (1) with the acceptance angle Δρ = 120°.
Figure 6. Comparison between the angular sensitivity of a photosensor embedded on HyperCube (red solid line) and the cosine-like angular sensitivity of the model (blue dashed dot line) given by Equation (2) and the Gaussian angular sensitivity (dashed green line) given by Equation (1) with the acceptance angle Δρ = 120°.
Sensors 15 16484f6 1024
Figure 7. Experimental recording of the dynamic response of the demodulated output signal SPh1 at 1 kHz. The settling time at 95% t95% is equal to 7.25 ms, and the rise time tm (the time required for the response to rise from 5% to 95% of its final value) is equal to 1.6 ms.
Figure 7. Experimental recording of the dynamic response of the demodulated output signal SPh1 at 1 kHz. The settling time at 95% t95% is equal to 7.25 ms, and the rise time tm (the time required for the response to rise from 5% to 95% of its final value) is equal to 1.6 ms.
Sensors 15 16484f7 1024
Figure 8. (A) Front view, which gives an illustration of the azimuth φ and the reference plane Πφ; (B) front view, which gives an illustration of the elevation angle ψ and the reference plane Πψ; (C) top view with the reference plane Πφ, which is the plane through the IR LED and the optical axes of the photosensor Ph1 and the photosensor Ph2; (D) side view with the reference plane Πψ, which is the plane through the IR LED and the optical axes of the virtual photosensor (Ph1 + Ph2) and the photoreceptor Ph3.
Figure 8. (A) Front view, which gives an illustration of the azimuth φ and the reference plane Πφ; (B) front view, which gives an illustration of the elevation angle ψ and the reference plane Πψ; (C) top view with the reference plane Πφ, which is the plane through the IR LED and the optical axes of the photosensor Ph1 and the photosensor Ph2; (D) side view with the reference plane Πψ, which is the plane through the IR LED and the optical axes of the virtual photosensor (Ph1 + Ph2) and the photoreceptor Ph3.
Sensors 15 16484f8 1024
Figure 9. Sketch diagram of the HyperCube signal processing algorithm. (A) Top view: the sensor measures the azimuth φ. The left part shows the IR LED modulated at a frequency noted fi (1 kHz, 3.5 kHz or 11.5 kHz). In this view, HyperCube is composed of two photosensors Ph1 and Ph2 with their respective cosine-like angular sensitivities corresponding to Sides A and C of the sensor (see Figure 2A). An analog band-pass filter acts as a demodulator to extract the signal corresponding to the frequency fi of the IR LED, and an analog low-pass filter section with the cut-off frequency of 100 Hz reduces the high-frequency noise and prevents the subsequent analog-to-digital conversion from any aliasing effects. The digital processing consists of computing for each frequency fi the ratio of the relative difference to the sum between the two signals ( S P h 1 S P h 2 S P h 1 + S P h 2 ) to yield the HyperCube sensor output signal; (B) Side view: the same signal processing is applied on the signal provided by the photosensor Ph3 (Side B in Figure 2A) of HyperCube and a virtual photosensor, which is the sum of the photosensors Ph1 and Ph2.
Figure 9. Sketch diagram of the HyperCube signal processing algorithm. (A) Top view: the sensor measures the azimuth φ. The left part shows the IR LED modulated at a frequency noted fi (1 kHz, 3.5 kHz or 11.5 kHz). In this view, HyperCube is composed of two photosensors Ph1 and Ph2 with their respective cosine-like angular sensitivities corresponding to Sides A and C of the sensor (see Figure 2A). An analog band-pass filter acts as a demodulator to extract the signal corresponding to the frequency fi of the IR LED, and an analog low-pass filter section with the cut-off frequency of 100 Hz reduces the high-frequency noise and prevents the subsequent analog-to-digital conversion from any aliasing effects. The digital processing consists of computing for each frequency fi the ratio of the relative difference to the sum between the two signals ( S P h 1 S P h 2 S P h 1 + S P h 2 ) to yield the HyperCube sensor output signal; (B) Side view: the same signal processing is applied on the signal provided by the photosensor Ph3 (Side B in Figure 2A) of HyperCube and a virtual photosensor, which is the sum of the photosensors Ph1 and Ph2.
Sensors 15 16484f9 1024
Figure 10. XY table driven by DC motors. The HyperCube position is controlled in closed-loop by means of two incremental encoders. The distance ZLED was assumed here to be known and constant. Additional active motion markers were used to compare the precision of HyperCube with the ground truth acquired from the VICON motion capture system.
Figure 10. XY table driven by DC motors. The HyperCube position is controlled in closed-loop by means of two incremental encoders. The distance ZLED was assumed here to be known and constant. Additional active motion markers were used to compare the precision of HyperCube with the ground truth acquired from the VICON motion capture system.
Sensors 15 16484f10 1024
Figure 11. XLED and YLED are measured by an incremental encoder for each direction and are plotted as a function of Sφ and Sψ. Linear regressions used to determine aφ, aψ, bφ and bψ are also plotted. We note that the curves are linear on the range displayed.
Figure 11. XLED and YLED are measured by an incremental encoder for each direction and are plotted as a function of Sφ and Sψ. Linear regressions used to determine aφ, aψ, bφ and bψ are also plotted. We note that the curves are linear on the range displayed.
Sensors 15 16484f11 1024
Figure 12. Height estimation description while HyperCube assesses the azimuthal angles φA and φB thanks to the signal output Sφ.
Figure 12. Height estimation description while HyperCube assesses the azimuthal angles φA and φB thanks to the signal output Sφ.
Sensors 15 16484f12 1024
Figure 13. (A) Position estimation for a circular trajectory with a radius of 50 mm in the XY plane. Note that the absolute value of the position error is lower than 1 cm and that the standard deviation is about 3 cm; We also note that the noise is more important for the experiments at 3.5 kHz (B) and 11.5 kHz (C).
Figure 13. (A) Position estimation for a circular trajectory with a radius of 50 mm in the XY plane. Note that the absolute value of the position error is lower than 1 cm and that the standard deviation is about 3 cm; We also note that the noise is more important for the experiments at 3.5 kHz (B) and 11.5 kHz (C).
Sensors 15 16484f13 1024
Figure 14. Mean of the height estimate for three pairs of IR LEDs (see Equation (18)) while the position of the HyperCube sensor is fixed. Note that the height estimation is more accurate at 300 mm, since this was the height of calibration.
Figure 14. Mean of the height estimate for three pairs of IR LEDs (see Equation (18)) while the position of the HyperCube sensor is fixed. Note that the height estimation is more accurate at 300 mm, since this was the height of calibration.
Sensors 15 16484f14 1024
Figure 15. Height estimation for three pairs of LEDs and with a 10-cm shift in the X and Y directions. Note that a small coupling appears on the height estimation with respect to the position of HyperCube in the XY plane.
Figure 15. Height estimation for three pairs of LEDs and with a 10-cm shift in the X and Y directions. Note that a small coupling appears on the height estimation with respect to the position of HyperCube in the XY plane.
Sensors 15 16484f15 1024
Figure 16. (A) Block diagram of the IR LED position tracking system characterized by moving by hand one IR LED (flickering frequency of 1 kHz) placed 30 cm above HyperCube; (B) experimental recording of the position tracking along the X direction versus time. The DC motor dynamics in charge of moving HyperCube along the X axis was a bit too slow to follow faithfully the quick variations of the reference input signal; (C) experimental recording of the position tracking along the Y direction versus time; (D) error of the IR LED position tracking. Note that the absolute value of the position error never exceeds 10 cm.
Figure 16. (A) Block diagram of the IR LED position tracking system characterized by moving by hand one IR LED (flickering frequency of 1 kHz) placed 30 cm above HyperCube; (B) experimental recording of the position tracking along the X direction versus time. The DC motor dynamics in charge of moving HyperCube along the X axis was a bit too slow to follow faithfully the quick variations of the reference input signal; (C) experimental recording of the position tracking along the Y direction versus time; (D) error of the IR LED position tracking. Note that the absolute value of the position error never exceeds 10 cm.
Sensors 15 16484f16 1024

Share and Cite

MDPI and ACS Style

Raharijaona, T.; Mignon, P.; Juston, R.; Kerhuel, L.; Viollet, S. HyperCube: A Small Lensless Position Sensing Device for the Tracking of Flickering Infrared LEDs. Sensors 2015, 15, 16484-16502. https://doi.org/10.3390/s150716484

AMA Style

Raharijaona T, Mignon P, Juston R, Kerhuel L, Viollet S. HyperCube: A Small Lensless Position Sensing Device for the Tracking of Flickering Infrared LEDs. Sensors. 2015; 15(7):16484-16502. https://doi.org/10.3390/s150716484

Chicago/Turabian Style

Raharijaona, Thibaut, Paul Mignon, Raphaël Juston, Lubin Kerhuel, and Stéphane Viollet. 2015. "HyperCube: A Small Lensless Position Sensing Device for the Tracking of Flickering Infrared LEDs" Sensors 15, no. 7: 16484-16502. https://doi.org/10.3390/s150716484

Article Metrics

Back to TopTop