^{1}

^{*}

^{2}

^{1}

^{2}

This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).

An alternative method for distance measurement is presented, based on a radiometric approach to the image formation process. The proposed methodology uses images from an infrared emitting diode (IRED) to estimate the distance between the camera and the IRED. Camera output grey-level intensities are a function of the accumulated image irradiance, which is also related by inverse distance square law to the distance between the camera and the IRED. Analyzing camera-IRED distance, magnitudes that affected image grey-level intensities, and therefore accumulated image irradiance, were integrated into a differential model which was calibrated and used for distance estimation over a 200 to 600 cm range. In a preliminary model, the camera and the emitter were aligned.

Distance estimation using vision sensors is an important aspect of robotics since many robot positioning algorithms use distance to calculate the robot's position as the basis for more complicated tasks.

Traditionally, distance measuring in robotics has been conducted by sonar (US) and infrared (IR) sensing. Several methods based on the line-of-sight (LOS) and echo/reflection models have also been used. The LOS model places the emitter and detector in different locations, and signals travel from emitter to detector. The Reflection model links emitter and detector physically (in the same place) and signals are reflected off an object or wall, following a round-trip-path.

In the Reflection model, the viability of IR as an accurate means of measuring distance depends on extensive prior knowledge of the surface (scattering, reflection, and absorption). [

Vision devices based on a geometrical model have been used for many positioning tasks. However, most vision positioning algorithms are based on geometrical imaging models, where 3D to 2D projection constitutes the main mathematical tool for analysis [

With geometrical models, a single camera and interest point can only estimate a 2D position, as projection based models cannot provide depth information. Nevertheless, depth can be calculated if additional information is included in the model. This entails the use of two vision sensors or some kind of active device [

To date, artificial vision has been one of most widely used positioning techniques since it gives accurate results. However, geometric modeling is more normally used in order to obtain distance to objects [

In intelligent spaces, smart living, etc., where a certain number of cameras are already installed, a simple method based on grey levels can be developed to determine the depth. The cameras already installed in the environment are used for performing other tasks necessary in smart living and intelligent spaces. For example, if a mobile robot carries an IRED, depth can be estimated from the cameras' pixel values, because there is a relationship between pixel grey-level intensity and the quantity of light that falls on the image sensor. Also, received light is related by inverse distance square law to the distance between the camera and the IRED.

When an image is captured by a digital camera, it provides a relative measure of the distribution of light within the scene. Thus, pixel grey-level intensities are a function of sensor surface irradiance accumulated during the exposition time.

The function that relates accumulated image irradiance to output grey-level intensities is known as the Radiometrical Camera Response Function (RCRF) [

In [

In most cases, an inverse-RCRF is required in order to obtain a direct relationship between pixel grey-level intensities and accumulated image irradiance by a Radiometric Camera Calibration Process (RCCP) [

Furthermore, if the camera takes an image of an IRED and this can be isolated, a point source model for the IRED can be used to estimate the irradiance on the surface of the camera lens using inverse distance square law [

Previous papers have been presented in the field of IR, in which the authors developed computer programs and methods. In [

The aim of the present study was to use a geometrical model for 2-D positioning, which would provide coordinates on a plane parallel to a mobile robot (for example), and then to use a radiometrical model in order to obtain the distance between the mobile robot and the plane. For our purposes, the LOS model was used in order to determine the distance between emitter (IR) and detector (CMOS camera).

The method using a single IR point to develop an effective and practical method is based on the assumption that the final application will be carried out in settings such as intelligent spaces, smart living spaces, etc., where a certain number of cameras are already installed. The cameras already installed in the environment are used for performing other tasks necessary in these smart living and intelligent spaces.

One possibility would be to use a photodiode. However, these devices provide an instantaneous response to the signals received, and given the distances involved in these applications (several meters), and working with an LOS link (the best alternative available), these can be less than pW. The low intensity of the received signal can, therefore, impede accurate, or even valid, measuring. Since distance estimation and robot position determination can involve measurements on a ms. timescale, considered as real time, the signal must be integrated into a determined time interval. The use of a photodiode would imply the need to design signal conditioning, integration and digital circuits, etc. All of these, in addition to a webcam, are already available in the proposed method; consequently, the design is simpler, implementation is quicker and final costs are lower, as a variety of existing models can be selected.

A further reason for using a camera is that by applying a differential method for measuring, and given that by using cameras the method can be selected digitally from a computer, automation and control, speed and safety of data acquisition is facilitated since two consecutive measurements are taken with different integration times.

Later, when the method for distance estimation using mathematical algorithms has been fully developed, it will be possible to combine this data with data obtained from the geometric calibration of cameras in order to improve generation of the variables and parameters involved in position and location of the device incorporating the IRED.

In order to define a model for the IRED-Camera set, the following aspects were established:

Estimation of accumulated image irradiance. When inverse RCRF is used, a measure of the accumulated image irradiance can be obtained. A sample of energy accumulated in the camera is shown in

The relationship between accumulated image irradiance and lens irradiance. This can be obtained using the camera's optical system model and also includes the camera exposition time. A linear behaviour is assumed for the camera's optical system model.

Behaviour of lens irradiance with the distance between the camera and the emitter. For the IRED Camera set a point source model can be used; in this case, lens irradiance can be estimated using inverse distance square law.

Image irradiance must be due to IRED light energy. Therefore, background illumination must be suppressed. An ideal implementation would be to test the algorithm in a dark room; however we used an interference filter to select a narrow wavelength band centered on the typical emitter wavelength in order to ensure that images were created only by IRED energy.

From a general point of view, accumulated image irradiance is a function of camera parameters and radiometric magnitudes which quantitatively affect the image formation process.

The fact that the emitter transmits up to 120°, or at a different angle, only influences the angle at which the detector can be situated in order to receive the emitter's signal, and does not affect the size of the image formed.

As regards image size, according to the laws of optical magnification this is only influenced by the size of the object and the distance at which it is located. In our case, as the diode size is both constant and very small, the image appears reduced. Nevertheless, as can be seen in

Reference [_{r}_{i}

To define the differential model, the magnitudes and relationships affecting _{r}

A differential method was selected because a measurement taken with a specific exposition time will include various errors due to camera irradiance, external illumination factors (spotlights, the sun, etc.), or the effect of temperature on the sensor, for example. The differential method enabled us to isolate the measurement from these effects.

Moreover, both the sensor and method are economic, since cameras are already installed in the application environment. Furthermore, the method is simple to launch and installation of the system is easy. The system is non-invasive and safe to operate, and the sensorial system is complementary to other methods, facilitating ease of data fusion

Assuming that the camera and the emitter are aligned initially, there are three magnitudes that affect the accumulated image irradiance for the camera-IRED set the camera exposition time, the IRED radiant intensity and the distance between the IRED and the camera [

In addition, and as in [_{r}_{r}_{r}

As in [_{r}_{1}, τ_{2}, δ_{1}, δ_{2}, ρ_{1}_{2}_{e}

From (_{j}_{1}, τ_{2}, δ_{1}, δ_{2}, ρ_{1}_{2}

If images captured with different camera exposition times are analyzed, then (_{n}_{r}_{r}

Once

The aim was to use the proposed differential model to estimate the distance between the camera and the IRED, where the model analyzes images of the IRED captured with different exposition times, and also assumes that distance and emitter radiant intensity are constant during the image capturing process.

The method described in this paper, based on differential image processing, presents several advantages, innovations and benefits related to previous studies, which can be summarized as follows:

the development of a sensor for distance measuring and pose detection based on grey level intensity of the images;

the development of a method for obtaining the distance between two points (IR-camera) using a differential;

the sensor and method are economic, since cameras are already in the application environment (requiring only one IRED for each object/mobile to be detected);

the method is simple to launch;

installation of the system is easy;

the system is non-invasive and safe;

the sensorial system is complementary to other methods, facilitating ease of data fusion; etc

For practical reasons, a Basler camera was used, with a SFH42XX High Powered IRED with 950 nm emission peak; thus an interference filter centered at 950 nm and with a bandwidth of 40 nm was added to the camera, which improved the signal/noise ratio since it eliminated background illumination: all visible light, and infrared light up to 930 nm and over 970 nm.

In order to use (_{i}_{r}_{n}

In _{4}

Two experiments were performed to test the validity of the differential model for distance estimation. In the first experiment, the 200 cm to 380 cm range was considered whilst the second considered the 400 cm to 600 cm range. The second experiment showed greater error since the distances were greater and were also beyond the range for which the sensor had been calibrated (that is, the range used in the first experiment). Nevertheless, the second experiment shows that even so, sufficiently accurate measurements can still be taken. For both experiments, distance was increased stepwise by 20 cm and the camera was aligned with the IRED.

In addition, to improve the efficiency of the methodology in practical applications, four images were analysed to estimate distance. The first image was captured with a _{r}_{1}_{2}_{3}_{e}2

Experiments were carried out using the method described above. The equipment used is indicated in Section 4. Once images from the optimal exposition time range had been selected, the deferential method for distance estimation was applied. The distance estimation results for each difference in exposition time, corresponding to the first experiment, are shown in

The final distance measurement is shown in

In the first distance range, errors in distance estimation using the differential model are less than 8 cm., representing a relative error of less than a 2.5%. The second experiment considered longer distances, and results are given in

In this case, the differential model was less accurate than in the first experiment.

An alternative method based on a radiometrical approach to the image formation process has been described for estimating the distance between a camera and an IRED. This method estimates the inverse RCRF in order to obtain a measure of the accumulated image irradiance, and shows that the accumulated image irradiance depends linearly on the emitter radiant intensity, the camera exposition time and the inverse square distance between the IRED and the camera. These behaviors are incorporated into a model, which can be re-written in a differential form.

The differential model has four parameters that must be estimated by means of a calibration process. Once the model's parameters have been calculated, the model's expression can be solved for distance estimation. Two distance ranges were considered for model validation. In the first range, errors were less than 8 cm. However, in the second experiment the errors were higher than in the first.

In conclusion, the proposed differential model represents an alternative method for estimating the distance between an IRED and a camera through analysis of image grey-level intensities.

This study was made possible thanks to the SILPAR II DPI2006-05835 projects sponsored by the DGI of the Spanish Ministry of Education at the University of Alcalá. We would also like to thank the Spanish Agency for International Development Cooperation (AECID), under the aegis of the Ministry of Foreign Affairs and Cooperation (MAEC).

Representative samples of images used in the emitter-to-camera distance characterization.

Representative results for the differential model calibration process. Three different emitter radiant intensities were considered by changing the diode polarization current.

Representative distance estimations and absolute error for Experiment 1.

Final distance estimation result for Experiment 1.

200 | 204.4 | 4.4 | 2.2 |

220 | 223.5 | 3.5 | 1.6 |

240 | 245.1 | 5.1 | 2.1 |

260 | 262.7 | 2.7 | 1.0 |

280 | 284 | 4.0 | 1.4 |

300 | 306 | 6.0 | 2.0 |

320 | 327.5 | 7.5 | 2.3 |

340 | 345.6 | 5.6 | 1.6 |

360 | 367.7 | 7.7 | 2.1 |

380 | 387.6 | 7.6 | 2.0 |

Final distance estimation result for Experiment 2.

400 | 412.2 | 12.2 | 3.1 |

420 | 425.6 | 5.6 | 1.3 |

440 | 452.6 | 12.6 | 2.9 |

460 | 487.5 | 27.5 | 6.0 |

480 | 495.1 | 15.1 | 3.1 |

500 | 508.3 | 8.3 | 1.7 |

520 | 543.1 | 23.1 | 4.4 |

540 | 567.5 | 27.5 | 5.1 |

560 | 568.0 | 8.0 | 1.4 |

580 | 595.1 | 15.1 | 2.6 |

600 | 615.7 | 15.7 | 2.6 |