Next Article in Journal
Mo-Based Layered Nanostructures for the Electrochemical Sensing of Biomolecules
Previous Article in Journal
Underground Coal Mine Fingerprint Positioning Based on the MA-VAP Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Photometric Stereo-Based Depth Map Reconstruction for Monocular Capsule Endoscopy

1
Intelligent Robotics Institute, School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
2
Beijing Advanced Innovation Center for Intelligent Robots and Systems, Beijing Institute of Technology, Beijing 100081, China
3
Key Laboratory of Biomimetic Robots and Systems, Beijing Institute of Technology, Ministry of Education, Beijing 100081, China
4
The BioRobotics Institute, Scuola Superiore Sant’Anna, 56025 Pisa, Italy
5
Department of Excellence in Robotics & AI, Scuola Superiore Sant’Anna, 56025 Pisa, Italy
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5403; https://doi.org/10.3390/s20185403
Submission received: 25 July 2020 / Revised: 15 September 2020 / Accepted: 16 September 2020 / Published: 21 September 2020
(This article belongs to the Section Biomedical Sensors)

Abstract

:
The capsule endoscopy robot can only use monocular vision due to the dimensional limit. To improve the depth perception of the monocular capsule endoscopy robot, this paper proposes a photometric stereo-based depth map reconstruction method. First, based on the characteristics of the capsule endoscopy robot system, a photometric stereo framework is established. Then, by combining the specular property and Lambertian property of the object surface, the depth of the specular highlight point is estimated, and the depth map of the whole object surface is reconstructed by a forward upwind scheme. To evaluate the precision of the depth estimation of the specular highlight region and the depth map reconstruction of the object surface, simulations and experiments are implemented with synthetic images and pig colon tissue, respectively. The results of the simulations and experiments show that the proposed method provides good precision for depth map reconstruction in monocular capsule endoscopy.

1. Introduction

The capsule endoscopy robot is a novel endoscopic device that can implement non-invasive digestive tract inspection. Limited by the dimension of the digestive tract, capsule endoscopy robots can only use monocular vision, which lacks depth perception. Therefore, depth map reconstruction is an important research topic for monocular capsule endoscopy robots [1,2,3,4].
Photometric stereo is a promising technology for depth map reconstruction in capsule endoscopy, since it only needs monocular vision and several light sources. For a Lambertian object surface, the relationship between surface depth and illuminance can be described by the photometric irradiance equation set based on the Lambertian reflection property [5], and the depth map can be reconstructed when the scale factor is known [6].
However, to apply photometric stereo in monocular capsule endoscopy, there are two key problems to be solved. First, the object surface for endoscopy, i.e., the digestive tract, is not an ideal Lambertian surface. As shown in Figure 1, specular highlights always exist in the captured images; thus, photometric stereo cannot be applied directly for such object surfaces. Second, to reconstruct the depth map of the whole object surface with photometric stereo, the ground truth of the depth of at least one object surface point is needed to serve as a boundary condition.
To solve these problems, in this paper, we study the reflection property of the specular highlight on a non-ideal Lambertian surface and propose a method that estimates the depth of the specular highlight point and reconstructs the depth map of the digestive tract surface by photometric stereo.
The main contributions of this paper are: (1) we combine the specular property and Lambertian property of the object surface to estimate the depth of the specular highlight point, and (2) the estimated depth of the specular highlight point can serve as the boundary condition of the subsequent depth map reconstruction; therefore, no extra depth measurement devices are needed for the capsule endoscopy robot.
In Section 2, we introduce the state-of-the-art depth map reconstruction solutions for monocular capsule endoscopy. In Section 3, we propose a near point light source photometric stereo framework for the capsule endoscopy robot system. In Section 4, we present a depth map reconstruction method, which is based on the specular reflection constraint and photometric stereo. In Section 5 and Section 6, the proposed method is evaluated by simulation and experiment, respectively. The conclusion of this paper is given in Section 7.

2. Related Work

2.1. Depth Map Reconstruction in Capsule Endoscopy

Traditional methods for vision-based depth map reconstruction employed in minimally invasive surgery require the use of stereo cameras [7]. While present on modern stereo laparoscopes and robotic platforms, they are not found in capsule endoscopy systems due to physical constraints. Monocular 3D reconstruction techniques such as MonoSLAM [8] rely on completely static target structures, visual features that can be efficiently tracked, and a fast frame rate generating an abundance of viewpoints that can be used for structure recovery; a set of requirements that is not suitable for a typical capsule endoscopy setting. As a novel computer vision technology, machine learning has also been employed to explore the endoscopic and surgical applications [9,10,11]. However, training samples of endoscopy images need to be labeled with professional knowledge, and the measurement of the ground truth is also time consuming, which makes this route difficult to achieve currently.
Another vision-based 3D reconstruction method that has seen early applications to endoscopy and minimally invasive surgery is Shape-From-Shading (SFS) [12], where the 3D structure of the visualized scene is reconstructed from a single monocular camera without any further information required. Nevertheless, the reconstructed structures are in a metric space, i.e., they are linearly scaled with an unknown factor related to the surface albedo. Furthermore, the SFS problem is not a well-posed mathematical problem, which means it needs more restrictions to determine a unique solution for 3D reconstruction.
Photometric stereo is a monocular vision technology developed from SFS. By illuminating the object with multiple light sources from different directions, multiple irradiance equations can be obtained, and the well-posedness of the problem can be improved [13]. To employ photometric stereo in a capsule endoscopy application, an important property of the object surface needs to be noticed: the digestive tract is not an ideal Lambertian surface; covered by digestive juice, it shows specular reflection when the incident angle of illumination is close to the reflected angle. Therefore, to avoid the influence of the specular highlight in photometric stereo, researchers have explored different strategies.

2.2. Photometric Stereo-Based Depth Map Reconstruction of a Non-Ideal Lambertian Surface

The most straightforward strategy to avoid the specular highlight is to eliminate the specular highlight by adding external optical filters. From a physical point of view, visible light is a kind of electromagnetic wave with the polarization property. According to the Fresnel reflectance model, when the light is reflected, the specular component shows obvious directionality [14]. Based on this theory, Parot et al. added a pair of polarizers assembled in front of the light source and camera to reduce the specular reflection highlight [15]. However, the external polarization device is not good for minimizing the size ofcapsule endoscopy robots.
Another frequently-used strategy for the specular highlight is by separating the specular component of reflection by image processing. References [16,17] tried to eliminate the specular component in the pre-processing step. Dichromatic reflectance models have also been considered for diffuse and specular reflection separation. Reference [18] employed a dark channel prior, while [19] iteratively compared the intensity logarithmic differentiation of the input image. Yang and Abuja [20] used both the diffuse and specular components for the reconstruction based on the assumption that the illumination chromaticity is known and at least one of the input images is free of specularity. Nevertheless, these image processing-based methods are not robust enough when the amount of input images is not sufficient.
Except for eliminating or separating the specular component of reflection, some researchers also attempted to integrate the Lambertian and specular component into one general model. Ikehata et al. [21] used the purely diffuse irradiance equation for general surfaces, considered the specular component as a sparse error, and introduced a regression procedure, which requires tens of images to improve accuracy. However, the expensive algorithms based on energy minimization tend to be slow and cannot provide real-time shape reconstruction, and the light attenuation during propagation has not been considered. The Blinn–Phong model is an empirical light reflectance model that integrates diffuse reflection and specular reflection [22], and it is also studied for the photometric stereo problem of non-ideal Lambertian surfaces. Reference [23] combined the Blinn–Phong reflectance model and perspective projection to solve the 3D reconstruction problem; nevertheless, the infinite point light source assumption limits its usage for more general applications. Reference [24] integrated the Blinn–Phong reflectance model with the close point light source model; by performing iteration from one ground truth value of the surface depth, the surface depth map can be reconstructed. However, to reconstruct the whole object surface, extra devices are needed to measure the ground truth depth of the object surface as a boundary condition.
In this paper, we present a photometric stereo-based depth map reconstruction solution for a monocular capsule endoscopy robot, which makes use of the geometric property of specular highlight. The specular property and Lambertian property of the object surface are employed to estimate the depth of the specular highlight point, which can serve as the boundary condition of the subsequent depth map reconstruction.

3. Photometric Stereo Framework for the Capsule Endoscopy Robot

To ensure the precision of depth map reconstruction, an appropriate photometric stereo framework needs to be established based on the characteristics of the capsule endoscopy robot system. In this section, we first outline the basic structure of the capsule endoscopy robot system. Then, we propose a near point light source photometric stereo framework based on the characteristics of the capsule endoscopy robot.

3.1. Capsule Endoscopy Robot System

Magnetic force can transmit without a medium, and it therefore has become the most frequently used driving force for capsule endoscopy robots [25,26,27]. A typical magnetically-driven capsule endoscopy robot system developed by the authors is shown in Figure 2. An Internal Permanent Magnet (IPM) is integrated in the capsule endoscopy robot as the locomotion unit, and an External Permanent Magnet (EPM) is attached at the end of a manipulator as a locomotion driver. By operating the EPM with the manipulator, the pose of the capsule endoscopy robot can be modified by magnetic force and torque.
The capsule endoscopy robot is designed to have an outer diameter of 14 mm and a length of 30 mm to ensure its maneuverability in the digestive tract of human beings [28]. The vision unit of the capsule robot is comprised of a monocular endoscopic camera, and it is located at the center of the capsule robot’s front side. The endoscopic camera has an ultra-short focal length, 640 × 480 pixels of resolution, and a 70.6 degree field of view and works with a depth of field from 3 mm to 100 mm. The illumination unit of the capsule endoscopy robot is comprised of four Prolight PS2P-TFPE FMR9 (3 mm × 2 mm × 0.8 mm, 0.2 W powered) surface-mounted LEDs, which are assembled around the camera with a 5.5 mm centrifugal distance. The four LEDs are driven by a TMS320F28335 DSP controller, and the illuminance of each light source can be modified independently by changing the duty cycle of the Pulse Width Modulation (PWM) signal.

3.2. Photometric Stereo Framework

For the capsule endoscopy robot system described in the previous part, photometric stereo is a promising solution for the depth map reconstruction of the object surface. Therefore, we propose a photometric stereo framework based on an endoscopic camera and surface-mounted LED light sources.

3.2.1. Near Point Light Source Model

In order to provide the necessary ingredients of the photometric stereo framework, we first employ a pinhole camera model to describe the monocular endoscopic camera [5]. As shown in Figure 3a, the optical center of the camera is located at point C, and the Pixel Reference Frame (PRF) O x y z is defined at the plane ξ = f . For a given point M on the object surface Σ , it can be described by an unknown function z x , y as:
M x , y = ξ , η , ζ = x z x , y f , y z x , y f , z x , y ,
where f is the focal length of the camera and the triple ξ , η , ζ = ξ x , y , η x , y , ζ x , y comprises the coordinate in the Camera Reference Frame (CRF) C ξ η ζ .
Since the surface-mounted LEDs are close to the object surface, for different points on the object surface, the incident directions of illumination cannot be considered as parallel. Therefore, we employ the near point light source model to describe the surface-mounted LED light sources of the capsule endoscopy robot [12,29]. As shown in Figure 3b, when point M is illuminated by the surface-mounted LED source S i with position ξ i , η i , ζ i (where i 1 , 4 ), the direction of incident light can be given by vector l i x , y , z , and the image of point M is projected at x , y , f on the Image Reference Frame (IRF).
The illumination direction vector l i x , y , z can be given by the ratio of direction vector l ¯ i x , y , z and its module q i x , y , z as:
l i x , y , z = ξ i f z + x , η i f z + y , ζ i f z f ξ i f z + x 2 + η i f z + y 2 + ζ i f z f 2 = l ¯ i x , y , z q i x , y , z ,
where:
l ¯ i x , y , z = ξ i f z + x , η i f z + y , ζ i f z f ,
q i x , y , z = ξ i f z + x 2 + η i f z + y 2 + ζ i f z f 2 .

3.2.2. Light Source Attenuation

The illuminance of the surface-mounted LED light source is inhomogeneous and anisotropic; therefore, the attenuation of illuminance needs to be considered in the photometric stereo framework. The two main sources of illuminance attenuation are the propagation distance and the radial angle with respect to the principle direction [5].
  • Attenuation due to propagation distance:
    For a point light source, the energy absorbed by a given area on the object surface attenuates with respect to the inverse square of propagation distance. For light source S i and object point M, the illuminance attenuation factor a i d x , y , z due to propagation distance can be described as:
    a i d x , y , z = 1 MS i 2 = 1 ξ i + x z f 2 + η i + y z f 2 + ζ i z 2 = f 2 z 2 q i 2 x , y , z .
  • Radial attenuation:
    The surface-mounted LED is a directional light source and has its maximal illuminance along a certain direction, which is known as the principle direction. When deviating from the principle direction, the illuminance of the surface-mounted LED usually attenuates with a cosine tendency with respect to the principle direction, as shown in Figure 4.
    Therefore, for light source S i and object point M, if we assume that the principle direction is along the z direction, the radial attenuation factor a i r x , y , z can be described as:
    a i r x , y , z = c o s θ i x , y , z = l i x , y , z · 0 , 0 , 1 = f q i x , y , z .
By combining the aforementioned two kinds of attenuations into a general illuminance attenuation factor, we get:
a i x , y , z = a i d x , y , z · a i r x , y , z = f 3 z 2 q i 3 x , y , z

4. Photometric Stereo-Based Depth Map Reconstruction

In this section, we introduce the depth map reconstruction method for the capsule endoscopy robot based on the photometric stereo framework proposed in the previous section. First, we analyze the property of the object surface of the capsule endoscope and introduce the reflection model of the non-ideal Lambertian surface. Then, we propose a method that combines the specular property and Lambertian property of reflection to estimate the depth of the specular highlight point. Finally, we introduce the depth map reconstruction method based on the depth estimation of the specular highlight point.

4.1. Reflection Model of the Non-Ideal Lambertian Surface

The object surface of the capsule endoscope is generally the inner wall of the digestive tract of human beings, and it is always covered by a film of digestive juice; thus, it has a two-fold reflection property: it can show both the specular property and Lambertian property, depending on whether the incident angle is close to the reflected angle or not.
As shown in Figure 5a, from a microscopic viewpoint, the surface of the digestive tract is not flat. On the one hand, the concave parts of the surface can be covered by liquid more easily and thus show significant specular reflection when illuminated by incident light. The convex parts, on the other hand, show more Lambertian reflection because of the roughness.
Therefore, from a macroscopic viewpoint, because the concave parts and convex parts of the liquid covered surface are distributed evenly, we can unify the specular reflectance and Lambertian reflectance on the object surface into a general reflection model, as shown in Figure 5b, in which the concave parts contribute the specular reflection component along a certain direction, while the convex parts contribute the Lambertian reflection component along all outward directions. Note that, though our model is similar to the Blinn–Phong reflection model, the difference is that our model discusses the geometric property of the specular component, i.e., the incident angle equals the reflected angle, which is different from the photometric property of the specular component discussed in the Blinn–Phong model.
To parameterize the aforementioned reflection model, we define a scenario in which the specular reflection and Lambertian reflection are created by different light sources, respectively, as shown in Figure 6.
For point M on the object surface Σ , the normalized outgoing normal vector n x , y can be given by:
n x , y = n ¯ x , y n ¯ x , y ,
where:
n ¯ x , y = z f 2 f z x , y , z x , y + x , y · z x , y .
For the ease of notation, here, we assume that among the four point light sources of the photometric stereo framework, the light source S σ creates the specular highlight at point M, while the other three light sources S δ 1 to S δ 3 create Lambertian reflection at point M. Then, we derive the two key equations in this paper, i.e., the specular reflection equation and the photometric irradiance equation set, from the specular reflection constraint and Lambertian reflection property.
  • Specular reflection equation:
    For light source S σ , because the incident angle equals the reflected angle, we get the specular reflection equation as:
    MC MC + l ¯ σ x , y , z q σ x , y , z = λ n ¯ x , y ,
    where:
    MC = x z f , y z f , z
    and λ is a scale factor. Then, by substituting the terms n ¯ x , y and MC in the specular reflection equation with the expressions in Equations (9) and (11), the specular reflection equation can be further notated as:
    x , y , f x 2 + y 2 + f 2 + l ¯ σ x , y , z q σ x , y , z = λ z f 2 f z x , f z y , z + x z x + y z y .
  • Photometric irradiance equation set:
    For light source S δ 1 to S δ 3 , because of the Lambertian reflection, the photometric irradiance equation set can be obtained as:
    I δ 1 x , y = a δ 1 x , y , z ρ x , y l δ 1 x , y , z · n x , y I δ 2 x , y = a δ 2 x , y , z ρ x , y l δ 2 x , y , z · n x , y I δ 3 x , y = a δ 3 x , y , z ρ x , y l δ 3 x , y , z · n x , y
    where a δ x , y , z is the attenuation factor and ρ x , y is the albedo of the object surface. By substituting the terms l δ x , y , z , n x , y and the attenuation factor a δ x , y , z with the expressions in Equations (2), (7), and (8), the photometric irradiance equation set Equation (13) can be notated as:
    I δ 1 x , y = f 3 ρ x , y z 2 n ¯ x , y · l ¯ δ 1 x , y , z · n ¯ x , y q δ 1 4 x , y , z I δ 2 x , y = f 3 ρ x , y z 2 n ¯ x , y · l ¯ δ 2 x , y , z · n ¯ x , y q δ 2 4 x , y , z I δ 3 x , y = f 3 ρ x , y z 2 n ¯ x , y · l ¯ δ 3 x , y , z · n ¯ x , y q δ 3 4 x , y , z .

4.2. Depth Estimation of Specular Highlight Point

To estimate the depth of specular highlight point z, we construct a Partial Derivative Equation (PDE) about z from the photometric irradiance equation set, Equation (13), and eliminate its partial derivative terms z x and z y by combining it with the specular reflection Equation (11).
First, for the photometric irradiance equation set, we eliminate the term f 3 ρ x , y z 2 n ¯ x , y by making ratios between different photometric irradiance equations, and an image ratio equation can be obtained as:
I δ 1 x , y I δ 2 x , y = q δ 2 4 x , y , z q δ 1 4 x , y , z · l ¯ δ 1 x , y , z · n ¯ x , y l ¯ δ 2 x , y , z · n ¯ x , y .
For the ease of notation, we denote the terms q i x , y , z and I i x , y as q i and I i in the following expressions, respectively. Considering that:
l ¯ δ x , y , z · n ¯ x , y = z f 2 · ξ δ f z + x , η δ f z + y , ζ δ f z f · f z x , y , z x , y + x , y · z x , y = ξ δ + ζ δ f x z x + η δ + ζ δ f y z y + ζ δ z z f ,
a PDE about depth z can be notated as:
F ξ , δ 1 , 2 x , y , z · z x + F η , δ 1 , 2 x , y , z · z y = z · F ζ , δ 1 , 2 x , y , z ,
where:
F ξ , δ 1 , 2 x , y , z = I δ 1 q δ 1 4 ξ δ 2 f + ζ δ 2 x I δ 2 q δ 2 4 ξ δ 1 f + ζ δ 1 x F η , δ 1 , 2 x , y , z = I δ 1 q δ 1 4 η δ 2 f + ζ δ 2 y I δ 2 q δ 2 4 η δ 1 f + ζ δ 1 y F ζ , δ 1 , 2 x , y , z = I δ 2 q δ 2 4 ζ δ 1 z I δ 1 q δ 1 4 ζ δ 2 z .
Then, to eliminate the partial derivative terms in Equation (17), we derive the relationship between depth z and partial derivatives z x and z y from the specular reflection Equation (12). By substituting l ¯ σ x , y , z with the expression in Equation (3), the scale factor λ can be eliminated, and after some algebra, two expressions about z x and z y can be obtained as:
z x = x z q σ x , y , z + x z + ξ σ f b x , y b x , y q σ x , y , z + b 2 x , y + f x ξ σ + y η σ f ζ σ z
and:
z y = y z q σ x , y , z + y z + η σ f b x , y b x , y q σ x , y , z + b 2 x , y + f x ξ σ + y η σ f ζ σ z ,
where:
b x , y = x 2 + y 2 + f 2 .
Finally, by substituting the partial derivative terms z x and z y in Equation (17) with the expressions in Equations (19) and (20), we get:
F ξ , δ 1 , 2 x , y , z · G ξ , σ x , y , z + F η , δ 1 , 2 x , y , z · G η , σ x , y , z = F ζ , δ 1 , 2 x , y , z · G ζ , σ x , y , z
where:
G ξ , σ x , y , z = x z q σ + x z + ξ σ f b x , y G η , σ x , y , z = y z q σ + y z + η σ f b x , y G ζ , σ x , y , z = z q σ b 2 x , y + z b 3 x , y + f x ξ σ + y η σ f ζ σ b x , y .
Since the only unknown of Equation (22) is the depth z of the specular highlight point, we move all terms to the same side of the equation and obtain an energy function as:
E x , y , z = F ξ , δ 1 , 2 x , y , z · G ξ , σ x , y , z + F η , δ 1 , 2 x , y , z · G η , σ x , y , z + F ζ , δ 1 , 2 x , y , z · G ζ , σ x , y , z .
By solving the minimization problem of this energy function, the depth z of the specular highlight point can be estimated.

4.3. Depth Map Reconstruction of the Object Surface

To reconstruct the depth map of the whole object surface, we consider solving the Partial Differential Equation (PDE) (17). Since the PDE (17) has three unknowns z, z x , and z y , the unique solution cannot be obtained with only one Dirichlet boundary condition. Therefore, to improve the well-posedness of this problem, we extend it to an equation set based on three image ratio equations, then an iteration scheme can be obtained [5]. By using the estimated depth of the specular highlight point as the boundary condition (seed point) of the iteration scheme, the depth map of the whole object surface can be then reconstructed.
In order to simplify the notation, we denote Equation (17) as:
F δ 1 , δ 2 x , y , z · z = s δ 1 , δ 2 x , y , z ,
where:
F δ 1 , δ 2 x , y , z = F ξ , δ 1 , 2 x , y , z , F η , δ 1 , 2 x , y , z
and:
s δ 1 , δ 2 x , y , z = z · F ζ , δ 1 , 2 x , y , z .
In the discretization of the PDEs, we define the position of point M to be described by discrete spatial coordinate i , j as x i , y j , Z j , i , and the step length along the x and y directions as Δ x and Δ y , respectively, then the forward upwind scheme can be obtained as:
F ξ , δ 1 , 2 i , j Z i + 1 , j Z i 1 , j 2 Δ x + F η , δ 1 , 2 i , j Z i , j + 1 Z i , j 1 2 Δ y = F ξ , δ 1 , 2 i , j Δ x 2 Z i + 1 , j 2 Z i , j + Z i 1 , j Δ x 2 + F η , δ 1 , 2 i , j Δ y 2 Z i , j + 1 2 Z i , j + Z i , j 1 Δ y 2 + s i , j .
The artificial diffusion introduced in the right-hand side of Equation (28) allows us to follow the vector field F by considering the most appropriate discretization for the first derivative in order to track the characteristic lines [30,31]. In particular, it is a consistent numerical scheme of order equal to one with respect to both Δ x and Δ y . After some algebra, Equation (28) can be simplified as:
F ξ , δ 1 , 2 i , j Z i s g n F ξ , δ 1 , 2 i , j , j + F η , δ 1 , 2 i , j Z i , j s g n F η , δ 1 , 2 i , j = Z i , j F ξ , δ 1 , 2 i , j + F η , δ 1 , 2 i , j Δ s δ 1 , δ 2 i , j ,
where Δ x = Δ y = Δ , and we emphasize again the dependence of the functions F and s on z. Using a fast-marching procedure, we couple the pairs of Equation (29) from different light source pairs in order to compute the directional derivative according to eight principal directions (2 horizontals, 2 verticals, and 4 diagonals), which span the two-dimensional image domain [32]. To reduce the error, we chose the specific equation by taking the pair of images that have the highest gray scale value at pixel x i , y j .

5. Simulation

The proposed depth estimation method for the specular highlight point in the previous section is based on the assumption that the incident angle equals the reflected angle. Nevertheless, in practical scenarios, the specular highlight is not always displayed as a certain point with the ideal specular property, but as a highlight region in which the pixels show the specular property in different degrees, i.e., the incident angle may not exactly equal the reflected angle. For this reason, the depth estimation errors of pixels in the specular highlight region are different from each other, and this can further influence the precision of the subsequent depth map reconstruction of the whole object surface. Therefore, to evaluate the effectiveness of the proposed method in a more practical scenario, we define a photometric stereo framework of monocular capsule endoscopy in a simulation environment and conduct simulation tests.

5.1. Simulation Configuration

A general view of our simulation configuration is given in Figure 7a. Here, we introduce the details of configuration in three parts: photometric stereo framework, object surface, and optical environment.

5.1.1. Photometric Stereo Framework

As shown in Figure 7b, the photometric stereo framework consists of a camera and four surface-mounted LEDs.
The camera is defined as an ideal pinhole camera, i.e., there is no geometric, chromatic, or photometric aberration during the imaging procedure. The focal length of the camera is defined as 565 pixels, the same as that in real experiments.
The four surface-mounted LED light sources are defined as near point light sources with two sources of illuminance attenuation, and they are placed surrounding the camera with the same distance d to the camera. To simplify the calculation, we assume that the four light sources and the camera are in the same plane, and the illuminance of the four light sources is the same. The principle directions of the light sources are all along the principle direction of the camera, and the distance d from the camera to each light source is defined as 5.5 mm. To obtain color images, we define the light sources in simulation to emit white light.
For the image processing configuration, we assume that all images created in the simulation environment are well exposed, and all image post-processing procedures, e.g., gamma correction, contrast, saturation, brightness, etc., are all closed; therefore, the pixel intensity in the generated images is proportional to the illuminance of the light sources.

5.1.2. Object Surface

Digestive tract polyps are the most typical objects for capsule endoscopy screening. Therefore, in our simulation environment, we define a polyp-shaped object surface to evaluate the precision of the proposed method. The shape and dimension of the object surface are shown in Figure 7c. The surface with a convex profile prevents from shadow and self-blocking, and the diameter of the spherical part is defined with reference to the real digestive tract polyps of human beings.
To simulate the reflectance of a liquid-covered rough surface in a practical scenario, we use the Blinn–Phong model to describe the reflectance of the object surface [22]. The factor for glossiness in the Blinn–Phong model is elaborately selected; therefore, the distribution of the specular highlight region is close to that in real experiments. The color of the object surface is defined as a real human digestive tract.
Since the digestive tract of human beings has homogeneous reflectance, we define that the reflectance of the object surface is also homogeneous. To optimize the focus of the endoscopic camera, we place the object surface aligned with the photometric stereo framework, and we define the distance from the camera to the object surface as 21.37 mm in the simulation and the subsequent real experiments.

5.1.3. Optical Environment

To simulate the dark environment in the digestive tract of human beings, we define the photometric stereo framework and the object surface to be isolated from external lighting. All illuminations in the simulation environment only come from the surface-mounted LEDs of the photometric stereo framework, and they can only be reflected by the object surface.
We further define that the secondary reflection does not occur on the object surface, and the propagation medium of light in the simulation is transparent and homogeneous.

5.2. Simulation Procedures and Results

5.2.1. Image Generation and Specular Highlight Region Detection

Based on the aforementioned simulation configuration, four photometric stereo images were generated by successively enabling the four light sources, as shown in the first row of Figure 8. The illuminance maps of the generated photometric stereo images are then obtained, as shown in the second row of Figure 8.
Considering that the specular highlight regions have low saturation and high intensity [2], we set the thresholds of saturation and intensity for detecting specular highlight in the four images, and the results of the detection are shown in the third row of Figure 8.

5.2.2. Depth Estimation of Specular Highlight Points

For all pixels in the detected specular highlight regions, we conducted the proposed depth estimation method. Since the four generated images are symmetric to each other, here we only discuss the result of Figure 8a. The depth estimation error of each specular pixel is obtained by comparing the estimation value and the ground truth, and the depth estimation error distribution of the specular highlight region is shown in Figure 9. Note that the relative depth errors are obtained by calculating the ratios of the absolute values of depth estimation errors to the ground truth depth of each specular highlight pixel.
According to the obtained result, the pixel at the centroid of the specular highlight region has the minimal depth estimation error 0.0488 mm and relative depth error 0.2808%, which come from the estimated depth −17.3700 mm and ground truth −17.4188 mm. The Root Mean Squared Error (RMSE) of the depth estimation of all pixels in the specular highlight region is 0.9355 mm, and the corresponding relative depth error is 5.3857%. Obviously, the depth estimation of the centroid pixel of the specular highlight region is the most suitable boundary condition for the subsequent depth map reconstruction, since it has the best estimation of depth among all points in the specular highlight region.

5.2.3. Depth Map Reconstruction of the Object Surface

To evaluate how the depth estimation error of the specular highlight point influences the precision of subsequent depth map reconstruction, we use the depth estimation results of the specular highlight pixels in Section 5.2.2 as the boundary conditions of the forward upwind scheme in Section 4.3. The depth map reconstruction tests were implemented based on specular highlight pixels with minimal depth estimation error and the RMSE of depth estimation. The results of the depth map reconstruction tests are shown in Figure 10 and Table 1.
For the centroid of the specular highlight region, i.e., the specular highlight pixel with minimal depth estimation error, the reconstructed depth map of the object surface almost has the same shape of the ground truth; the RMSE of the depth reconstruction of the object surface is less than 0.1 mm; and the relative RMSE of the depth reconstruction of the object surface is less than 0.5%, which is relatively precise for the capsule endoscopy application.
For the specular highlight pixels with the RMSE of depth estimation, the RMSEs of the reconstructed depth map of the object surface are less than 2 mm, and the relative RMSEs of the object surface depth reconstruction are less than 10%, which are also acceptable for the capsule endoscopy application.

6. Experiment

To further evaluate the effectiveness of the proposed method for the practical monocular capsule endoscopy robot, we conducted a depth map reconstruction experiment with pig colon tissue.

6.1. Experiment Devices

To guarantee the dimensional precision of the experiment, we established a test bench for depth map reconstruction, as shown in Figure 11a.
The monocular vision unit and illumination unit of the capsule endoscopy robot are fixed on a 2D linear stage with a 3D printed holder, as shown in Figure 11b. The pig colon tissue is fixed on a 3D printed base board to prevent from deformation, as shown in Figure 11c. For comparison with the simulation tests in Section 4, the profile of the 3D printed base board is designed with the same shape and dimensions as those in Figure 7c.

6.2. Experiment Configuration

6.2.1. Photometric Stereo Framework Calibration

Different from the ideal models in the simulation, the camera and light sources in the practical photometric stereo framework need to be calibrated before experiments.
  • Camera calibration:
    For the practical endoscopic camera, different kinds of aberrations exist in the imaging system, and they can be classified into two main categories: geometric aberration and photometric aberration.
    The geometric aberration of the camera is caused by the inherent optical property and assembling error of the spherical lens. To compensate the geometric aberration, the internal parameters of the camera must be calibrated first. By calibration with a chessboard pattern [33], the intrinsic parameters of the endoscopic camera are obtained as shown in Table 2. The average reprojection error of the calibration is 0.157224 pixels.
    The photometric aberration of the camera is caused by the post-processing procedures of the image sensor. Procedures such as gamma correction, modification of saturation, contrast, white balance, etc., can make the pixel intensity in the image not proportional to the incoming illuminance. To obtain the precise illuminance map of each image, we calibrate the non-linear relationship between the illuminance on the image sensor and the pixel intensity in the captured images.
  • Light source calibration:
    The surface-mounted LED light sources need to be calibrated, as their illuminance and assembling are not perfect.
    With respect to illuminance, though all LED light sources in the photometric stereo framework are powered by the same voltage, the illuminance of each LED may still be different since the diode voltage drop of each individual LED is different. Therefore, it is necessary to calibrate the illuminance difference between the LED light sources.
    With respect to assembling, errors in the position and principle direction of the LED light sources are inevitable. Therefore, to obtain more precise positions and principle directions of the light sources, we employ the light source position calibration method and principle direction calibration method in [34,35] to calibrate the light sources.

6.2.2. Optical and Imaging Preparation

  • External light shield:
    To avoid the disturbance of external light, all experiments were conducted in a black box.
  • Exposure
    To ensure that the captured images were well exposed, we closed the auto exposure function of the camera and matched the illuminance of the light sources with the camera’s exposure setting by modifying the duty cycle of the PWM signal.
  • Image post-processing:
    Since the photometric-based method is very sensitive to the noise of illuminance, the Gaussian filter and medium filter are used in image post-processing to decrease the white noise and salt-and-pepper noise.

6.3. Experiment Procedure and Results

6.3.1. Image Capture and Specular Highlight Region Detection

Based on the experimental configuration described in Section 6.2, we captured four photometric stereo images under the illumination of four different light sources, as shown in the first row of Figure 12. Then, the illuminance maps of the photometric stereo images were obtained by using the camera calibration parameters, as shown in the second row of Figure 12. After that, the specular highlight regions in the illuminance maps were detected by means of saturation and intensity, as shown in the third row of Figure 12.

6.3.2. Depth Estimation of the Specular Highlight Points

For all pixels in the detected specular highlight regions, we implemented the proposed depth estimation method. For comparison with the simulation, here, we only discuss the results of the depth estimation of pixels in the specular highlight region of Figure 12a. The depth estimation error distribution is shown in Figure 13.
According to the obtained results, the pixel at the centroid of the specular highlight region has the minimal depth estimation error 0.2830 mm and relative depth error 1.6272%, which comes from the estimated depth −17.1100 mm and ground truth −17.3930 mm. The RMSE of the depth estimation of all pixels in the specular highlight region is 3.1892 mm, and the corresponding relative depth error is 18.3361%.
From a quantitative viewpoint, the depth estimation errors of the specular highlight region in the experiments and simulation show a similar distribution, but for pixels that are not at the center of the specular highlight region, the depth estimation errors are much more significant and not suitable for the subsequent depth map reconstruction. To meet the need of the practical application, the points with minor depth estimation need to be chosen from the specular highlight region. According to our results in the simulation and experiment, the centroid of the specular highlight region can be a good candidate.

6.3.3. Depth Map Reconstruction of the Object Surface

In the subsequent depth map reconstruction test, we used the depth estimation result of the specular highlight points in Section 6.3.2 as the boundary condition of the forward upwind scheme in Section 4.3. The centroid pixel of the specular highlight region, i.e., the specular highlight point with minimal depth estimation error, was used as the seed point of depth map reconstruction. The experiment result is compared with the ground truth depth map and the simulation results in Figure 14 and Table 3.
The reconstructed depth map of the object surface in the experiment has a similar shape as the ground truth depth map; the RMSE of the depth reconstruction of the object surface is less than 0.6 mm, and the relative RMSE of the depth reconstruction of the object surface is less than 3%. For general screenings in capsule endoscopy, this precision of the depth map reconstruction can be a good accessory means for polyp diagnosis.

7. Conclusions

This paper proposes a photometric stereo-based depth map reconstruction method for monocular capsule endoscopy. By combining the specular property and Lambertian property under different illumination conditions, the depth of the specular highlight point is estimated, and the depth map of the whole object surface is reconstructed by a forward upwind scheme based on photometric image ratio equations. The results of the simulations and experiments show that the proposed method can provide sufficient precision in the depth estimation of specular highlight points and the depth map reconstruction of the object surface. When the centroid of the specular highlight region is used as the seed point of the forward upwind scheme, the precision of depth map reconstruction can meet the need of the practical application.
For future works, more challenging problems will be considered both in theoretical and experimental terms. For theoretical modeling, a more precise reflection model of a non-ideal Lambertian surface will be considered to improve the depth estimation of the specular highlight. For experiments, more comparable experiments will be verified not only by the camera, but also by MRI and CT devices. For object surfaces with more practical reflection properties, more complex phantom structures [36,37] and experimental techniques [38,39] will also be considered in experiments.

Author Contributions

Conceptualization, methodology, and writing, original draft preparation, Y.H.; validation, P.Z.; writing, review and editing, J.L., F.M., and G.C.; supervision, P.D. and Q.H. All authors read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Key R&D Program of China with Grant Number 2019YFB1312400.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fan, Y.; Meng, M.Q.H.; Li, B. 3D reconstruction of wireless capsule endoscopy images. In Proceedings of the 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, Buenos Aires, Argentina, 1–4 September 2010; pp. 5149–5152. [Google Scholar]
  2. Ciuti, G.; Visentini-Scarzanella, M.; Dore, A.; Menciassi, A.; Dario, P.; Yang, G.Z. Intra-operative monocular 3D reconstruction for image-guided navigation in active locomotion capsule endoscopy. In Proceedings of the 2012 4th IEEE RAS & EMBS International Conference on Biomedical Robotics and Biomechatronics (BioRob), Rome, Italy, 24–27 June 2012; pp. 768–774. [Google Scholar]
  3. Prasath, V.S.; Figueiredo, I.N.; Figueiredo, P.N.; Palaniappan, K. Mucosal region detection and 3D reconstruction in wireless capsule endoscopy videos using active contours. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 4014–4017. [Google Scholar]
  4. Karargyris, A.; Karargyris, O.; Bourbakis, N. 3D representation of the digestive tract surface in Wireless Capsule Endoscopy videos. In Proceedings of the 2010 IEEE International Conference on BioInformatics and BioEngineering, Philadelphia, PA, USA, 31 May–3 June 2010; pp. 279–280. [Google Scholar]
  5. Mecca, R.; Wetzler, A.; Bruckstein, A.M.; Kimmel, R. Near field photometric stereo with point light sources. SIAM J. Imaging Sci. 2014, 7, 2732–2770. [Google Scholar] [CrossRef]
  6. Agrawal, A.; Raskar, R.; Chellappa, R. What is the range of surface reconstructions from a gradient field? In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 578–591. [Google Scholar]
  7. Stoyanov, D.; Scarzanella, M.V.; Pratt, P.; Yang, G.Z. Real-time stereo reconstruction in robotically assisted minimally invasive surgery. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Beijing, China, 20–24 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 275–282. [Google Scholar]
  8. Newcombe, R.A.; Davison, A.J. Live dense reconstruction with a single moving camera. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1498–1505. [Google Scholar]
  9. Su, H.; Yang, C.; Ferrigno, G.; De Momi, E. Improved human—Robot collaborative control of redundant robot for teleoperated minimally invasive surgery. IEEE Robot. Autom. Lett. 2019, 4, 1447–1453. [Google Scholar] [CrossRef] [Green Version]
  10. Su, H.; Qi, W.; Yang, C.; Sandoval, J.; Ferrigno, G.; De Momi, E. Deep neural network approach in robot tool dynamics identification for bilateral teleoperation. IEEE Robot. Autom. Lett. 2020, 5, 2943–2949. [Google Scholar] [CrossRef]
  11. Qi, W.; Su, H.; Aliverti, A. A smartphone-based adaptive recognition and real-time monitoring system for human activities. IEEE Trans. Hum. Mach. Syst. 2020, 5, 414–423. [Google Scholar] [CrossRef]
  12. Okatani, T.; Deguchi, K. Shape reconstruction from an endoscope image by shape from shading technique for a point light source at the projection center. Comput. Vis. Image Underst. 1997, 66, 119–131. [Google Scholar] [CrossRef]
  13. Woodham, R.J. Photometric stereo: A reflectance map technique for determining surface orientation from image intensity. In Image Understanding Systems and Industrial Applications I; International Society for Optics and Photonics: Bellingham, WA, USA, 1979; Volume 155, pp. 136–143. [Google Scholar]
  14. Wolff, L.B.; Boult, T.E. Constraining object features using a polarization reflectance model. Phys. Based Vis. Princ. Pract. Radiom. 1993, 1, 167. [Google Scholar] [CrossRef]
  15. Parot, V.; Lim, D.; González, G.; Traverso, G.; Nishioka, N.S.; Vakoc, B.J.; Durr, N.J. Photometric stereo endoscopy. J. Biomed. Opt. 2013, 18, 076017. [Google Scholar] [CrossRef]
  16. Barsky, S.; Petrou, M. The 4-source photometric stereo technique for three-dimensional surfaces in the presence of highlights and shadows. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1239–1252. [Google Scholar] [CrossRef] [Green Version]
  17. Yang, Q.; Wang, S.; Ahuja, N. Real-time specular highlight removal using bilateral filtering. In Proceedings of the European Conference on Computer Vision, Crete, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 87–100. [Google Scholar]
  18. Kim, H.; Jin, H.; Hadap, S.; Kweon, I. Specular reflection separation using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1460–1467. [Google Scholar]
  19. Tan, R.T.; Ikeuchi, K. Separating reflection components of textured surfaces using a single image. In Digitally Archiving Cultural Objects; Springer: Berlin/Heidelberg, Germany, 2008; pp. 353–384. [Google Scholar]
  20. Yang, Q.; Ahuja, N. Surface reflectance and normal estimation from photometric stereo. Comput. Vis. Image Underst. 2012, 116, 793–802. [Google Scholar] [CrossRef]
  21. Ikehata, S.; Aizawa, K. Photometric stereo using constrained bivariate regression for general isotropic surfaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2179–2186. [Google Scholar]
  22. Blinn, J.F. Models of light reflection for computer synthesized pictures. In Proceedings of the 4th Annual Conference on Computer Graphics and Interactive Techniques, San Jose, CA, USA, 20–22 July 1977; pp. 192–198. [Google Scholar]
  23. Khanian, M.; Boroujerdi, A.S.; Breuß, M. Photometric stereo for strong specular highlights. Comput. Vis. Media 2018, 4, 83–102. [Google Scholar] [CrossRef] [Green Version]
  24. Mecca, R.; Rodolà, E.; Cremers, D. Realistic photometric stereo using partial differential irradiance equation ratios. Comput. Graph. 2015, 51, 8–16. [Google Scholar] [CrossRef]
  25. Arezzo, A.; Menciassi, A.; Valdastri, P.; Ciuti, G.; Lucarini, G.; Salerno, M.; Di Natali, C.; Verra, M.; Dario, P.; Morino, M. Experimental assessment of a novel robotically-driven endoscopic capsule compared to traditional colonoscopy. Dig. Liver Dis. 2013, 45, 657–662. [Google Scholar] [CrossRef] [PubMed]
  26. Slawinski, P.R.; Obstein, K.L.; Valdastri, P. Capsule endoscopy of the future: What’s on the horizon? World J. Gastroenterol. WJG 2015, 21, 10528. [Google Scholar] [CrossRef] [PubMed]
  27. Li, J.; Barjuei, E.S.; Ciuti, G.; Hao, Y.; Zhang, P.; Menciassi, A.; Huang, Q.; Dario, P. Magnetically-driven medical robots: An analytical magnetic model for endoscopic capsules design. J. Magn. Magn. Mater. 2018, 452, 278–287. [Google Scholar] [CrossRef]
  28. Sliker, L.J.; Ciuti, G. Flexible and capsule endoscopy for screening, diagnosis and treatment. Expert Rev. Med. Devices 2014, 11, 649–666. [Google Scholar] [CrossRef]
  29. Collins, T.; Bartoli, A. 3d reconstruction in laparoscopy with close-range photometric stereo. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 634–642. [Google Scholar]
  30. Quarteroni, A.; Valli, A. Numerical Approximation of Partial Differential Equations; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008; Volume 23. [Google Scholar]
  31. Strikwerda, J.C. Finite Difference Schemes and Partial Differential Equations; SIAM: Philadelphia, PA, USA, 2004; Volume 88. [Google Scholar]
  32. Mecca, R.; Tankus, A.; Wetzler, A.; Bruckstein, A.M. A direct differential approach to photometric stereo with perspective viewing. SIAM J. Imaging Sci. 2014, 7, 579–612. [Google Scholar] [CrossRef]
  33. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef] [Green Version]
  34. Hao, Y.; Visentini-Scarzanella, M.; Li, J.; Zhang, P.; Ciuti, G.; Dario, P.; Huang, Q. Light source position calibration method for photometric stereo in capsule endoscopy. In Advanced Robotics; Taylor & Francis: Abingdon, UK, 2020; pp. 1–13. [Google Scholar]
  35. Visentini-Scarzanella, M.; Kawasaki, H. Simultaneous camera, light position and radiant intensity distribution calibration. In Image and Video Technology; Springer: Berlin/Heidelberg, Germany, 2015; pp. 557–571. [Google Scholar]
  36. Kłodowski, K.; Krzyżak, A.T. Innovative anisotropic phantoms for calibration of diffusion tensor imaging sequences. Magn. Reson. Imaging 2016, 34, 404–409. [Google Scholar] [CrossRef]
  37. Fieremans, E.; Lee, H.H. Physical and numerical phantoms for the validation of brain microstructural MRI: A cookbook. Neuroimage 2018, 182, 39–61. [Google Scholar] [CrossRef]
  38. Krzyżak, A.; Jasiński, A.; Kwieciński, S.; Kozłowski, P.; Adamek, D. Quantitative assessment of injury in rat spinal cords in vivo by MRI of water diffusion tensor. Appl. Magn. Reson. 2008, 34, 3. [Google Scholar] [CrossRef]
  39. Sulaiman, S.; Marciani, L. MRI of the Colon in the Pharmaceutical Field: The Future before us. Pharmaceutics 2019, 11, 146. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Digestive tract surface showing the specular highlight when illuminated by a light source from a certain direction.
Figure 1. Digestive tract surface showing the specular highlight when illuminated by a light source from a certain direction.
Sensors 20 05403 g001
Figure 2. Magnetically-driven capsule endoscopy robot system. IPM, Internal Permanent Magnet; EPM, External Permanent Magnet.
Figure 2. Magnetically-driven capsule endoscopy robot system. IPM, Internal Permanent Magnet; EPM, External Permanent Magnet.
Sensors 20 05403 g002
Figure 3. Photometric stereo framework. (a) Pinhole camera model of the endoscopic camera. (b) Near point light source model of the surface-mounted LED.
Figure 3. Photometric stereo framework. (a) Pinhole camera model of the endoscopic camera. (b) Near point light source model of the surface-mounted LED.
Sensors 20 05403 g003
Figure 4. Radial attenuation of the surface-mounted LED with respect to the principle direction.
Figure 4. Radial attenuation of the surface-mounted LED with respect to the principle direction.
Sensors 20 05403 g004
Figure 5. Lambertian reflection and specular reflection on the digestive tract. (a) Reflection on the digestive tract surface observed from a microscopic viewpoint. (b) General reflection model of the non-ideal Lambertian surface.
Figure 5. Lambertian reflection and specular reflection on the digestive tract. (a) Reflection on the digestive tract surface observed from a microscopic viewpoint. (b) General reflection model of the non-ideal Lambertian surface.
Sensors 20 05403 g005
Figure 6. Lambertian reflection and specular reflection under illumination from different light sources. The light source S δ creates Lambertian reflection, while the light source S σ creates specular reflection.
Figure 6. Lambertian reflection and specular reflection under illumination from different light sources. The light source S δ creates Lambertian reflection, while the light source S σ creates specular reflection.
Sensors 20 05403 g006
Figure 7. Simulation configuration of photometric stereo in monocular capsule endoscopy. (a) General view of the simulation environment. (b) Camera and light sources’ distribution in the photometric stereo frame (where d is the distance from the camera to each light source). (c) Section view of the object surface (dimensions in millimeters).
Figure 7. Simulation configuration of photometric stereo in monocular capsule endoscopy. (a) General view of the simulation environment. (b) Camera and light sources’ distribution in the photometric stereo frame (where d is the distance from the camera to each light source). (c) Section view of the object surface (dimensions in millimeters).
Sensors 20 05403 g007
Figure 8. Photometric stereo images, illuminance maps, and detected specular highlight regions in the simulation environment. (ad): generated photometric stereo images in the simulation environment. (eh): illuminance maps of the photometric stereo images in the simulation environment. (il): detected specular highlight regions in the simulation environment.
Figure 8. Photometric stereo images, illuminance maps, and detected specular highlight regions in the simulation environment. (ad): generated photometric stereo images in the simulation environment. (eh): illuminance maps of the photometric stereo images in the simulation environment. (il): detected specular highlight regions in the simulation environment.
Sensors 20 05403 g008
Figure 9. Depth estimation error distribution of the specular highlight region. (a) Depth estimation error map (absolute value). (b) Relative depth error map (absolute value).
Figure 9. Depth estimation error distribution of the specular highlight region. (a) Depth estimation error map (absolute value). (b) Relative depth error map (absolute value).
Sensors 20 05403 g009
Figure 10. Results of the depth map reconstruction tests based on the specular highlight pixels with different depth estimation precisions in the simulation. (a) Ground truth of the object surface. (b) Depth map reconstruction results based on the specular highlight pixel with minimal depth estimation error. (c) Depth map reconstruction results based on the specular highlight pixel with positive RMS depth estimation error. (d) Depth map reconstruction results based on the specular highlight pixel with negative RMS depth estimation error.
Figure 10. Results of the depth map reconstruction tests based on the specular highlight pixels with different depth estimation precisions in the simulation. (a) Ground truth of the object surface. (b) Depth map reconstruction results based on the specular highlight pixel with minimal depth estimation error. (c) Depth map reconstruction results based on the specular highlight pixel with positive RMS depth estimation error. (d) Depth map reconstruction results based on the specular highlight pixel with negative RMS depth estimation error.
Sensors 20 05403 g010
Figure 11. Test bench for the experiment. (a) General view of the test bench. (b) Front view of the monocular capsule endoscope. (c) Pig colon tissue fixed on the 3D printed base board. (Considering the institutional and governmental regulations about the ethical use of animals, all pig colon tissues used in this study are bought from market, as food grade.)
Figure 11. Test bench for the experiment. (a) General view of the test bench. (b) Front view of the monocular capsule endoscope. (c) Pig colon tissue fixed on the 3D printed base board. (Considering the institutional and governmental regulations about the ethical use of animals, all pig colon tissues used in this study are bought from market, as food grade.)
Sensors 20 05403 g011
Figure 12. Photometric stereo images, illuminance maps, and detected specular highlight regions in the experiment. (ad): captured photometric stereo images (after distortion correction). (eh): illuminance maps of photometric stereo images in the experiment. (al): detected specular highlight regions in the experiment.
Figure 12. Photometric stereo images, illuminance maps, and detected specular highlight regions in the experiment. (ad): captured photometric stereo images (after distortion correction). (eh): illuminance maps of photometric stereo images in the experiment. (al): detected specular highlight regions in the experiment.
Sensors 20 05403 g012
Figure 13. Depth estimation error distribution of the specular highlight region. (a) Depth estimation error map (absolute value). (b) Relative depth error map (absolute value).
Figure 13. Depth estimation error distribution of the specular highlight region. (a) Depth estimation error map (absolute value). (b) Relative depth error map (absolute value).
Sensors 20 05403 g013
Figure 14. Comparison of the results in the simulation and experiment. (a) Ground truth depth map of the object surface. (b) Reconstructed depth map based on the centroid pixel of the specular highlight region in the simulation. (c) Reconstructed depth map based on centroid pixel of the specular highlight region in the experiment.
Figure 14. Comparison of the results in the simulation and experiment. (a) Ground truth depth map of the object surface. (b) Reconstructed depth map based on the centroid pixel of the specular highlight region in the simulation. (c) Reconstructed depth map based on centroid pixel of the specular highlight region in the experiment.
Sensors 20 05403 g014
Table 1. RMSEs of the depth map reconstruction of the object surface in the simulation.
Table 1. RMSEs of the depth map reconstruction of the object surface in the simulation.
Depth Estimation Error of Seed Point (mm)RMSE of Reconstructed Depth Map of Object Surface (mm)Relative RMSE of Reconstructed Depth Map of Object Surface
Minimal error0.04880.09220.4545%
RMSE of specular region, positive0.93551.64838.1280%
RMSE of specular region, negative−0.93551.78438.7982%
Table 2. Calibrated intrinsic parameters of the endoscopic camera.
Table 2. Calibrated intrinsic parameters of the endoscopic camera.
ComponentEstimation
f x 564.9467
f y 564.9467
x 0 322.4128
y 0 263.3339
s0.0
Table 3. Comparison between the simulation result and the experimental result.
Table 3. Comparison between the simulation result and the experimental result.
TypeDepth Estimation Error of Seed Point (mm)RMSE of Reconstructed Depth Map of Object Surface (mm)Relative RMSE of Reconstructed Depth Map of Object Surface
Simulation0.04880.09220.4545%
Experiment0.28300.51452.5730%

Share and Cite

MDPI and ACS Style

Hao, Y.; Li, J.; Meng, F.; Zhang, P.; Ciuti, G.; Dario, P.; Huang, Q. Photometric Stereo-Based Depth Map Reconstruction for Monocular Capsule Endoscopy. Sensors 2020, 20, 5403. https://doi.org/10.3390/s20185403

AMA Style

Hao Y, Li J, Meng F, Zhang P, Ciuti G, Dario P, Huang Q. Photometric Stereo-Based Depth Map Reconstruction for Monocular Capsule Endoscopy. Sensors. 2020; 20(18):5403. https://doi.org/10.3390/s20185403

Chicago/Turabian Style

Hao, Yang, Jing Li, Fei Meng, Peisen Zhang, Gastone Ciuti, Paolo Dario, and Qiang Huang. 2020. "Photometric Stereo-Based Depth Map Reconstruction for Monocular Capsule Endoscopy" Sensors 20, no. 18: 5403. https://doi.org/10.3390/s20185403

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop