You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

10 January 2023

Design and Implementation of a Video-Frame Localization System for a Drifting Camera-Based Sewer Inspection System

,
,
and
1
Graduate School of Integrated Science and Technology, Shizuoka University, Hamamatsu 432-8011, Japan
2
Department of Information Science, Aichi Institute of Technology, Toyota 470-0392, Japan
3
College of Engineering, Academic Institute, Shizuoka University, Hamamatsu 432-8011, Japan
*
Authors to whom correspondence should be addressed.
This article belongs to the Special Issue Internet of Things for Smart City Application

Abstract

To reduce the cost of inspecting old sewer pipes, we have been developing a low-cost sewer inspection system that uses drifting wireless cameras to record videos of the interior of a sewer pipe while drifting. The video’s data are transmitted to access points placed in utility holes and further transmitted to a video server where each video frame is linked to its capturing position so that users can identify the damaged areas. However, in small-diameter sewer pipes, locating drifting nodes over the full extent of the pipeline using Wi-Fi-based localization is difficult due to the limited reach of radio waves. In addition, there is the unavailability of a GNSS signal. We propose a function to link each video frame to a position based on linear interpolation using landmarks detected by the camera and image processing. Experiments for testing the accuracy of the localization in an underground sewer pipe showed that all utility holes were successfully detected as landmarks, and the maximum location estimation accuracy was less than 11.5% of the maximum interval of landmarks.

1. Introduction

1.1. Background: Sewer-Pipe Inspection with Drifting Wireless Cameras

Aged sewer pipes are susceptible to severe problems such as clogging by sand entering from cracks or by tree roots; the deterioration of sewer function; and road collapse. The year 2022 marks more than 30 years since the installation of 40% of a total of 490,000 km of sewer pipes in Japan [1,2]. Therefore, it is necessary to inspect aging sewer pipes regularly and repair or replace any damaged parts. However, current sewer-pipe-inspection methods, such as visual inspection [3], fiberscope inspection [4], self-propelled robot inspection [5], and boat-type camera inspection [6], are costly and/or time-consuming.
To shorten the time and reduce the monetary costs, we propose a sewer-pipe-inspection method using drifting wireless camera nodes (CNs) [7], as shown in Figure 1. When inspecting a sewer pipe with this system, an inspector firstly places one or more CN into the sewer pipe. Then, the CNs start recording videos of the interior of the pipe. When each of the CNs reaches the wireless communication range of one of the access points (APs) installed at some utility hole, it sends the recorded video data to the AP via wireless LAN. Finally, a server in the cloud aggregates the data from the APs via cellular networks. Then, users, i.e., sewer inspectors, access the server and browse the video to identify the positions of damage.
Figure 1. Sewer inspection system using drifting wireless camera nodes: An inspector places the drifting wireless camera nodes (CNs) into the pipe. Then, each CN starts drifting and recording videos inside the sewer pipe. Each CN sends video data to access points (APs) when it is within the wireless communication range of one of the APs. The AP receives the data and forwards the received data to the video server. Users can access the server and identify damaged points of the pipe via web browsing.
This system is safe because inspectors do not need to enter the pipes. In addition, unlike fiberscopes and self-propelled robot methods, the inspection range is not limited by the cables. Furthermore, unlike the boat-shaped camera technique, it is possible to watch the captured images before the camera drifts downstream through the sewer pipe. It is also possible to identify where the video frame was captured. These features reduce labor costs and are convenient.
This system uses wireless video data transmission from CNs in a sewer pipe. To identify the performance of the wireless communication in narrow ( ϕ 200–250 mm) sewer pipes, which are the most widely used public sewer pipes in Japan, we conducted an experiment. Moreover, we designed a wireless communication protocol for a narrow underground pipe environment. In [8], we evaluated the performance of the wireless communication according to IEEE 802.11 and IEEE 802.15.4 at 2.4/5 GHz and ARIB STD-T108 at 920 MHz in a 200–250 mm in diameter underground sewer pipe. The measurement results revealed that the maximum communication distance is approximately 10 m in the sewer pipe, whereas the interval between utility holes is typically 30 m or more. Ishihara et al. [7] developed a protocol to collect video data from multiple CNs to compensate for the short wireless communication range.

1.2. Motivation: Difficulty of Localization in Sewer Pipes

Even when video data can be reliably collected via wireless communication in the neighborhood of APs, identifying the positions of pipe damage, such as cracks and clogged tree roots from the captured video, is difficult because it is unclear where the video was taken. To detect the positions of pipe damages by looking at the aggregated video data, it is necessary to identify where each video frame was taken. However, GNSS cannot be used in sewer pipes, and location estimations of CNs using radio waves are also difficult because the range of the radio wave’s communication in a narrow pipe (diameter of 200–250 mm) is limited. Additionally, the CN’s weight should be light so that it can easily drift down a sewer pipe. Thus, the additional devices for the location estimation should be minimum. To tackle this problem, it is necessary to link each video frame to the position where it was captured with the minimum additional devices and without depending on radio communication.

1.3. Contributions

The contributions of this paper are summarized as follows. Firstly, we propose a method to estimate CN positions based on linear interpolation using utility holes and pipe joints as landmarks in a sewer pipe. We also implemented the proposed method together with a browsing system that links location information to each frame of the video and conducted experiments to evaluate the accuracy of the proposed localization function using a drifting CN in a real underground pipe.

1.4. Structure of the Paper

The remainder of this paper is structured as follows: we introduce related work on the self-position estimation of wireless mobile devices in Section 2. In Section 3, we describe a sewer-pipe-inspection method using drifting wireless cameras that we developed. In Section 4, we propose a method to link the location of a CN to a timestamp of a video frame. Section 5 describes the implementation of the proposed localization system, and Section 6 describes experiments for the verification of the system in an underground pipe. Section 7 summarizes this paper.

3. Sewer-Pipe-Inspection System with Drifting Wireless Cameras

In this section, we provide an overview of the sewer inspection system with drifting wireless CNs that we developed for sewer pipes with diameters of 200 to 250 mm, which are the most popular pipes in Japan. We inspect several hundred meters to several kilometers at a time. A CN and an AP communicate with each other via wireless LAN, and their communication range is about 5–10 m, which is much shorter than the distance between utility holes, which is typically 30 m or more.

3.1. Basic Design

Figure 1 shows an overview of the system. The inspection flow of the system is as follows. First, an inspector places APs in multiple utility holes in the inspection section. Then, the inspector inserts multiple CNs into the sewer pipe from a utility hole at appropriate intervals. After being inserted into a pipe, they start recording the inside of the pipe. When each of the CNs enters the communication range of the wireless LAN of an AP, it transmits the video data to the AP. Then, the AP, in turn, forwards the video data to the video server via a cellular network. Finally, the inspector accesses the video server and views the videos recorded by multiple CNs. Since the inspector need not enter the pipe and wait until the CNs reach the end of the inspection range to view the video, this system reduces sewer inspection labor costs. Generally, by using a self-propelled camera robot in a sewer pipe, the pipe needs to be cleaned after stopping sewer water flows. However, our system does not require stopping the water flow, and it does not require cleaning the pipe. Therefore, the inspection operation can be performed safely and quickly. Moreover, the inspector can respond quickly to problems such as the failure of CNs and stalling in the pipe. In addition, when severe damage is discovered with respect to a pipe, the inspector can quickly address the failure.
Each CN consists of a camera, sensor, light, wireless LAN interface, battery, and a small computer installed in a waterproof transparent capsule. We developed a prototype CN for this system. Figure 2 shows a photograph of the prototype CN.
Figure 2. CN prototype with a waterproof capsule.
Each AP shown in Figure 3 was equipped with a transceiver, a battery, antennas, and a computer unit. The APs were installed inside utility holes and communicate with CNs drifting inside the pipe via wireless LAN. They also communicate with the server in the cloud and outside the utility holes via a cellular network and transfer videos received from the CNs.
Figure 3. An access point installed in a hole.

3.2. Transmission of Video Data from a Drifting Wireless Camera to an AP

As described in Section 1, the communication range of IEEE802.11n wireless LAN in a small diameter sewer pipe is quite short. Therefore, the amount of video data transferred from a CN to an AP can be quite small. Let us assume that a CN drifts at a speed of 0.3 m/s and that the CN stores the video at a bit rate of 2.5 Mbps, frame size of 1920 × 1080, 30 fps, and in H.264 format; and that APs are installed at intervals of 200 m in a sewer pipe. Let us also assume that the video data transfer rate over a wireless LAN is 20 Mbps and that the communication range of the AP is 8 m, which is the distance from which a CN can communicate with an AP while moving within a range of 16 m. Under these conditions, the CN accumulates approximately 200 MB ( 2.5 Mbps × 200 m / 0.3 m / s / 8 bit ) of video data while drifting between two neighboring APs, say A P k 1 and A P k in Figure 4. The CN takes approximately 100 s ( = 200 MB / 20 Mbps ) to transmit the 200 MB video data to the downstream AP ( A P k ). Since the CN drifts at 0.3 m/s, the period of time in which the CN is within the communication range of the AP is about 53 s ( = 16 m / 0.3 m / s ). This duration is insufficient for transmitting all video data. The CN cannot transmit all video data captured between APs within the interval of communicating with the AP. Therefore, Yasuda et al. proposed a protocol sewer video transmission protocol 2019 (SVTP2019), which utilizes multiple CNs to transmit videos covering all sections between the APs.
Figure 4. Example of transmission of video data from multiple CNs to A P k .

3.3. How Multiple CNs Transmit Video Data of a Different Part of the Section Relative to an AP

SVTP allows multiple CNs to transmit videos of different parts of a section to a downstream AP based on the timestamp of each CN when they capture the videos. Figure 4 shows an example of the transmission of video data from two CNs to an AP. C N i is the i th CN released into a sewer pipe, A P k is the k th AP from the released point of the CNs, and t i , k is the timestamp when C N i first receives a beacon packet from A P k . The following is a brief description of the SVTP. A P k periodically broadcasts beacon packets, including τ : the elapsed time after receiving the first beacon from A P k 1 that corresponds to the position of the upper stream edge of a section of the sewer pipe that the AP wants to receive from the next CN. If A P k has never received video frames captured between the section between itself and A P k 1 , then τ = 0 . Precisely, A P k receives video frames after each CN receives a beacon from A P k 1 . For example, if C N 1 sends Δ t seconds of video frames captured between t 1 , k 1 and t 1 , k 1 + Δ t while passing through the wireless communication range of A P k , then A P k starts to broadcasts a beacon packet, including τ = Δ t . The next CN, C N 2 , will send the video of the rest of the sections between A P k 1 and A P k . C N 2 receives a beacon packet from A P k . The beacon packets include the updated τ = Δ t . Thus, it sends video frames that it captured after t 2 , k 1 + Δ t δ t , where δ t is a margin for the overlap of the sections included in the videos sent by C N 1 and C N 2 .

3.4. Where Was the Frame Captured?

If the velocity of the water in the sewer pipe is constant and known and the position of the CN when it first received a beacon from an AP is known, then estimating the position of the CN when it captured a video frame between the AP and the next AP is easy. However, the speed of the water flow is not generally constant, and APs are installed at only some of the utility holes in the inspection area. Thus, if the distance between APs is long, the position estimation error caused by assuming a constant water speed will be significant. In addition, even if the water flow is constant, the positions where C N i and C N j firstly receive the beacon packet from some APs differs because of the interval of beacons from the AP and the interval between the release times of C N i and C N j . To minimize this error, in the following section, we propose a method to locate CNs based on the known locations of landmarks such as APs, utility holes, and pipe joints.

4. Linking a Video Frame to the CN’s Position

In this section, we propose a method for estimating the location of a CN based on the elapsed time after the CN passes the closest landmark. We use the linear interpolation technique based on the location of landmarks, such as APs and utility holes with known positions.

4.1. Problem Definition and System Model

The problem we solve in this paper involves identifying the position of a drifting camera node (CN) moving in a sewer pipe, and this is performed by assuming the following system model.
  • The CN cannot use a GNSS. The CN and APs do not have sensors that can be used to estimate the position of the CN except for cameras and wireless communication interfaces.
  • The CN can communicate with access points (APs) via a radio communication link when it is close to one of them (e.g., the distance between the CN and an AP is less than 10 m). The APs are installed in some of the utility holes. The distances between neighboring utility holes are longer than twice the maximum communication distance between the CN and an AP. The received signal strength of the radio signal from an AP is sufficiently low when the distance between the CN and the AP is longer than the maximum communication distance. Thus, the CN cannot use the received signal strength of the radio signal from an AP to estimate its position when the distance between the CN and the AP is long.
  • One camera is installed in the CN. The camera points straight up and records videos of the pipe wall, ceiling, and the interior of a utility hole that is close to the CN.
  • The CN records videos while it drifts in the sewer pipe and records the timestamp when each video frame is captured. When the CN can communicate with an AP, the CN sends the recorded video data with timestamps of the video frames to the AP. The video server that receives all data received by the APs estimate the CN’s position when each frame is captured.
  • The starting position of the CN and positions of all utility holes and joints of pipe segments are known.

4.2. Localization Strategy

As described in Section 2, ultrasonic and radio-wave-based indoor localization methods are unsuitable for estimating the location of CNs drifting in sewer pipes. Considering the cost and weight, estimating the position of CNs using only wireless LAN functionality and cameras attached to the CN is difficult without adding any new devices to the CN. In addition, the CN’s position estimation error should be less than the minimum unit relative to pipe maintenance so that it does not lead to construction work in the incorrect areas. The typical segment size of a 200 mm in diameter sewer pipe is 2 m [39]. Thus, the maximum error of the estimate of the CN’s position should be less than 2 m. Since the shapes of joints of neighboring pipe segments and utility holes are easily detected by using classical image-processing techniques, as we explain later, they can be used as landmarks to estimate the position of the CN. Therefore, the proposed method uses a camera and the wireless LAN function to detect landmarks in a sewer pipe and to estimate the position of a CN using linear interpolation based on the elapsed time after passing the last landmark.

4.3. CN Localization Based on Linear Interpolations Using Landmarks

For the sake of simplicity, the following description assumes that only one CN is used; the same method can be used to locate two or more CNs. The timestamp used to localize a CN is assumed to be obtained from the clock on the CN. Let P m be the position of the m th landmark that the CN encountered, since it started drifting, and t m be the time when the CN passes P m , as shown in Figure 5. Then, the position of the CN P ( t ) at any time t during interval [ P m 1 , P m ] (the linear interpolated route as shown in Figure 6 is expressed as follows.
P ( t ) = P m 1 + ( P m P m 1 ) t t m 1 t m t m 1
Figure 5. Landmarks (AP and utility holes) and a CN drifting between them.
Figure 6. Linear interpolation using landmarks.
The equation uses only the time difference of local clocks on the same device. Therefore, the local clocks of the CNs need not be synchronized with other devices. When aggregating video data, the video server can estimate the position at which the video frame was captured based on the local timestamp of the CN linked to each frame. Since the estimation error monotonically increases with the distance between the last landmark that the CN passed and the CN, the more landmarks there are, the more accurate the estimation.
The potential landmarks in sewer pipes include utility holes, sewer pipe segment joints, and APs. The typical distance between the utility holes of small-diameter sewer pipes ( ϕ 200 mm–250 mm) targeted in this study is several tens of meters, whereas the length of a sewer pipe segment is 990–2000 mm [39]. Therefore, if all joints can be detected, the estimated position error of the CN by linear interpolation will be less than 2000 mm.

4.4. Landmark Detection

We can easily detect the positions of the utility holes and joints of pipe segments using classic image-processing techniques.
Utility holes 
Since the shape of the lid of a utility hole is a circle, we can identify the position of a utility hole by detecting a circle of a suitable size from a video frame captured by a CN, as shown in Figure 7. The time when the center of the detected circle is the closest to the center of the frame is recorded as the time when the CN passes the utility hole. A circular shape can be detected using the Hough transform technique [40].
Joints of pipe segments 
Since the shape of the joint of two pipe segments is a straight line perpendicular to the pipe’s direction, we can identify the position of a pipe’s joint by detecting a linear shape from a frame recorded by a CN and checking the relative angle of the line and the pipe’s direction. The pipe’s direction can be detected by finding two dark areas corresponding to the upstream side and downstream side, as shown in Figure 8. The time when the center of the detected line is the closest to the center of the frame is recorded as when the CN passes the landmark: a pipe joint. Such a linear shape can generally be detected using the Hough transform technique [41].
The detailed operation for detecting a circle or straight-line shape is as follows. Firstly, we convert the color of each captured video frame to grayscale and apply histogram flattening processes. Then, we apply image edge detections based on the Canny method [42] implemented in openCV [43]. The Hough transform is applied to this binary image to detect straight lines and circles.
First beacon reception points from an AP 
Each AP sends beacon packets periodically. When a CN receives a beacon packet from an AP with an ID that it has not received before, it stores the current timestamp.
Figure 7. Example of a video frame captured at the bottom of a utility hole.
Figure 8. Example of a video frame captured at a pipe segment joint.

5. Implementation

In this section, we describe the implementation of a function for estimating the CN’s position and for linking it to each video frame, as explained in Section 4. We implemented a function for the CN’s position estimation on the video server. Video data with timestamps were sent from each CN to APs. Then, APs forwarded the video data to the video server. Finally, the video server executed the CN localization function with the received video data.

5.1. Implementation of a Drifting Wireless Camera

We implemented a prototype of a CN, as shown in Figure 9. The CN prototype was equipped with a Raspberry Pi 3B, a camera with a fish-eye lens, LED lights inside a plastic container, and a battery. It records videos at a resolution of 1080p, frame rate 30 fps, and in H264 while moving inside of the sewer pipe.
Figure 9. A prototype CN.

5.2. Implementation of Utility Hole Detection

The implementation of image-processing functions required for CN localization was based on OpenCV [43]. Although we proposed three landmark detection methods (utility holes, pipe joints, and APs), we implemented a function for the utility holes to demonstrate the basic performance of the proposed method. To find the shape of a utility hole’s lid in a captured video frame, the function detects a circle with a suitable radius as a utility hole lid using Hough transforms (cv2.HoughCircles( )). Figure 7 shows a frame in which a CN drifts inside a sewer pipe and captures a 200 mm diameter utility hole lid. As shown in the figure, multiple circular-shaped objects (the edge of a utility hole’s lid and the edge of the joint connecting the utility hole and the sewer pipe) were included in the same frame when the utility hole lid was detected. Since the size of the circle of the edge of the joint is clearly larger than the size of the circle of the lid, we could easily and only detect the circle of the lid based on the radius of the circle. Thus, we determined the circle’s radius for detecting the lid using the detection system based on preliminary experiments. Since multiple consecutive frames included the images of the utility hole lid, we needed to identify the best frame from them; i.e., we needed to select a frame in which the lid was located closest to the center of the frame. We selected a frame that had the timestamp closest to the average of the timestamps of frames that included the shape of the utility hole’s lid.

5.3. Implementation of Linear Interpolation

After the video server finds the timestamp of the moment, the CN passes the utility hole (referred to as the passing time); it links the timestamp of each frame of the received video data to the location based on the linear interpolation, as follows. Let t be the timestamp of the frame. The video server finds the number m that satisfies the following condition: t m t t m + 1 , where t m is the time the CN passes the m-th landmark. From the locations of the m-th and m + 1 -th landmarks, P m and P m + 1 , which are provided for the video server in advance, the video server estimates the CN’s position, P ( m ) , using Equation (1). We assume t 0 = 0 and P 0 = 0 because the first landmark is the point where the CN is inserted into the pipe. If t is larger than t m max ( m max is the farthest landmark number detected by the CN), then the passing time of the last landmark, m max 1 , is given to m in Equation (1).

6. Operation Verification

We verified the accuracy of the proposed localization system using an underground pipe testbed. First, we explain the environment for the verification of the localization accuracy and verification method. Then, we show the results of the verification. Finally, we discuss the localization accuracy of the proposed method based on the results.

6.1. Verification

6.1.1. Verification Environment

We evaluated the performance of the proposed system in an underground reinforced concrete sewer pipe on our campus, as shown in Figure 10a,b. The length and diameter were 11 m and 250 mm, respectively. The depth from the ground surface to the bottom of the pipe was 520 mm. The pipe had vertical holes every 1 m, and we set lids and placed sandbags on all vertical holes during the experiments. The pipe had two water pools at both ends of the pipe. We placed a pump at the bottom of the pool (right side in Figure 10a) to pump water to the pool at the upper side through a pipe in order to generate a water flow. The pump was adjusted to generate a flow velocity of approximately 0.26 m/s. In addition, we set a rope with a scale from the upper side of the sewer pipe (the CN insertion point) to the removal point in advance so that we could visually confirm the captured location of each video frame.
Figure 10. Overview of the testbed. (a) Structure; (b) Picture.

6.1.2. Verification Method

We conducted the experiments using the following procedure. First, to control the CN, we pre-connected it to a PC via SSH in the 2.4 GHz IEEE 802.11n ad hoc mode. As soon as we released the CN from the first upstream handhole point, we sent a raspivid command [44] from the PC to the CN to start the camera recording. The recording duration was set to 40 s, which was longer than the time taken for the CN to drift down the sewer pipe. After the CN drifted down the pipe, we retrieved the CN. Since conducting more trials renders the measurement results reliable, we repeated this process 15 times. The results show that the 15 trials seem to be sufficient for understanding the accuracy of the proposed system.
After the trials, we transferred the videos recorded by the CN to a Linux machine, which ran the video frame localization program explained in Section 5 for each video frame and recorded the position of capture of each frame. We then checked the estimated positions of the captured video frames (via a web page we developed, as shown in Figure 11). To confirm the ground truth of the captured positions, we read the scale value in the image of the rope measure in each frame that estimated the captured positions at 0.5 × n m ( n = 0 , 1 , 2 , , 17 ) . We then checked the positions where the frames were captured by looking at the image of the rope scale included in the frames.
Figure 11. Web page for comparing the video frames and the captured position (displaying a result of the first trial experiment).

6.2. Results

Figure 12a shows the relationship between the estimated location of the CN and the ground truth over the 15 trial experiments in our proposed method. To evaluate the effect of detecting landmarks, we demonstrate the location estimation without using landmarks except for the start point and the last handhole. Figure 12b shows the relationship between the estimated location of the CN and the ground truth when under linear interpolation with only two landmarks at the 0 and 8 m points. For each trial, we plotted the relationship from 0 to 8.5 m, every 0.5 m, as we described in Section 6.1.2. The result of each trial was depicted with the different colors on the graph. The red dotted lines represent the positions of the handholes at 1, 5, 7, and 8 m in the sewer pipe. In Figure 12a, in all the trials, there were no significant shifts in the drifting positions. Furthermore, the localization error was within the distances relative to the adjacent landmarks. Comparing Figure 12a,b, we can observe that the variation in trajectories is larger in Figure 12b.
Figure 12. Relationship between the ground truth and the estimated captured position of the video frame under the linear interpolation. Each color corresponds to each of the 15 trials. (a) Landmarks at 0 m, 1 m, 5 m, 7 m, and 8 m are used; (b) Landmarks at 0 m and 8 m are used.
Figure 13a shows the distribution of the position estimation errors at various positions, and Figure 13b shows the error under the condition with linear interpolation by only two landmarks (landmark with 0 and 8 m points). For each trial in the figures, we plotted the error between the estimated location and the ground truth from 0 to 8.5 m, every 0.5 m. The result of each trial was depicted with the different colors on the graph in the same manner in Figure 12. In Figure 13, the errors at the handhole locations (1, 5, 7, and 8 m) were within 0.1 m. As shown in Figure 7, the entire circular shape of the handhole (shot from the bottom of the pipe to the top) was included in the video frame. Considering that the diameter of all the handholes was 200 mm, even with a 0.1 m error, the system successfully detected the shapes of all the handholes included in the camera view. The maximum error was approximately 0.46 m at the 2.5 m point, which was almost at the middle of the longest interval of handholes.
Figure 13. Video frame estimated position error. The error between estimated location and ground truth is depicted with different colors for each of the 15 trials. (a) Landmarks at 0 m, 1 m, 5 m, 7 m, and 8 m are used; (b) Landmarks at 0 m and 8 m are used.
Figure 13a presents that the longer the distance between the CN and the closest handhole, the more significant the variance in the position estimation’s error. This indicates that the velocity of the CN was affected by the water flow and the collisions of the CN with the pipe wall. The errors in the section between 1 and 5 m were all negative, indicating that the server estimated the point in front of the ground truth, as the CN may be increasing its velocity from the released point until it reaches a velocity equal to the velocity of the water flow. Therefore, we can estimate that the drifting velocities around 1 and 5 m points were slower than the average section velocity. Figure 13a has a smaller overall error in position estimations than Figure 13b. Furthermore, the maximum error in Figure 13a was −0.46 m at the 2.5 m point, and Figure 13b had a larger error of −1.15 m at 3.5 m and 4.0 and 4.5 m points. It can be seen that the more landmarks our proposed method uses for linear interpolation, the smaller the error in position estimations.
Figure 14a shows the cumulative distribution function (CDF) of the estimated position error observed at all estimated positions for the 15 trials. Figure 14b shows the CDF under the linear interpolation with only two landmarks at 0 and 8 m. In Figure 14, more than 90% of the estimated drifting position errors were within 0.3 m. On the other hand, Figure 14b shows that more than 90% of the estimated drifting position errors were within 0.9 m.
Figure 14. CDF of errors in the video frame’s estimated position. (a) Landmarks at 0 m, 1 m, 5 m, 7 m, and 8 m are used; (b) Landmarks at 0 m and 8 m are used.

6.3. Discussion

It was observed that the estimated drifting position errors were within the distance between neighboring handholes, which prevented the possibility of inspecting and excavating the incorrect section of the sewer pipe.
In 2-meter-long sections, between 5 and 7 m and between 7 and 8 m, the localization error was within 0.15 m, as shown in Figure 13. Since the length between the joints of sewer pipes is typically 2 m, as described in Section 4, it can be said that if both utility holes and the joints of the sewer pipe are added as landmarks, the error will be less than 0.15 m.
In the experiment, the CN detected all handholes in all trials. However, if some of the landmarks are not detected in the inspection area, it is impossible to use consecutive landmarks to estimate the CN’s position between them, causing a significant error during linear interpolations. To avoid errors, a function for detecting landmark detection failure is needed. As the typical joint interval and the time of passing the utility holes are known, and the moving speed of the CN can be estimated from the history of the time of passage, the landmark detection failure can be easily detected. However, the CN may collide with obstacles in a pipe, e.g., tree roots, and temporarily stop, causing a significant error in linear interpolation. To tackle the drawbacks of linear interpolation, we propose adding IMU and the odometry function described in Section 2 to the CN to reduce the estimation error at locations far from landmarks.
In these experiments, the drifting speed of the CN was set to 0.26 m/s. However, in a real sewer pipe environment, a CN drifts at a variable speed. If the speed is fast, the video frames captured by the CN cannot reflect all scenes of the inspection area. Let us consider a case when a CN captures video frames at 30 fps, one video frame covers a sewer pipe of length ranging 0.2 m (according to Section 6.2), and the shutter speed is sufficiently high for neglecting the drifting speed of the CN; then, the drifting speed of the CN is v m / s . The CN should satisfy condition v 0.2 m / f × 30 fps = 6 m / s to capture the entire scene of the inspection area. If the CN captures at 60 fps , it should satisfy condition v 0.2 m / s × 60 fps = 12 m / s as well. Considering the possible water speed in sewer pipes, this requirement for the velocity of the CN is easily achievable.

7. Conclusions

This paper described the proposal, design, and implementation of a video frame localization system for a drifting-camera-based sewer-pipe-inspection system. The main contributions of this paper are summarized as follows. Firstly, we proposed a method for estimating CN positions based on linear interpolation using utility holes and pipe joints as landmarks in a sewer pipe. We also implemented the proposed method together with a browsing system that links location information to each frame of the video and conducted experiments to evaluate the accuracy of the proposed localization function using a drifting CN in a real underground pipe. The evaluation results show that the utility holes (handholes) included in the video frames captured by a camera on a drifting camera node (CN) were successfully detected by using the Hough transform technique and contributed to accurately estimating the position of the CN, i.e., the position where each video frame was captured. The proposed localization system only requires the minimum required components on a CN, a camera, a light, and a Wi-Fi module, which are originally needed for the drifting-camera-based sewer-pipe-inspection system.
The results showed that all handholes, installed at 1, 2, and 4 m intervals, were successfully detected, and the maximum localization error was less than 0.46 m, which was 11.5% of the maximum interval of the handholes. Assuming that the pipe joints are used as landmarks, the accuracy of the video frame’s localization with the proposed method will be enough to identify damaged pipe segments, since the intervals are typically 2 m long. To the best of the authors’ knowledge, this is the first study presenting the localization of a drifting camera in a narrow sewer pipe and not a self-propelled robot, and it is based on image processing.
In the future, we plan to (1) achieve accurate detection of pipe joints; (2) conduct experiments using actual sewer pipes with large utility holes (manholes), obstacles, and damage; (3) combine the proposed method with IMU for more accurate localization of CNs; and (4) propose a method to correct the localization of CNs in cases some of the landmarks are failed to be detected.

Author Contributions

Conceptualization, Y.C. and S.I.; data curation, Y.C.; formal analysis, Y.C. and S.I.; funding acquisition, S.I.; investigation, Y.C., Y.T. and S.I; methodology, Y.C. and S.I.; project administration, S.I.; resources, S.I.; software, Y.C.; supervision, S.I.; validation, S.I.; visualization, Y.C.; writing—original draft, Y.C.; writing—review and editing, H.S. and S.I. All authors have read and agreed to the published version of the manuscript.

Funding

Part of this work was funded by SCOPE (Strategic Information and Communications Research and Development Promotion Programme), grant number 172306007, by the Ministry of Internal Affairs and Communications in Japan and JSPS KAKENHI, grant number 21K18746.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ministry of Land, Infrastructure, Transport and Tourism. Implementation of Planned Reconstruction. Available online: https://www.mlit.go.jp/crd/sewerage/policy/03.html (accessed on 24 October 2022).
  2. Ministry of Land, Infrastructure, Transport and Tourism. Implementation of Planned Reconstruction (Japanese Version). Available online: https://www.mlit.go.jp/mizukokudo/sewerage/crd_sewerage_tk_000135.html (accessed on 24 October 2022).
  3. Tohri Co. Ltd. Business Field. Available online: http://www.tohri-net.co.jp/field/survey-analysis/ (accessed on 24 October 2022).
  4. Fiberscope.net. 2022. Drain camera VIPER. Available online: https://www.fiberscope.net/portable-push-camera-video-inspection-pipe-drain-sewer-cctv/ (accessed on 24 October 2022).
  5. Proteus 2022. Proteus Systems–Minicam. Available online: https://www.minicam.co.uk/systems/20crawlers/proteus/ (accessed on 24 October 2022).
  6. Sekino, M. Just searching for self-propelled camera inspection system, monthly. J. Sew. 2011, 34, 55–58. [Google Scholar]
  7. Ishihara, S.; Liu, Z.; Tachibana, S.; Yasuda, T. Implementation of video data transmission protocol for a narrow sewer pipe screening system using drifting wireless cameras. In Proceedings of the 2020 International Conference on Information Networking (ICOIN), Barcelona, Spain, 7–10 January 2020; pp. 384–389. [Google Scholar] [CrossRef]
  8. Nagashima, T.; Tanaka, Y.; Ishihara, S. Measurement of Wireless LAN Characteristics in Sewer Pipes for Sewer Inspection Systems Using Drifting Wireless Sensor Nodes. IEICE Trans. Commun. 2016, 99, 1989–1997. [Google Scholar] [CrossRef]
  9. Zafari, F.; Gkelias, A.; Leung, K.K. A Survey of Indoor Localization Systems and Technologies. IEEE Commun. Surv. Tutor. 2019, 21, 2568–2599. [Google Scholar] [CrossRef]
  10. Zhang, R.; Zhong, W.; Kemao, Q.; Zhang, S. A Single LED Positioning System Based on Circle Projection. IEEE Photonics J. 2017, 9, 7905209. [Google Scholar] [CrossRef]
  11. Li, H.; Huang, H.; Xu, Y.; Wei, Z.; Yuan, S.; Lin, P.; Wu, H.; Lei, W.; Fang, J.; Chen, Z. A Fast and High-Accuracy Real-Time Visible Light Positioning System Based on Single LED Lamp with a Beacon. IEEE Photonics J. 2020, 12, 1–12. [Google Scholar] [CrossRef]
  12. Guan, W.; Chen, S.; Wen, S.; Tan, Z.; Song, H.; Hou, W. High-Accuracy Robot Indoor Localization Scheme Based on Robot Operating System Using Visible Light Positioning. IEEE Photonics J. 2020, 12, 1–16. [Google Scholar] [CrossRef]
  13. Jackson, J.C.; Summan, R.; Dobie, G.I.; Whiteley, S.M.; Pierce, S.G.; Hayward, G. Time-of-flight measurement techniques for airborne ultrasonic ranging. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2013, 60, 343–355. [Google Scholar] [CrossRef]
  14. Chew, M.T.; Alam, F.; Legg, M.; Sen Gupta, G. Accurate Ultrasound Indoor Localization Using Spring-Relaxation Technique. Electronics 2021, 10, 1290. [Google Scholar] [CrossRef]
  15. Hoeflinger, F.; Saphala, A.; Schott, D.J.; Reindl, L.M.; Schindelhauer, C. Passive Indoor-Localization using Echoes of Ultrasound Signals. In Proceedings of the 2019 International Conference on Advanced Information Technologies (ICAIT), Yangon, Myanmar, 6–7 November 2019; pp. 60–65. [Google Scholar]
  16. Zhang, R.; Höflinger, F.; Reindl, L. TDOA-Based Localization Using Interacting Multiple Model Estimator and Ultrasonic Transmitter/Receiver. IEEE Trans. Instrum. Meas. 2013, 62, 2205–2214. [Google Scholar] [CrossRef]
  17. Worley, R.; Yu, Y.; Anderson, S. Acoustic Echo-Localization for Pipe Inspection Robots. In Proceedings of the 2020 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), Karlsruhe, Germany, 14–16 September 2020; pp. 160–165. [Google Scholar]
  18. Dabove, P.; Pietra, V.D.; Piras, M.; Jabbar, A.A.; Kazim, S.A. Indoor positioning using Ultra-wide band (UWB) technologies: Positioning accuracies and sensors’ performances. In Proceedings of the 2018 IEEE/ION Position, Location and Navigation Symposium (PLANS), Monterey, CA, USA, 23–26 April 2018; pp. 175–184. [Google Scholar]
  19. Wang, C.; Xu, A.; Kuang, J.; Sui, X.; Hao, Y.; Niu, X. A High-Accuracy Indoor Localization System and Applications Based on Tightly Coupled UWB/INS/Floor Map Integration. IEEE Sens. J. 2021, 21, 18166–18177. [Google Scholar] [CrossRef]
  20. Queralta, J.P.; Almansa, C.M.; Schiano, F.; Floreano, D.; Westerlund, T. UWB-based System for UAV Localization in GNSS-Denied Environments: Characterization and Dataset. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), LAS Vegas, NV, USA, 24–30 October 2020; pp. 4521–4528. [Google Scholar]
  21. Krishnan, S.; Sharma, P.; Guoping, Z.; Woon, O.H. A UWB based Localization System for Indoor Robot Navigation. In Proceedings of the 2007 IEEE International Conference on Ultra-Wideband, Singapore, 24–26 September 2007; pp. 77–82. [Google Scholar]
  22. Maghdid, H.S.; Lami, I.A.; Ghafoor, K.Z.; Lloret, J. Seamless outdoors-indoors localization solutions on smartphones: Implementation and challenges. ACM Comput. Surv. 2016, 48, 53. [Google Scholar] [CrossRef]
  23. Poulose, A.; Kim, J.; Han, D.S. A Sensor Fusion Framework for Indoor Localization Using Smartphone Sensors and Wi-Fi RSSI Measurements. Appl. Sci. 2019, 9, 4379. [Google Scholar] [CrossRef]
  24. Hashem, O.; Harras, K.A.; Youssef, M. Accurate indoor positioning using IEEE 802.11mc round trip time. Pervasive Mob. Comput. 2021, 75, 101416. [Google Scholar] [CrossRef]
  25. Kim, J.; Lim, J.S.; Friedman, J.; Lee, U.; Vieira, L.; Rosso, D.; Gerla, M.; Srivastava, M.B. SewerSnort: A Drifting Sensor for In-situ Sewer Gas Monitoring. In Proceedings of the 2009 6th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks, Rome, Italy, 22–26 June 2009; pp. 1–9. [Google Scholar]
  26. Maletic, N.; Sark, V.; Ehrig, M.; Gutiérrez, J.; Grass, E. Experimental Evaluation of Round-Trip ToF-based Localization in the 60 GHz Band. In Proceedings of the 2019 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Pisa, Italy, 30 September–3 October 2019; pp. 1–6. [Google Scholar]
  27. NJR4652 Series (60 GHz Smart Sensor Micro-Module, Digital Output Type). Available online: https://www.nisshinbo-microdevices.co.jp/en/products/micro/sensor/njr4652.html (accessed on 24 October 2022).
  28. Datasheet: NJR4652F2S1 Series. Available online: https://www.nisshinbo-microdevices.co.jp/en/pdf/micro/datasheet/sensor/NJR4652JS1_Datasheet_Pre01e.pdf (accessed on 24 October 2022).
  29. Aitken, J.M.; Evans, M.H.; Worley, R.; Edwards, S.; Zhang, R.; Dodd, T.; Mihaylova, L.; Anderson, S.R. Simultaneous Localization and Mapping for Inspection Robots in Water and Sewer Pipe Networks: A Review. IEEE Access 2021, 9, 140173–140198. [Google Scholar] [CrossRef]
  30. Murtra, A.C.; Tur, J.M.M. IMU and cable encoder data fusion for in-pipe mobile robot localization. In Proceedings of the 2013 IEEE Conference on Technologies for Practical Robot Applications (TePRA), Woburn, MA, USA, 22–23 April 2013; pp. 1–6. [Google Scholar]
  31. Al-Masri, W.M.F.; Abdel-Hafez, M.F.; Jaradat, M.A. Inertial navigation system of pipeline inspection gauge. IEEE Trans. Control Syst. Technol. 2020, 28, 609–616. [Google Scholar] [CrossRef]
  32. Yan, Y.; Zhang, B.; Zhou, J.; Zhang, Y.; Liu, X. Real-Time Localization and Mapping Utilizing Multi-Sensor Fusion and Visual–IMU–Wheel Odometry for Agricultural Robots in Unstructured, Dynamic and GPS-Denied Greenhouse Environments. Agronomy 2022, 12, 1740. [Google Scholar] [CrossRef]
  33. Hansen, P.; Alismail, H.; Rander, P.; Browning, B. Pipe mapping with monocular fish eye imagery. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 5180–5185. [Google Scholar]
  34. Hansen, P.; Alismail, H.; Browning, B.; Rander, P. Stereo visual odometry for pipe mapping. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 4020–4025. [Google Scholar]
  35. Hansen, P.; Alismail, H.; Rander, P.; Browning, B. Visual mapping for natural gas pipe inspection. Int. J. Robot. Res. 2015, 34, 532–538. [Google Scholar] [CrossRef]
  36. Zhang, R.; Evans, M.H.; Worley, R.; Anderson, S.R.; Mihaylova, L. Improving SLAM in pipe networks by leveraging cylindrical regularity. In Proceedings of the Annual Conference Towards Autonomous Robotic Systems, Lincoln, UK, 8–10 September 2021; Volume 13054. [Google Scholar] [CrossRef]
  37. Alejo, D.; Caballero, F.; Merino, L. RGBD-based robot localization in sewer networks. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4070–4076. [Google Scholar]
  38. Alejo, D.; Caballero, F.; Merino, L. A Robust Localization System for Inspection Robots in Sewer Networks. Sensors 2019, 19, 4946. [Google Scholar] [CrossRef]
  39. NORMADOC. Documents Normatifs et Reglementaires. Available online: https://www.normadoc.com/english/jis-a-5372-2016.html (accessed on 24 October 2022).
  40. Ioannou, D.; Huda, W.; Laine, A. Circle recognition through a 2D Hough Transform and radius histogramming. Image Vis. Comput. 1999, 17, 15–26. [Google Scholar] [CrossRef]
  41. Duda, R.; Hart, P. Use of the Hough Transformation To Detect Lines and Curves in Pictures. Commun. ACM 1972, 15, 11–15. [Google Scholar] [CrossRef]
  42. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, PAMI-8, 679–698. [Google Scholar] [CrossRef]
  43. OpenCV. Open Source Computer Vision Library. Available online: https://opencv.org (accessed on 24 October 2022).
  44. Bellard, F. ffmpeg Contributers. FFmpeg—A Complete, Cross-Platform Solution to Record, Convert and Stream Audio and Video. Available online: https://ffmpeg.org (accessed on 24 October 2022).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.