Method and Device of All-in-Focus Imaging with Overexposure Suppression in an Irregular Pipe

To avoid depth-of-field mismatches caused by the changes in pipe structure and image overexposures caused by highly reflective surfaces while radial imaging irregular pipes, this paper proposes a novel all-in-focus, adaptable, and low scene-coupling method that suppresses overexposures in support of fault detection. Firstly, the pipeline’s radial depth distribution data are obtained by sensors, and an optimal all-in-focus imaging scheme is established by combining camera parameters. Secondly, using digital imaging technology, the high reflection effect produced by disparate light sources is comprehensively evaluated for overexposure suppression. Thirdly, a device is designed for imaging non-Lambertian free-form surface scenes under low illumination, providing the sequence images needed for the next step. Lastly, specific digital fusions are made to the sequential images to obtain an all-in-focus final image without overexposure. An image-quality analysis method is then used to measure the efficacy of the system in obtaining the characteristic information of the inner surfaces of an irregular pipe. Results of the experiment show that the method and device used are able to distinguish small 0.5 mm wide lines ranging from 40–878 mm depth and are capable of providing efficient image support for defect inspection of irregular pipes and free-form surfaces amongst other irregular surfaces.


Introduction
Irregular pipes (e.g., those with S-shaped inlets) have special profile and variable cross-sectional features [1,2] characterized by continuous and irregular changes in the axial and radial directions, and the special non-Lambertian coating sprayed on its inner surface shows high reflectivity [3]. Thus, it is difficult to acquire the high-quality images needed for rapid digital fault detection.
Among the contemporary techniques, axial and radial imaging are the most common. Axial imaging refers to single-viewpoint methods (e.g., closed-circuit television [4], fisheye [5], single-reflection [6,7], and catadioptric panoramic annular lens [8,9]), in which the imaging device is placed along the axial direction of the inner pipeline. This results in a circular ring image whose effects directly depend on the structural regularity and coincidence degrees between the visual and pipeline axes, resulting in an image distortion-correction algorithm needing to be applied [6,10]. Any sudden changes in pipeline appearance or deviations along the visual axis can reduce imaging effects and detection accuracy. Thus, it is difficult to directly apply axial methods to irregular pipes. Radial imaging implies that the visual axis of the imaging device is perpendicular to the inner surface of the pipeline, from which information is obtained directly [11]. Although this method can be adapted to circular pipes of varying sizes by changing lenses, irregular pipes typically provide frequent and large radial depth-of-field (DoF) changes that make information acquisition prohibitively difficult and time-consuming. Imaging systems with expanded DoFs are helpful, and researchers have offered several related solutions (e.g., wavefront coding [12,13], special flat lensing [14], and light-field imaging [15,16]) that improve DoF handling at the expense of spatial resolution. However, these results are unsuitable for reliable defect detection. To achieve the all-in-focus imaging requirements needed for irregular pipes, more recent studies have applied multifocus fusion methods using sequence images [17,18]. These methods have good fusion effects that can be applied to a wide range of tasks, but they rely on passive focusing algorithms that are also unsuitable for irregular pipes, particularly due to the lack of texture gleaned in the low illumination [19,20]. Later, Liu [21] proposed another method of depth-image segmentation, which achieved good all-in-focus results, but the graph-based algorithm was unsuitable for continuous imaging. Notably, when imaging surfaces have large depth changes, it tends to be difficult to control the segmentation depth. Thus, a multicamera fusion scheme with a multifocal lens [22] can help resolve the large DoF problem. However, the required equipment is bulky, the solution is overly complex, and the system does not work well inside irregular and small pipelines requiring large depth spans.
Active illumination is often required when imaging pipe interiors. However, the highly reflective and freely changing structural features quickly lead to overexposures, which reduce the utility of the acquired information for fault detection. To mitigate these negative effects, Shao [23] devised a digital micro-mirror device (DMD) to realize pixel-level spatiotemporal modulations of illumination, the principle of which is that the reflected light (adaptive stripes with spatiotemporal modulation) modulated by DMD is projected through a triangular prism, mirror, and projection lens onto a workpiece with a highly reflective surface, and a uniform illumination is formed on the workpiece surface to highlight the defect information. Liu [24] introduced a study that uses a coaxial light source to obtain the first coaxial optical images to identify the bearing center coordinates and text information; then, under the multiangle light, images are obtained at different angles in turn, and all the defects of the bearing can be highlighted after image fusion. Feng [25] realized the acquisition of surface defect information of small highly reflective parts through high dynamic imaging technology and improved the quality of acquired images. Chen [26] designed a dual lighting system with a large area and high-brightness illumination from both sides of the steel ball; this design improved the light uniformity on the steel ball surface, thereby avoiding the generation of light spots and creating good conditions for subsequent detection. Although these methods addressed specific highly reflective objects and achieved decent results, most are meant to be used with the outer surfaces of objects. Notably, when imaging devices and algorithms are designed for specific tough jobs, their performance tends to be tightly coupled to the unique configuration; thus, they poorly adapt to changing scenes.
In summary, to obtain a method for solving the DoF mismatch problem while supporting overexposure suppression imaging of the insides of irregular pipes, this paper first proposes a cross-modal all-in-focus imaging device and strategy based on the target surface depth using a lighting and imaging method of overexposure suppression that depends on high-reflection prior information. Then, an imaging device is introduced, and an imaging experiment of 0.5 mm wide fine lines is carried out on the highly reflective surface with a maximum depth of 878 mm. Subsequently, an image fusion is carried out and the best all-in-focus image with no overexposure is obtained. The proposed method and device can provide efficient image support for defect inspection of special-shaped pipes, free-form surfaces, and other unconventional scenes, and it has good practical ap-plication value.
The structure of the paper is organized as follows: firstly, the background and key issues of all-in-focus imaging and imaging on high reflect surface are presented in Section 1, along with an overview of the relevant literature. Section 2 elaborates the method of all-infocus imaging through fusing the depth date of imaging surface, as well as the method for overexposure suppression on surface which has high reflection character. In Section 3, an imaging device, which has a high-resolution camera, depth sensors, light sensors, and four evenly spaced light sources, is designed, and three imaging experiments and evaluations of the imaging effects are applied to verify the effectiveness and superiority of the above proposed method. Lastly, Section 4 summarizes the key technologies and achievements of this paper.

Cross-Modal All-in-Focus Imaging Strategy
The DoF is the key imaging feature of a pinhole camera, as shown in Figure 1, and it is determined by the diameter of the allowable circle of confusion, δ, focal length, f, and lens F-value, which can be characterized by the front and back DoFs: where ∆L front represents the front DoF, ∆L back represents the back DoF, L represents the shooting distance from the focal plane to the photosensitive element, and δ is a hyperparameter set according to the sharpness requirement. Generally, the photosensitive elements size is selected in [1, n] pixels for the δ number. In this study, to clearly demonstrate the law of front and back scenes along an object plane of distance L, we let where D front (L) is the depth function of the DoF's front edge (Figure 1), and D back (L) is the depth function of its back edge. According to Equation (3)'s Gaussian imaging formula, for a fixed-focus pinhole imaging system, the image distance, v, can be changed by moving the lens focus. Thus, changes in v allow objects to be clearly imaged in the corresponding DoF under different object distances, u.
an imaging device, which has a high-resolution camera, depth sensors, light sensors, and four evenly spaced light sources, is designed, and three imaging experiments and evaluations of the imaging effects are applied to verify the effectiveness and superiority of the above proposed method. Lastly, Section 4 summarizes the key technologies and achievements of this paper.

Cross-Modal All-in-Focus Imaging Strategy
The DoF is the key imaging feature of a pinhole camera, as shown in Figure 1, and it is determined by the diameter of the allowable circle of confusion, δ, focal length, f, and lens F-value, which can be characterized by the front and back DoFs: where ΔLfront represents the front DoF, ΔLback represents the back DoF, L represents the shooting distance from the focal plane to the photosensitive element, and δ is a hyperparameter set according to the sharpness requirement. Generally, the photosensitive elements size is selected in [1, n] pixels for the δ number. In this study, to clearly demonstrate the law of front and back scenes along an object plane of distance L, we let where Dfront(L) is the depth function of the DoF's front edge (Figure 1), and Dback(L) is the depth function of its back edge. According to Equation (3)'s Gaussian imaging formula, for a fixed-focus pinhole imaging system, the image distance, v, can be changed by moving the lens focus. Thus, changes in v allow objects to be clearly imaged in the corresponding DoF under different object distances, u. The distance from the inner surface of the irregular pipe to the geometric centroid of the respective radial cross-sections is large. When the camera is at the center of the crosssection from different viewpoints, the large DoF differences required by each viewpoint cause single-or multifocus imaging techniques to fail to match the appropriate DoF. To The distance from the inner surface of the irregular pipe to the geometric centroid of the respective radial cross-sections is large. When the camera is at the center of the crosssection from different viewpoints, the large DoF differences required by each viewpoint cause single-or multifocus imaging techniques to fail to match the appropriate DoF. To capture images in such scenes, a DoF strategy of pinhole and cross-mode all-in-focus imaging is needed, as shown in Figure 2.
where the current depth range is included in any DoF with LFS as the object-plane depth, including the camera imaging range at View point A. To ensure good imaging quality, the intermediate depth of the scene can be used to focus the image in the single-focus mode, as follows: where DFsingle(dfA, dbA) denotes the depth of the object plane when the depth range is (dfA dbA). Correspondingly, when the camera is at Viewpoint B for any LFM ∈ (Dback −1 (dbB), Dfront −1 (dfB)) there are (dfB, dbB) ⊄ (Dfront(LFM), Dback(LFM)), such that when Dfront −1 (dfB) < Dback −1 (dbB), the DoF corre sponding to any object plane, LFM, cannot completely cover the current surface, and multipl segmented focusing images are required. The mathematical relationship of the corresponding conditions is expressed as where <Dn> is the depth range of each segment after the target depth is segmented, and N is the number of segmentations, where Nmin is the minimum value. The object plane depth Dobj(LFMj), corresponding to depth range <Dj>, is analyzed using the abscissa of point M j B where j = 1, 2, …, n, and n = 3 in Figure 2b where DFMj represents the depth of each object plane corresponding to <Dj> of multifocu imaging, and DFMj = Dobj(LFMj) = LFMj. In this paper, the process of dividing (df, db) into {<D1> <D2>, …, <Dn>} is called "depth segmentation". The process of calculating a {<D1>, <D2> for any L FS ∈(D back −1 (d bX ), D front −1 (d fX )), there is (d fX , d bX ) ∩ (D front (L FS ), D back (L FS )) where the current depth range is included in any DoF with L FS as the object-plane depth, including the camera imaging range at Viewpoint A. To ensure good imaging quality, the intermediate depth of the scene can be used to focus the image in the single-focus mode, as follows: where D Fsingle (d fA , d bA ) denotes the depth of the object plane when the depth range is (d fA , d bA ). Correspondingly, when the camera is at Viewpoint B for any L FM ∈ (D back −1 (d bB ), where DFsingle(dfA, dbA) denotes the depth of the object plane when the depth range is (dfA dbA). Correspondingly, when the camera is at Viewpoint B for any LFM ∈ (Dback −1 (dbB), Dfront −1 (dfB)) there are (dfB, dbB) ⊄ (Dfront(LFM), Dback(LFM)), such that when Dfront −1 (dfB) < Dback −1 (dbB), the DoF corre sponding to any object plane, LFM, cannot completely cover the current surface, and multiple segmented focusing images are required. The mathematical relationship of the corresponding conditions is expressed as where <Dn> is the depth range of each segment after the target depth is segmented, and N is the number of segmentations, where Nmin is the minimum value. The object plane depth Dobj(LFMj), corresponding to depth range <Dj>, is analyzed using the abscissa of point M j B where j = 1, 2, …, n, and n = 3 in Figure 2b where DFMj represents the depth of each object plane corresponding to <Dj> of multifocu imaging, and DFMj = Dobj(LFMj) = LFMj. In this paper, the process of dividing (df, db) into {<D1> <D2>, …, <Dn>} is called "depth segmentation". The process of calculating a {<D1>, <D2> ..., <Dn>} that matches {DFM1, DFM2, ..., DFMn} is called "DoF matching". Accordingly, in view of the large radial depth span of an irregular pipe, a cross-modal adaptive all-in-focu imaging strategy is proposed, as shown in Equation (7).
where {DF(df, db)} is the set of depth positions of the object plane after solving the (df, db (D front (L FM ), D back (L FM )), such that when D front −1 (d fB ) < D back −1 (d bB ), the DoF corresponding to any object plane, L FM , cannot completely cover the current surface, and multiple segmented focusing images are required. The mathematical relationship of the corresponding conditions is expressed as where <D n > is the depth range of each segment after the target depth is segmented, and N is the number of segmentations, where N min is the minimum value. The object plane depth, D obj (L FMj ), corresponding to depth range <D j >, is analyzed using the abscissa of point M j B , where j = 1, 2, . . . , n, and n = 3 in Figure 2b. Then, for Viewpoint B, when , all sub-focal plane positions in the multifocus mode can be obtained. {D Fmulti (d fB , d bB )} = {D FM1 , D FM2 , · · · · · · , D FMj }, and where D FMj represents the depth of each object plane corresponding to <D j > of multifocus imaging, and D FMj = D obj (L FMj ) = L FMj . In this paper, the process of dividing (d f , d b ) into {<D 1 >, <D 2 >, . . . , <D n >} is called "depth segmentation". The process of calculating a {<D 1 >, <D 2 >, . . . , <D n >} that matches {D FM1 , D FM2 , . . . , D FMn } is called "DoF matching". Accordingly, in view of the large radial depth span of an irregular pipe, a cross-modal adaptive all-in-focus imaging strategy is proposed, as shown in Equation (7).
where {D F (d f , d b )} is the set of depth positions of the object plane after solving the (d f , d b ) depth interval. By imaging the object planes in the set, the all-in-focus sequence images in the current field of view can be obtained.

Lighting and Imaging Strategy of Overexposure Suppression
The classic Phong model [27] can be used to quantitatively describe the relationship between the amount of light observed as a function of the surface profile and light and viewing angle. As shown in Figure 3a, L, N, R, and V are the incident light, imaging surface normal, reflected light, and observation vectors, respectively. When the small-area diffuse light source (LS) is independent and unique, Equation (8) is obtained as follows: where k d × I pd × cosi and k s × I ps × cos m θ are the diffuse and specular reflection components, respectively, and m is the reflection light convergence index related to surface smoothness. As the observation angle, θ, decreases, there is a region named area-of-reflect (Ar), in which the specular reflection component increases sharply by the m-th power, which is a high-reflection region.  As shown in Figure 3d, A-1, A-2, and B are imaging systems with illuminations in which the positional relationships between the camera and the small light source are relatively fixed. C-LS1 and C-LS2 are imaging systems with illuminations that lack fixed positional relationships between the camera and light source. The reflection problem of free- When a small area diffuse light source is used for direct illumination, the effect of the specular reflection component of the non-Lambertian plane is shown in Figure 3b. Under the action of LS1, Ar1, Ar2, Ar3, and Ar4 are the direct high-reflection, high-reflection transition, conventional reflection, and low-reflection areas, respectively. Consequently, the specular reflection intensity decreases. According to Equation (8) and Figure 3b, the high-reflection areas created by light sources of different paths do not completely overlap. Therefore, under a fixed viewing angle, the high-reflection areas, Ar1 and Ar2, can only be generated by light source LS2 in the incident optical path, IL-RL; however, this is not true for other light sources, such as LS3. Therefore, it is possible to obtain image sequences with highly reflective position differences by separately illuminating and imaging with small light sources at different positions to recover surface information through image fusion [26,28].
As shown in Figure 3d, A-1, A-2, and B are imaging systems with illuminations in which the positional relationships between the camera and the small light source are relatively fixed. C-LS1 and C-LS2 are imaging systems with illuminations that lack fixed positional relationships between the camera and light source. The reflection problem of free-form surfaces mainly results from the coupling of three factors: the surface feature structure of the free-form surface, camera pose, and light-source pose. The reflection type is set by the bright (high reflection) or dark (non-high reflection) field-forward caused by these three conditions [29]. See Figure 3d for poses A-1 and A-2 of the same viewpoint position and A-1, or A-2 and B, facing the same area but with different poses. Even under the same camera pose, C, when lighting conditions LS4 and LS5 differ, the imaging effects are not highly reflective at C-LS4, but are highly reflective at C-LS5, and different surface and lighting conditions can produce the same effects. Under freely changing imaging conditions and the mutual influence of reflective factors, the system must adopt different lighting schemes to adapt to the various changes.
Because the surfaces and camera poses are separately determined for each image, we designed a dynamic lighting device comprising four independently controllable small diffuse light sources in the same plane. The device produces low coupling effects between the imaging characteristics and surface features. As shown in Figure 3c, each light source is combined using the same intermediate camera to create a relatively fixed position, and a lighting and imaging overexposure suppression matching strategy is proposed as follows: • Four light sources are used to separately provide illumination and pre-imaging.

•
The reflection of the image formed under the illumination of each light source is calculated to obtain each source's prior information.

•
The prior information that produces the least high reflection is chosen.
If high reflections cannot be avoided in this fashion, the complementary information between images formed by two light sources is used to suppress the high reflection effect via fusion. The decision conditions are determined as follows: • Condition-Ca: If a single light source, a or b(a, b∈{T light , B light , L light , R light }), can avoid the high-reflection overexposure, it is selected for supplementary light, • Condition-Cb: If highly reflective overexposure cannot be avoided, a combined lighting scheme with dual light sources, a and b(a, b∈{T light , B light , L light , R light }), is selected for supplementary illumination, and the decision equation is expressed as where IMG a represents the image obtained when light source a provides illumination. Size(IMG a ) indicates the size of the area of the overexposure in IMG a , T light , B light , L light , and  (10), LC(IMG a ,IMG b ) is the decision value of the light-source combination (LC), and the overall reflective intensity (ORI = 2 − 2/(1+ e (−kα × α) )) represents the overall highly reflective imaging condition under two light sources. The spot overlap intensity (SOI = 3/(3 + 10 × e (kβ × (β − 1)) ) is used to characterize the overlap between the image spots formed under the two light sources, where α and β are the normalized and coincident spot areas after the AND operation (S and (IMG a ,IMG b )). The OR operation (S or (IMG a ,IMG b )) pertains to spots α = S and (IMG a ,IMG b )/S img and β = S or (IMG a ,IMG b )/S img , and k O and k M are the ORI and SOI weights, respectively, in the decision-making system, where k O + k M = 1. kα and kβ represent the total and coincident spot area penalty coefficients, respectively. A larger penalty coefficient results in a stronger attenuation effect of the increase in α and β on the overall decision-making result. A larger LC value of the decision result of the decision-making formula leads to a better final imaging effect. The important function of this formula is the selection of the two light sources that provide the smallest total spot area for illumination and imaging under the condition that the spot overlap is the smallest possible. The sequence of images obtained by this decision are obtained after fusion to produce the image with the best high-reflection suppression from the current viewpoint.

Imaging Control and Fusion Scheme
Owing to the complex relationships among surface features, illumination angles, and camera poses, sample quality is often significantly reduced. To remedy this problem, the adaptability of the imaging device is improved, as shown in Figure 4a, using the following adaptive decision-making control strategy:

1.
Depth data acquisition and all-in-focus scheme: Using the depth sensor to obtain the depth distribution of the unknown curved surface in the current field of view, depth data are applied to Equation (7).

2.
Pre-imaging and lighting plan: Using the depth data, different light sources are used for single-exposure pre-imaging at a fixed photosensitive level at the middle depth of the current surface and analyzing quality by judging whether there is overexposure or calculating the exposure conditions. The acquired prior information of high reflectivity under each light source is used to form the lighting plan using Equations (9) and (10).

3.
Full-focus imaging under an optimal lighting scheme: The combination of four imaging scenarios is present (i.e., I: single-focus single-illumination scene with small DoF; II: single-focus dual-illumination scene with small DoF; III: multifocus singleillumination scene with large DoF; and IV: multifocus dual-illumination scene with large DoF). Then, the all-in-focus scheme of Step 1 and the lighting scheme of Step 2 are used to match one of the four preset combinations, execute the illumination scheme once, and image each object plane separately.

4.
Sequence image fusion: After acquiring the sequence images, a specific fusion scheme (e.g., multifocus or wavelet) is performed on the sequence images according to the combined method chosen in Step 3. The image fusion process in the four scenarios proceeds as described below.
As shown in Figure 4b, the output images of each mode are marked as Img-Type-I, Img-Type-II, Img-Type-III, and Img-Type-IV. During object plane imaging under each light source, multi-exposures of the four photosensitive levels are performed by changing the exposure time, and fusion is performed to obtain the high-dynamic range (HDR) image needed to solve the problem of insufficient dynamics and poor brightness uniformity under single imaging. This then facilitates the expression of internal defects and other information needed to obtain high-definition images in the full DoF. This process is indicated by the red arrow in Combination I in Figure 4b.
Img-Type-II, Img-Type-III, and Img-Type-Ⅳ. During object plane imaging under each light source, multi-exposures of the four photosensitive levels are performed by changing the exposure time, and fusion is performed to obtain the high-dynamic range (HDR) image needed to solve the problem of insufficient dynamics and poor brightness uniformity under single imaging. This then facilitates the expression of internal defects and other information needed to obtain high-definition images in the full DoF. This process is indicated by the red arrow in Combination I in Figure 4b.
(a) (b) Figure 4. Imaging and fusion process: (a) the process of image acquisition by the imaging method in this paper; (b) the method and process of image fusion in this paper. I, the process in the red box indicates that the sequence images obtained under the imaging combination I in step 3 are subjected to multi-exposure fusion to obtain the Img-Type-1; the processes in the blue box and the orange box in II and III respectively indicate that, under imaging combinations II and III, by using wavelet fusion or multifocus fusion, the images of Img-Type-2 or Img-Type-3 can be obtained; the process shown in the purple box IV shows that image of Img-Type-4 can be obtained by fusing two images of Img-Type-3 under combination IV by multifocus fusion.

All-in-Focus Imaging Device with Overexposure Suppression
Spatial resolution is an important performance parameter for measuring the ability of an imaging system to resolve fine defects. According to past research [30], the spatial resolution, Rspace, of radial imaging can be described as follows: indicates that the sequence images obtained under the imaging combination I in step 3 are subjected to multi-exposure fusion to obtain the Img-Type-1; the processes in the blue box and the orange box in II and III respectively indicate that, under imaging combinations II and III, by using wavelet fusion or multifocus fusion, the images of Img-Type-2 or Img-Type-3 can be obtained; the process shown in the purple box IV shows that image of Img-Type-4 can be obtained by fusing two images of Img-Type-3 under combination IV by multifocus fusion.

All-in-Focus Imaging Device with Overexposure Suppression
Spatial resolution is an important performance parameter for measuring the ability of an imaging system to resolve fine defects. According to past research [30], the spatial resolution, R space , of radial imaging can be described as follows: where f is the focal length, l is the distance between lens and object, l ≈ L, size pixel is the single-pixel size of the occupied photosensitive element, and size obj is the minimum width of the defective object. To achieve the imaging requirements of size obj with 0.5 mm wide fine lines at L = 1000 mm in the irregular pipe, SONY IMX477 is selected as the photosensitive imaging element. According to Equation (12), the lens has a focal length, f, of 3.9 mm, aperture value, F, of 2.8, and field of view (FoV) of 75 × 52. For the depth sensor selection, the ultra-miniature multiarea depth sensor, VL53L5CX, with a resolution of 8 × 8 and an FoV of 45 • × 45 • was selected, and the two were combined as shown in Figure 5. Due to the excellent multizone ranging ability of the sensor, this device can measure the depth data of a plane, and its measurement accuracy reached 5% under the condition of this paper. A high-brightness light-emitting diode (LED) lamp bead with a color temperature of 5500 K was selected as the light source, and a polymethylmethacrylate diffuser lens with a diameter of 15 mm and an illumination angle of 120 • were selected as the condensing element. The light source array, ultra-miniature area depth sensor, and geometric installation relationship with the camera are shown in Figure 5. 8 and an FoV of 45° × 45° was selected, and the two were combined as shown in Figure 5. Due to the excellent multizone ranging ability of the sensor, this device can measure the depth data of a plane, and its measurement accuracy reached 5% under the condition of this paper. A high-brightness light-emitting diode (LED) lamp bead with a color temperature of 5500 K was selected as the light source, and a polymethylmethacrylate diffuser lens with a diameter of 15 mm and an illumination angle of 120° were selected as the condensing element. The light source array, ultra-miniature area depth sensor, and geometric installation relationship with the camera are shown in Figure 5. Controlling the exposure of sequential images is crucial to the accuracy of pre-imaging information and the acquisition of multiple-exposure sequential images. Exposure control is determined by the aperture value, International Standards Organization (ISO) sensitivity, and exposure time, and it is closely related to the stable light-field intensity. Because the overall reflection of incident light on the curved surface differs greatly per light source, an independent four-channel constant-current source-drive circuit and proportional-integral-derivative (PID) control algorithm are used to achieve the constant control of light intensity, as shown in Figure 6. The CN5711 constant-current LED driver, which has excellent current stability and brightness retention over a wide temperature range, is used to maintain light-level consistency between images. The algorithm uses light-intensity feedback to modulate the brightness of each source using pulse-width modulation. The PID control formula is shown in Equation (13), where e is the current error, and et−1 is the last error originating from the feedback data of the two light intensity sensors ( Figure 5). Simultaneously, to ensure the consistency of final sequence images, the process is controlled according to the consistency parameters presented in Table 1. Controlling the exposure of sequential images is crucial to the accuracy of pre-imaging information and the acquisition of multiple-exposure sequential images. Exposure control is determined by the aperture value, International Standards Organization (ISO) sensitivity, and exposure time, and it is closely related to the stable light-field intensity. Because the overall reflection of incident light on the curved surface differs greatly per light source, an independent four-channel constant-current source-drive circuit and proportional-integralderivative (PID) control algorithm are used to achieve the constant control of light intensity, as shown in Figure 6. The CN5711 constant-current LED driver, which has excellent current stability and brightness retention over a wide temperature range, is used to maintain lightlevel consistency between images. The algorithm uses light-intensity feedback to modulate the brightness of each source using pulse-width modulation. The PID control formula is shown in Equation (13), where e is the current error, and e t−1 is the last error originating from the feedback data of the two light intensity sensors ( Figure 5). Simultaneously, to ensure the consistency of final sequence images, the process is controlled according to the consistency parameters presented in Table 1.

Imaging Experiments
To verify the high-quality adaptability of the proposed imaging device under arbitrary DoFs, the experimental setup shown in Figure 6a was applied. A thin, gray, specular reflective iron sheet with diffuse reflection characteristics was used to build the scene (Figure 6e). The device was placed on a height-adjustable x-y coordinate slide to simulate different positional and attitudinal perspectives. The film ruler shown in the rectangle was attached to the curved surface and used to judge the final clarity. The device was designed to adaptively match the scene to combination Types I-IV based on the acquired depth and pre-imaging information. In this study, three viewpoints, x(x = i, ii, iii), of different combination types were imaged in the dark environment to verify the adaptability and efficacy of the proposed strategy and device. The viewpoint positions are shown in Figure 6b-d.

Single-Focus and Illumination under Small DoF
The positional relationship between the camera and scene was adjusted to Viewpoint i, as shown in Figure 6b. According to the process of Figure 4a, an imaging experiment was performed using the developed imaging device. The experimental steps and results are as follows:
To improve the operation speed, the obtained image is reduced to 10% of the original To improve the operation speed, the obtained image is reduced to 10% of the original for binarization where the threshold, TH, is 245. Then, a 5 × 5 opening operation is performed to obtain image with interference removed. The spot area Size(IMG a ) is calculated for a ∈ {T light , B light , L light , R light }, and the optimal spot area under different light sources is discerned as shown in Table 2. The results show that L light would not cause overexposure, which satisfies decision Condition Ca of Equation (9). Here, L light is the final decision. Typical image-processing is illustrated in Figure 8a. To improve the operation speed, the obtained image is reduced to 10% of the original for binarization where the threshold, TH, is 245. Then, a 5 × 5 opening operation is performed to obtain image with interference removed. The spot area Size(IMGa) is calculated for a ∈ {Tlight, Blight, Llight, Rlight}, and the optimal spot area under different light sources is discerned as shown in Table 2. The results show that Llight would not cause overexposure, which satisfies decision Condition Ca of Equation (9). Here, Llight is the final decision. Typical image-processing is illustrated in Figure 8a. The final imaging scheme from Viewpoint i combines the full-focus scheme of Step 1 and the lighting scheme of Step 2, and DFS = 386 mm is selected as the object plane for HDR imaging under Llight. The imaging consistency control parameters are listed in Table 1.
As shown in Figure 8b, the multi-exposure fusion algorithm [31] is used on the sequence images to obtain the HDR image.

Single-Focus and Dual Illumination under Small DoFs
As shown in Viewpoint ii (Figure 6c), while single-focus imaging with a small DoF, if single-illumination Condition Ca cannot be satisfied, a dual light-source combined image is required. First, as with Viewpoint i, Step 1 is performed with the depth distribution data (Figure 9a-c), and the object plane depth, DFS = 538 mm, is chosen to achieve the allin-focusing scheme.
Step 2 is performed next to obtain the pre-image single-illumination overexposure information from four light sources (Figure 9d,e). It can be seen that all single-light-source images suffer from overexposure. Hence, the ORI, SOI, and LC of the overexposed image (Table 3) are examined to determine the optimal combinations of two light sources under Condition Cb (Figure 9f). The black-filled area represents the overlapping overexposed area. A smaller area indicates a better final fusion effect. From the decision result, LC, in Table 4, the Tlight and Llight combination obtains the best high reflection suppression effect.
Step 3 is then performed; a DFS of 538 mm is chosen for the object plane, and multi- The final imaging scheme from Viewpoint i combines the full-focus scheme of Step 1 and the lighting scheme of Step 2, and D FS = 386 mm is selected as the object plane for HDR imaging under L light . The imaging consistency control parameters are listed in Table 1.
As shown in Figure 8b, the multi-exposure fusion algorithm [31] is used on the sequence images to obtain the HDR image.

Single-Focus and Dual Illumination under Small DoFs
As shown in Viewpoint ii (Figure 6c), while single-focus imaging with a small DoF, if single-illumination Condition Ca cannot be satisfied, a dual light-source combined image is required. First, as with Viewpoint i, Step 1 is performed with the depth distribution data (Figure 9a-c), and the object plane depth, D FS = 538 mm, is chosen to achieve the all-in-focusing scheme. For example, the image with a red color character represents the image obtained under Tlight, and the blue, orange, and purple colors represent Blight, Llight, and Rlight respectively. (f) Inferred images after "operation" between two different pre-imaging images, where the meaning of the color is similar to that shown in Figure 9e, and the area with black color represents the inferred area where the spots of different light sources coincide.  For example, the image with a red color character represents the image obtained under T light , and the blue, orange, and purple colors represent B light , L light , and R light respectively. (f) Inferred images after "operation" between two different pre-imaging images, where the meaning of the color is similar to that shown in Figure 9e, and the area with black color represents the inferred area where the spots of different light sources coincide.
Step 2 is performed next to obtain the pre-image single-illumination overexposure information from four light sources (Figure 9d,e). It can be seen that all single-light-source images suffer from overexposure. Hence, the ORI, SOI, and LC of the overexposed image (Table 3) are examined to determine the optimal combinations of two light sources under Condition Cb (Figure 9f). The black-filled area represents the overlapping overexposed area. A smaller area indicates a better final fusion effect. From the decision result, LC, in Table 4, the T light and L light combination obtains the best high reflection suppression effect. Step 3 is then performed; a D FS of 538 mm is chosen for the object plane, and multiexposure sequence images are obtained under the separate illuminations of T light and L light .
Lastly, in Step 4, multi-exposure fusion is performed according to the Viewpoint i steps to obtain the optimal HDR images under two light sources. Due to the character smoothness of wavelet fusion method in image processing [32], wavelet-domain overexposure fusion is then performed on the obtained HDR images while retaining low-level components ( Figure 10).

Multifocus and Dual Illumination under Large DoF
The first two imaging experiments used small DoFs; however, for large cases, the problem of an insufficient DoF may be encountered. As shown in Figure 6d, Viewpoint iii has large depth multifocus features (Combination IV), including high reflectivity under Viewpoint ii. The efficacy of the proposed illumination imaging strategy under complex multifocus and combined illumination conditions is demonstrated using the Viewpoint iii experiment via the following steps: 1. According to Step 1, the ranges of the current depth as detected by the depth sensor were found to be df = 247 and db = 878. The depth matrix and surface diagram are shown in Figure 11a,b. As Dfront −1 (df) < Dback −1 (db), the current focus mode is selected. Then, according to Equation (7), the depth distribution data are used to segment the current surface with depth-matching (Figure 11c). The assignment results for the object plane depth of the final all-in-focusing scheme are listed in Table 5. Unlike Viewpoints i and ii, Viewpoint iii requires depths of DFM1 = 584 mm, DFM2 = 350 mm, and DFM3 = 250 mm, as shown in Table 5. 2. Pre-imaging steps such as those in the previous experiments are performed to determine the lighting scheme. The pre-imaging and decision parameters are listed in Tables 1 and 3, respectively, and the lighting decision-making process and data are shown in Figure 11d-f and Table 6, respectively. By comparing the LC results in Ta-

Multifocus and Dual Illumination under Large DoF
The first two imaging experiments used small DoFs; however, for large cases, the problem of an insufficient DoF may be encountered. As shown in Figure 6d, Viewpoint iii has large depth multifocus features (Combination IV), including high reflectivity under Viewpoint ii. The efficacy of the proposed illumination imaging strategy under complex multifocus and combined illumination conditions is demonstrated using the Viewpoint iii experiment via the following steps: 1.
According to Step 1, the ranges of the current depth as detected by the depth sensor were found to be d f = 247 and d b = 878. The depth matrix and surface diagram are shown in Figure 12a,b. As D front , the current focus mode is selected. Then, according to Equation (7), the depth distribution data are used to segment the current surface with depth-matching (Figure 12c). The assignment results for the object plane depth of the final all-in-focusing scheme are listed in Table 6. Unlike Viewpoints i and ii, Viewpoint iii requires depths of D FM1 = 584 mm, D FM2 = 350 mm, and D FM3 = 250 mm, as shown in Table 6.

2.
Pre-imaging steps such as those in the previous experiments are performed to determine the lighting scheme. The pre-imaging and decision parameters are listed in Tables 1 and 3, respectively, and the lighting decision-making process and data are shown in Figure 12d-f and Table 5, respectively. By comparing the LC results in Table 5, the combination of T light and B light are selected as the required combination to provide lighting.

3.
As with Viewpoint ii, under T light and B light , the multi-exposure image sequence at object plane depths of D FM1 = 584 mm, D FM3 = 350 mm, and D FM2 = 250 mm is acquired.

4.
According to the fusion process shown in Figure 4b, the multi-exposure sequence images are first fused to obtain the HDR image according to the combination of each object plane's depth and light-source condition (Figure 11a,b). For the multifocus HDR images obtained from the same light source, the method in [17] is used to perform multifocus fusion to obtain the all-in-focus HDR image under Viewpoint iii, and a wavelet fusion operation like the one from Viewpoint ii is used on the all-in-focused HDR images under light sources of T light and B light to suppress the high reflectivity and restore the information of the overexposed area. The final fusion results are shown in Figure 11c.

Effect Evaluation of All-in-Focus Imaging with Overexposure Suppression
The performance quality of the detailed imaging surface information at various depths was subjectively and objectively evaluated for the all-in-focus fusion of large DoFs and the information recovery effects after overexposure suppression.

Evaluation of the All-in-Focus Imaging Effect
The post-all-in-focus fusion image should have better texture feature expression than any single image taken prior to fusion due to the improvements to global sharpness. We compared the results before and after fusion of the three focus images under Viewpoint and ii, Viewpoint iii requires depths of DFM1 = 584 mm, DFM2 = 350 mm, and DFM3 = 250 mm, as shown in Table 5. 2. Pre-imaging steps such as those in the previous experiments are performed to determine the lighting scheme. The pre-imaging and decision parameters are listed in Tables 1 and 3, respectively, and the lighting decision-making process and data are shown in Figure 11d-f and    For example, the image with a red color character represents the image obtained under the T light , and the blue, orange, and purple colors represent B light , L light , and R light respectively. (f) Inferred images after "operation". Images are inferred between two different pre-imaging images, where the meaning of the color is similar to that shown in Figure 12e, and the area with black color represents the inferred area where the spots of different light sources coincide.

Effect Evaluation of All-in-Focus Imaging with Overexposure Suppression
The performance quality of the detailed imaging surface information at various depths was subjectively and objectively evaluated for the all-in-focus fusion of large DoFs and the information recovery effects after overexposure suppression.

Evaluation of the All-in-Focus Imaging Effect
The post-all-in-focus fusion image should have better texture feature expression than any single image taken prior to fusion due to the improvements to global sharpness. We compared the results before and after fusion of the three focus images under Viewpoint iii with light source T light , as shown in Figure 11a, and we objectively and subjectively evaluated the all-in-focus effects.

Objective Evaluation of the All-in-Focus Effects
The objective evaluation calculated image sharpness on the basis of statistical image gradient and sharpness evaluations. The image gradient evaluation adopted the energy of gradient (EOG) and Tenengrad methods, and the Vollath function was chosen on the basis of the statistical results. Furthermore, the improvement in detail expression between the final image and each original image could be individually compared as a function of the signal-to-noise ratio (SNR), and peak signal-to-noise ratio (PSNR) [32].

EOG function
The EOG function takes the square sum of the differences in gray value between a pixel and the adjacent pixels in the xand y-directions as the gradient value of each, which are accumulated as input to the sharpness evaluation function. After averaging all pixels, the expression is 2

Tenengrad function
The Tenengrad function uses the Sobel operator to calculate the gradient value representing the sharpness of the image edge in the horizontal and vertical directions. After averaging the gradient values of all pixels, the expression is where Gx(x,y) and Gy(x,y) are the gradient values of the pixel in the horizontal and vertical directions at (x,y), respectively, and its formula is as follows: where ⊗ is the convolution symbol, and g x and g y are the horizontal and vertical templates of the Sobel operator, which are defined as g x = The Vollath autocorrelation function reflects the similarity between two points in space. The edges of the clear texture details in an image are clear and sharp, and the correlation between the pixels is low, whereas the texture details in the out-of-focus area are blurred, and the correlation between image pixels is high. The calculation result reflects the similarity of all adjacent pixels, thus evaluating overall image quality. The Vollath function is expressed as follows: 4 Signal-to-noise ratio 5 Peak signal-to-noise ratio where MSE is the mean square error, In the evaluation functions of Sections above, g(x, y) is the pixel value at (x, y) of the image in 1 -3 , p(x, y) is the pixel value at (x, y) of the original image in 4 -5 , q(x, y) is the pixel value at (x, y) of the fused image in 4 -5 , and X and Y are the numbers of rows and columns in the image pixel matrix, respectively. The calculation results are listed in Tables 7 and 8. The results of the objective evaluation show that, compared with the original images, those obtained from multifocus sequence fusion were significantly improved according to the sharpness quantification index, indicating that the overall sharpness of the fused images was significantly improved.

Subjective Evaluation of the Effects of All-in-Focus Fusion
Subjective evaluation was performed by directly observing the details and expressing the power of the images on the surface at each depth before and after all-in-focus imaging. Subsequently, the efficacy of the control strategy was ascertained.
As shown in Figure 13, there are three regions, back region (br), middle region (mr), and front region (fr), representing different depths of the surface. The details of the br, mr, and fr regions in the IMG FMj (j = 1, 2, 3) at different object plane depths, D FMj (j = 1, 2, 3), under Viewpoint iii were analyzed subjectively. Taking the horizontal comparison of the detailed image of area br in the red frame of Figure 13 as an example, image IMG FM1 had the most prominent detail expression at this depth, and the expressivities of IMG FM2 and IMG FM3 decreased with the decrease in object plane depth. The blurriness of image region br was positively correlated to the depth of the IMG FM1 , IMG FM2 , and IMG FM3 object planes. Meanwhile, the fused image, IMG Fused , had a similar detailed performance capability to IMG FM1 , and depth regions mr and fr followed similar rules. By comprehensively comparing the clarity laws of br, mr, and fr, it can be found that, when facing the inner wall of the irregular pipe in the depth range of 247-878 mm from Viewpoint iii, compared with IMG FMj after a single focus, which only represents the surface texture of a certain depth range, the fused image, IMG Fused , successfully distinguished surfaces at different depths with fine streaks and good performance, indicating that the imaging device and strategy effectively obtained pipeline surface information at various depths. IMGFMj after a single focus, which only represents the surface texture of a certain depth range, the fused image, IMGFused, successfully distinguished surfaces at different depths with fine streaks and good performance, indicating that the imaging device and strategy effectively obtained pipeline surface information at various depths.

Evaluation of Information Recovery Effect after Suppressing Overexposure
To evaluate the efficacy of lighting and information recovery for the overexposed area, an example was provided via the selection judgment of the lighting source chosen under Viewpoint i, which showed that the proposed control strategy effectively selected the most suitable light source, thus ensuring lighting and imaging quality. However, when the overexposure problem cannot be overcome by changing the light-source positions, a sequence of images with non-overlapping overexposed areas can be generated using two light sources at different positions. As shown in Figure 14, detailed texture information of the overexposed area was recovered by fusion. Figure 13. Comparison of results before and after all-in-focus.

Evaluation of Information Recovery Effect after Suppressing Overexposure
To evaluate the efficacy of lighting and information recovery for the overexposed area, an example was provided via the selection judgment of the lighting source chosen under Viewpoint i, which showed that the proposed control strategy effectively selected the most suitable light source, thus ensuring lighting and imaging quality. However, when the overexposure problem cannot be overcome by changing the light-source positions, a sequence of images with non-overlapping overexposed areas can be generated using two light sources at different positions. As shown in Figure 14, detailed texture information of the overexposed area was recovered by fusion.
area, an example was provided via the selection judgment of the lighting source chosen under Viewpoint i, which showed that the proposed control strategy effectively selected the most suitable light source, thus ensuring lighting and imaging quality. However, when the overexposure problem cannot be overcome by changing the light-source positions, a sequence of images with non-overlapping overexposed areas can be generated using two light sources at different positions. As shown in Figure 14, detailed texture information of the overexposed area was recovered by fusion. Figure 14. Comparisons of multifocus renderings. Figure 14. Comparisons of multifocus renderings. Figure 14 compares the details of the all-in-focus images under different light sources (T light and B light ) and after suppressing overexposure (Fused) in the same area and range. Considering the horizontal contrast between the detailed images under T light as an example (outlined in red), T light -P1, T light -P2, and T light -P3 are the high-reflection overexposure and transition areas under the light source and the non-overexposed area, respectively. The overexposed area loses the ability to express texture details owing to pixel saturation. As the area transitions to a non-overexposed area, the detailed expression gradually increases. By longitudinally comparing the texture details of X light -P1, X light -P2, and X light -P3 (X = T light , B light , and Fused) areas, it can be seen that, after overexposure suppression, the overexposed spots were removed after magnifying the image details after suppression. Hence, it can be clearly seen that the texture information (e.g., scratches and dents) were recovered well, demonstrating that the device has a good overexposure suppression capability. These results show that overexposed spots in highly reflective areas can be restored to their detailed information.
On the basis of these subjective and objective evaluations, the all-in-focus imaging and lighting strategy with overexposure suppression effectively guides the developed imaging device to obtain highly detailed images with fully expressed global detail information for fault detection.

Conclusions
In this study, a cross-modal all-in-focus imaging method with overexposure suppression was proposed, which realizes its function by obtaining depth and pre-imaging information. Compared with existing methods, our method is unaffected by surface textures and ambient light, and it directly determines the focus required for the depth of the object plane, resulting in an all-in-focus imaging capability that ensures the large span depth imaging effects with efficient focusing. The low coupling device and method do not depend on the shape of the imaging surface, and the camera characteristics solve the problem of regional overexposures on highly reflective surfaces. Hence, drastic changes in the effects of the final overexposure suppression are avoided, showing good robustness.
Imaging experiments on non-Lambertian free-form surfaces were performed under no ambient light conditions, demonstrating that the new system adaptively obtains clear all-in-focus images without overexposure under all depth span and curved surface conditions inside non-Lambertian-shaped pipes. The system is small and demonstrates low coupling and self-adaptation. It is suitable for free-scene imaging in any irregular cavity requiring active lighting, pipes of different diameters, and highly reflective free-form surfaces.
Although the method of all-in-focus imaging with overexposure suppression in this paper can obtain the sequence images effectively, the fusion process of sequence images introduced in this paper needs three fusion algorithms, which are inefficient. In the future, further research will focus on studying more efficient special fusion algorithms to further improve the efficiency and robustness of image fusion.  Institutional Review Board Statement: Not applicable.

Informed Consent Statement: Not applicable.
Data Availability Statement: All data and code will be made available on request to the correspondent author's email with appropriate justification.

Conflicts of Interest:
The authors declare no conflict of interest.