Next Article in Journal
An Improved Bio-Orientation Method Based on Direct Sunlight Compensation for Imaging Polarization Sensor
Previous Article in Journal
Analyzing Data Modalities for Cattle Weight Estimation Using Deep Learning Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Fusion of RGB and Depth Data with Image Enhancement

by
Lennard Wunsch
1,*,†,
Christian Görner Tenorio
1,†,
Katharina Anding
1,
Andrei Golomoz
1 and
Gunther Notni
1,2
1
Group of Quality Assurance and Industrial Image Processing, Faculty of Mechanical Engineering, Technische Universität Ilmenau, 98693 Ilmenau, Germany
2
Fraunhofer Institute for Applied Optics and Precision Engineering IOF Jena, 07745 Jena, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(3), 73; https://doi.org/10.3390/jimaging10030073
Submission received: 29 January 2024 / Revised: 12 March 2024 / Accepted: 19 March 2024 / Published: 21 March 2024
(This article belongs to the Section Image and Video Processing)

Abstract

:
Since 3D sensors became popular, imaged depth data are easier to obtain in the consumer sector. In applications such as defect localization on industrial objects or mass/volume estimation, precise depth data is important and, thus, benefits from the usage of multiple information sources. However, a combination of RGB images and depth images can not only improve our understanding of objects, capacitating one to gain more information about objects but also enhance data quality. Combining different camera systems using data fusion can enable higher quality data since disadvantages can be compensated. Data fusion itself consists of data preparation and data registration. A challenge in data fusion is the different resolutions of sensors. Therefore, up- and downsampling algorithms are needed. This paper compares multiple up- and downsampling methods, such as different direct interpolation methods, joint bilateral upsampling (JBU), and Markov random fields (MRFs), in terms of their potential to create RGB-D images and improve the quality of depth information. In contrast to the literature in which imaging systems are adjusted to acquire the data of the same section simultaneously, the laboratory setup in this study was based on conveyor-based optical sorting processes, and therefore, the data were acquired at different time periods and different spatial locations. Data assignment and data cropping were necessary. In order to evaluate the results, root mean square error (RMSE), signal-to-noise ratio (SNR), correlation (CORR), universal quality index (UQI), and the contour offset are monitored. With JBU outperforming the other upsampling methods, achieving a meanRMSE = 25.22, mean SNR = 32.80, mean CORR = 0.99, and mean UQI = 0.97.

1. Introduction

Computer vision using artificial intelligence is being integrated into many industrial processes to improve their performance by supervising the processes, collecting information for defect reduction [1,2], object localization [3], and obtaining more knowledge about the process. With better access to depth data and its addition to pre-installed computer vision systems, the range of applications of computer vision within quality insurance and decision-making has increased. In applications such as bridge inspection [1], railway quality insurance [2], agriculture [3,4], and robotics [5], RGB-D data are being utilized. By adding RGB data to depth data, the collection of shape-based information in addition to texture- and color-based information RGB-D data is generated. These data can ensure a deeper understanding of industrial processes and their defect sources. Therefore, applications such as approximate weight determination, defect size determination, defect localization, and additional support in decision-making can benefit from RGB-D data.
In order to create RGB-D data from a multi-view system, a data fusion algorithm is necessary to combine RGB and depth data. By fusing data, information from different sources are combined into one data object. However, a combination of these data can be challenging since different information sources can have different attributes. Data fusion can be beneficial not only in terms of additional information but also for an improvement in data quality, as shown, for example, by Okafor et al. [6] and Nemati et al. [7]. Given the capabilities of different systems, a smart combination of these systems can result in the compensation of the disadvantages of each system respectively. In order to process images with different resolutions and boundaries, more complex methods are needed. Boström et al. [8] collected multiple definitions of information fusion. The goals of information fusion are defined as “the provision of a better understanding of a given scene” and to “obtain information of greater quality”, among others. That is why RGB-D data have such vast potential when implemented in industrial processes for quality insurance.
Setting up a high-resolution sensor system for RGB-D data imaging can be expensive since at least two cameras are necessary, with one of them imaging the depth data while the other images RGB data, depending on the imaging techniques, and costs can become high. In order to obtain the precise RGB-D data from an analyzed sample, the sensor systems have to be either spatially adjusted, e.g., inclined under a certain angle towards each other to capture the same section of the sample at the same time, or temporally adjusted, e.g., not capturing the same sample section and adjusting the time shift due to different spatial acquisition positions via object tracking. The adjustment of the sensor systems is carried out via a co-ordinate system transformation of one type of data into the co-ordinate system of another system. Regarding sensor systems, there are many different setups for color- and shape-based imaging. Siepmann et al. [9] used a setup that utilized pattern projection. Another method with which to obtain RGB-D images is the combination of known depth imaging systems, such as the time of flight (ToF) [10,11,12,13], confocal microscopy [14], light detection and ranging (LIDAR) [15], or laser scanning [16,17,18] in combination with an RGB camera system.
Hach et al. [19] present an RGB-D camera containing a combination of a ToF sensor and an RGB camera. Both systems are combined in the system without inclination to each other. A typical system used in the literature [20,21,22] for a combination of a ToF sensor and an RGB camera is the Microsoft Kinect system. Another possibility for generating multimodal data is the combination of thermal data acquired by thermal infrared (TIR) cameras and depth imaging systems [16,17,23,24,25]. Regarding the adjustment of the sensor systems, different acquisition structures can be differentiated. These structures are either stationary, e.g., fixing the different imaging systems onto a stationary structure [12,13,15,16,17], or dynamic structures, e.g., a terrestrial scanner moving in the ambient air [15,18,20,23,26]. In order to summarize the state-of-the-art multimodal imaging for obtaining depth and color information, different depth-imaging systems are viable depending on the specific application.
Setting up an RGB-D imaging system can also become quite complex [27,28]. The more the camera parameters are the same, such as resolution, the easier it is to implement the data fusion algorithm. However, high-resolution depth cameras are more expensive than RGB cameras of the same resolution. Adjusting both cameras is also a challenging but necessary task. By calibrating both systems, the imaged regions are able to overlap and, thus, enable more precise data fusion. Therefore, a lot of thought needs to be put into setting up a viable imaging system for fused RGB-D data before data processing.
This work will analyze the required samples with a laboratory setup based on optical sorting processes [29,30]. Industrial processes based on optical sorting using a conveyor belt need to analyze a high quantity of samples in a short period of time. Therefore, only two depth acquisition methods, ToF and laser triangulation, respectively, should be considered for laser scanning in more detail. Other methods, such as interferometry, confocal microscopy, pattern projection, or photogrammetry, have a higher resolution than ToF or laser scanning methods but require more measuring time and come with greater costs. Due to the higher spatial resolution of a non-time-based measurement compared to ToF, this work uses a triangulation-based laser line scanning system as an effective depth imaging system. After setting up the imaging system, a fitting algorithm was needed for data fusion. In the past, a wide range of such data fusion methods have been proposed. Eichhardt et al. [11] give an overview of data fusion methods. A brief overview of different data fusion classifications, techniques, and architectures is given by Castanedo [31] and Elmenreich [32].
The fitting algorithm, also called data fusion, consists of data acquisition and preprocessing in case of different camera resolutions, as well as up or downsampling and the co-ordinate system transformation of one data type into the co-ordinate system of the reference system [33]. Usually, the system with the higher resolution is the preferred reference system; however, due to hardware limitations, this is not always viable. When transforming one image into the co-ordinate system of the reference system, false values can occur. These are pixel values that are seen by one system but not the other. The way these false values are dealt with is managed by the upsampling algorithm used to adjust the image resolution before transforming the co-ordinate system [34]. Therefore, this paper focuses on the variation of upsampling algorithms.
In the following sections, first, the data and imaging setup used in this work are introduced, and some background on data preprocessing is given; then, different upsampling methods, such as joint bilateral upsampling (JBU) [35,36] and Markov random fields (MRFs) [37,38], and different interpolation algorithms [39,40] are briefly described. For further reading on the matter of data fusion, Junger et al. [41] also compare multiple data fusion methods. In Section 4, the complete process of data acquisition, preprocessing, and data fusion, including post-processing, is described in detail. Afterward, a closer look at the evaluation process and its parameters is given. Finally, the results of each data fusion method applied to the same raw images are shown and evaluated based on the root mean square error (RSME), correlation (CORR), signal-to-noise ratio (SNR), universal quality index (UQI), and contour offset ( Δ Off p o s i t i o n ).

2. Data Fusion Process

Data fusion is the process of combining data from different sources [42]. By doing so, one can gain deeper knowledge and access combined information. The resolution, offset, and regions of interest (ROI) of the images must be adjusted by the data fusion algorithm so that each pixel of the RGB image gets a corresponding depth value in addition to the color values. This progress is difficult due to an offset in images both in the translational and rotational direction caused by different resolutions, non-commensurability, and missing or inconsistent data, as described by Illmann et al. [43] and Lahat et al. [44].
The main steps of data fusion are the following [33]:
  • Data acquisition and cropping;
  • Feature extraction;
  • Data registration;
  • Post-processing.
The ROI is defined by a cropping algorithm that crops the objects and works as a segmentation algorithm. In order to adjust the resolution of the data, the up- or downsampling and the combination of both must be realized. For this process, a reference image must be established before the resolution of the other image is adjusted. In this paper, the RGB image was chosen as the reference image for the fused RGB-D image. In addition to the higher resolution, it contains fewer defects and contains sharper imaged edges. The offset can be dealt with by translating the depth image based on information from an object registration algorithm. In order to accomplish this, the image-based data must be transformed into the same co-ordinate system to represent correlating areas. For this data registration, extracting the corresponding features in each image is necessary. Illmann et al. [43] provide further details and strategies for merging data. Feature extraction and data registration also deal with the time delay between both imaging systems. Therefore, both offset and time delay are solved.
This paper compares three simple interpolation methods with more complex methods, such as MRF and JBU, to achieve RGB-D-fused data. In the following, each upsampling method is presented in more detail.

2.1. Interpolation Methods

The classic interpolation methods used in this paper are given by Dianyuan [39] and Nischwitz et al. [40]. These methods were used as direct fusion methods to obtain corresponding depth pixel values for each respective RGB pixel. First, nearest-neighbor interpolation is chosen, where the unknown value of a pixel is calculated based on the interpolation of the pixel closest to the new pixel position. The second interpolation method used is a bi-linear interpolation, where four weight pixels are used for value calculation. The third and last interpolation method uses a cubic interpolation. Here, the unknown pixel is calculated using 16 surrounding pixels. Moreover, the interpolation methods are basic methods in image processing and are used in MRF and JBU as well.

2.2. Joint Bi-lateral Upsampling

Joint bi-lateral upsampling [35] is a guided depth upsampling method that uses bi-lateral filters to calculate a high-resolution depth map, D h . It is considered a local depth upsampling method [11] because the weight for the convolution is calculated based on the local pixel position. JBU uses the weighted convolution of a high-resolution RGB image, I h , with a lower-resolution depth map, D l . The weight can be calculated with a range filter, r, which is a Gaussian kernel with σ r = 0.5 , and the intensity differences of the high-resolution image, I h , and the low-resolution image, I l , at their pixel positions. The high-resolution depth map can be obtained with the simplified Equation (1) [45]:
D p h = q S w ( p , q ) · D q l q S w ( p , q ) , with w ( p , q ) = r · ( I p h I q l )
where S is the neighborhood of the pixel, p is the pixel position of the high-resolution data, and q is the pixel position of the low-resolution data. Since JBU is a guided upsampling method, to apply this method, the resolution of both data needs to be an integer multiple of each other. In order to apply JBU, the resolution difference between D p h and D q l needs to be an integer multiple of two. Moreover, Riemens et al. [45] proposed a multistep upsampling method to smoothen the edges. The multistep upsampling approach utilizes a cascade of 2 × 2 upsampling factors to reach factors of 4 × 4 or 8 × 8 .

2.3. Markov Random Fields

Depth upsampling by Markov random fields (MRFs) [37,38] is a guided, global method [11] because it is used as an optimization method to find the global optimum. It utilizes a Markov network to calculate the pixel values of the upsampled depth map based on weighted high-resolution image data and estimated upsampled depth data. With an iterative process, the current calculated depth map can be optimized by considering the high-resolution image data. Liu et al. [46] proposed a model that considers a data term and a smoothing term to calculate the high-resolution depth map. The data term calculates the quadratic differences between the real and estimated data, and the smoothing term calculates the given potential with a weight that depends on the given image data. Equation (2) [46] shows the proposed model:
D h = a r g m i n D ( 1 β ) · i p ( D i D i 0 ) 2 + β · i p j S w i , j c ( D i D j ) 2 ,
where β is an empirical factor, i and j are indices, and D 0 is the initial cubic interpolated high-resolution depth map of the low-resolution depth map. The weight is given in Equation (3) [46]:
w i , j c = e x p k C I l k I j k 2 6 · σ c 2 ,
where C represents the three RGB values, k is an index, and σ c is an empirical factor. Deriving Equation (2) [46] and setting it equal to zero leads to Equation (4) [46]:
D i n + 1 = ( 1 β ) · D i 0 + 2 · β · j S w i , j c · D j n ( 1 β ) + 2 · β · i S w i , j c ,
where n corresponds to the current iteration.

3. Hardware and Data Preprocessing

Industrial processes are usually stationary systems; therefore, the lab setup used for this work is also a stationary system. Since the integration of an imaging system into an industrial process can be expensive and complex, a conveyor belt system was chosen. These systems are widely utilized in industry and provide simpler setup options compared to free-falling objects. However, one should note that when imaging the depth information of free-falling objects, the complete object can be imaged when using conveyor belts, but the bottom of the object cannot be imaged due to the top-down view of the system [47]. Therefore, only a 2.5-dimensional image can be generated.
In order to image objects in the same scene in an industrial environment, a setup consisting of a feeding system and an imaging system was built. In the feeding system, two conveyor belts are moved at different speeds, and five plates are arranged horizontally along the second belt to separate the incoming objects. The imaging system is positioned along the second belt. This system consists of an RGB line camera and a 3D laser line scanner. As described in Section 1, for depth imaging, a 3D laser line scanner was chosen since, when compared to systems such as the time of flight system of Microsoft Kinect, it obtains a better overall resolution (see Refs. [48,49]). The 3D laser line scanner images a 3D point cloud. However, JBU, MRFs, and interpolation upsampling were proposed for the depth map data. Therefore, the imaged point cloud was transferred into a depth map.
Both cameras were chosen based on each other’s parameters to ease the data fusion process later on. Although imaging objects placed on a conveyor belt results in a shadow below the object due to illumination, this is a cheaper alternative. The shadow below the object does not exist with planar objects and is not relevant for the detection of defects on the surface of the object. The important attributes of each camera are shown in Table 1 [49,50]. The number of active pixels for the laser line scanner was reduced by binning due to the necessary speed of the converter belt to ensure the proper separation of the objects. Using the full scanner would have resulted in missing pixels in the depth image.
Since the cameras optical axes are horizontally shifted, the number of pixels covering the ROI (the conveyor belt) is not the same as the number of active pixels. According to the definition of ROI, both sensor systems lose some pixels at the edges. Therefore, the ratio between the used pixels of both systems is no longer exactly two; it is a decimal number. This is important for the use of JBU and will be explained in Section 2.2.
The laboratory setup by Anding et al. [29,30] is shown in Figure 1. The feeding system with object separation can be seen on the right, and the imaging area is next to the separation plates. Both cameras are positioned above the illumination system and imaging area. The separation of objects is implemented by the combination of two conveyor belts moving at different speeds and the array of separation plates, which are arranged orthogonal to the movement direction of the second conveyor belt. In order to accomplish homogeneous illumination for the reduction of object shadows, the objects were illuminated using two angled light sources from above and one light source below the conveyor belt. For depth data acquisition, no additional illumination was needed.
This study features natural objects such as stones from quarries. These complex objects were chosen to show the limits of the presented data fusion approach. Since they don’t have a planar surface, the missing depth data below the objects need to be considered. A coin was used as a simple known round object for the validation of each individual process. In order to avoid losing pixels while imaging, the line frequency is calculated automatically. In this way, the speed of the conveyor belt in the imaging path and the line frequency of both cameras can be adjusted. An incorrect line frequency results in compressed or stretched images. A round normal object and Equation (5) were used to find the correct line frequency, since the round normal object is imaged as an ellipse at incorrect line frequencies:
x 2 a 2 + y 2 b 2 = 1
With a and b describing both ellipsoid axles. These parameters are calculated using elliptical Hough transformation. Figure 2 shows examples of the objects of the featured dataset and the coin used for validation.
By using a 3D laser line scanner, the depth images are subjected to artifacts called depth shadows (along the edges), as seen in Figure 3. These artifacts are an inaccuracy, as can be seen from the depth value reading zero in the image, causing a bigger offset between the RGB image and depth data. Removing these artifacts is necessary to fuse the images accurately. With the help of interpolation and averaging over neighboring pixels, in addition to a threshold-based edge detection algorithm, the removal of the artifacts was achieved. Using this algorithm, the edges of the depth shadow were removed, and, therefore, the edges of the depth data were improved, resulting in sharper edges closer to the real edges of the imaged object.

4. Data Acquisition and Processing

In order to understand the complete process, different levels of processing needed to be considered. First, one needs to understand what happens in data acquisition. This process is shown in Figure 4. In order to take a closer look at the data fusion process in the grey area in Figure 4, the flowchart in Figure 5 is shown.
Since the camera systems are positioned along the conveyor belt behind each other, there is a temporal difference, Δ t , in acquisition. Since both data need to be present to combine them, the processing of both data needs to be delayed until both data are acquired. For segmentation, a cropping algorithm is used. This, however, can be carried out right after acquisition.
An overview of the data fusion process is shown in Figure 5. Before the preprocessing algorithm is used, the incoming samples of the conveyor belt are cropped to remove the background. Afterward, the preprocessing algorithm is used on the acquired samples. On the one hand, the RGB data need to be synthesized to create an image of the full object without the RGB artifacts caused by the conveyor belt. Meanwhile, the depth shadows mentioned in Figure 3 are removed from the 3D data.
Next, resolution matching is implemented. This step uses different up- and downsampling steps to increase the resolution of the depth data and fit the RGB data accordingly. First, a combination of down- and upsampling methods based on classical interpolation methods is applied to the RGB data to create a resolution difference between RGB and depth data of a factor of 2 × 2.
The downsampling is necessary due to the ROI issue, as described in Section 3. The resolution of both data is not exactly an integer multiple of each other. The RGB data were chosen as reference data due to their higher source resolution. So, the RGB data were first downsampled to the resolution of the depth data. In the second step, they are upsampled by a factor of two to meet the requirement for guided upsampling methods such as JBU and MRF.
Next, the depth data need to be assigned to their corresponding RGB data. This is carried out by feature extraction and feature matching. Features such as area, convexity, etc., are extracted from the data and assigned over difference tolerances. The features used are the following:
  • Geometric dimensions: height, width, and area;
  • Circumference;
  • Center of gravity;
  • Convexity;
  • Upper-edge position.
After the assignment, the JBU, MRFs, and classical interpolation methods are finally applied, respectively, to match the RGB data resolution and the depth data resolution. Since the RGB data resolution was set to double the depth data resolution in the previous steps, the depth data were upsampled by a factor of 2 × 2. For classical interpolation methods, the RGB data are not used. However, since JBU and MRFs are guided upsampling methods, RGB data with double the resolution are necessary (see Section 2.2 and Section 2.3).
As a last step, data registration is carried out. First, the co-ordinate system of the RGB data and the 3D data must be transformed into the same co-ordinate system to ensure an offset-free fusion process. With the transformation of the co-ordinate system, the fusion itself can be carried out. After fusion, a morphological operation, more precisely, closing, and a threshold to minimize the created noise are executed. In order to obtain a fully overlapping fusion of depth data and RGB data, points outside of the matching edges have to be eliminated by applying an RGB contour mask. Pixels inside the edges have to be interpolated to extend to the boundaries.

5. Evaluation Process

In order to be able to compare the different methods used for upsampling in terms of their performance for data fusion, different evaluation methods were used. For evaluation, the RMSE, CORR, SNR, and UQI based on the work of Jagalingam et al. [51] and Naidu et al. [52] are used. In addition to these assessment parameters, an offset between the depth data and RGB data was used. N and L are the number of pixels in each spatial dimension, and I corresponds to the intensity of the pixels at positions i and j of the depth data or RGB data, respectively.
RMSE measures the quadratic error between the source images, the RGB and upsampled depth data, and the fused RGB-D image, with zero being the best possible score. Equation (6) [51,52] can be used to calculate the RMSE:
RSME = 1 N · L · i = 1 N j = 1 L I binary , upsampled , depth ( i , j ) I binary , RGB ( i , j ) 2
where I binary , RGB represents the intensity of the image acquired by the RGB line camera, and I binary , upsampled , depth is the intensity calculated using the acquired depth data and the respective upsampling method. In order to measure the information similarities between the source images and the fused image SNR, which can be seen in Equation (7) [51,52], this equation can be used. The higher the score, the better the method has performed:
SNR = 20 · log 10 i = 1 N j = 1 N I binary , upsampled , depth ( i , j ) 2 i = 1 N j = 1 N ( I binary , upsampled , depth ( i , j ) I binary , RGB ( i , j ) ) 2 ,
CORR is the correlation coefficient given in Equations (10) and (11) [51,52]. It is used as an evaluation method to measure the similarity of an RGB or depth image and an RGB-D image, with 1 being a perfect score:
C depth , RGB = i = 1 N j = 1 L I binary , upsampled , depth ( i , j ) · I binary , RGB ( i , j ) ,
C depth = i = 1 N j = 1 L I binary , upsampled , depth ( i , j ) 2 ,
C RGB = i = 1 N j = 1 L I binary , RGB ( i , j ) 2 ,
CORR = 2 · C depth , RGB C depth · C RGB
UQI is the universal quality index given in Equation (12) [51,52]. It is used to describe the information brought into the fused image, with 1 being the best possible score.
UQI = 4 · σ d e p t h , R G B · ( μ d e p t h + μ R G B ) ( σ d e p t h 2 + σ R G B 2 ) · ( μ d e p t h 2 + μ R G B 2 ) , with
  • μ being the mathematical expected value, respectively;
  • σ being the standard deviation, respectively.
The offset, Δ Off p o s i t i o n , is measured to evaluate the overlap of the object maps between the source and fused images. Equation (13) gives the if statement to calculate the offset:
Δ Off p o s i t i o n = + 1 if e d g e R G B ( i , j ) e d g e d e p t h ( i , j ) , 0 otherwise .
These assessment parameters are used to compare and evaluate the upsampling methods presented in Section 6.

6. Results

In total, more than 225 data fusion experiments using 11 different objects (10 stones and the validation coin) and upsampling parameters were conducted. Table 2, Table 3 and Table 4 show the mean of the evaluation parameters for each of the upsampling methods presented in Section 2. Based on the results, Table 5 shows a comparison using the best-performing parameters. The mean was calculated considering all analyzed objects. Due to computational costs, the upsampling of the factors to 2 × 2 the resolution of the RGB data was analyzed.
For MRFs, 30 different parameter scenarios were conducted to find the best parameters. Table 2 shows the mean evaluation results of Markov random field upsampling using a cubic interpolation algorithm [46] since the performance of the MRFs using nearest neighbor and bi-linear interpolation did not perform as well. σ c or β were varied while the other parameters were fixed. The variation in the parameters was chosen arbitrarily. A total of 60 iterations were carried out; for the majority of the dataset, this quantity of iterations was sufficient, and the performance was saturated. When comparing the negative and positive values of σ c , equal results were achieved. However, varying the sign of β does show an effect. Based on this, the best parameters were chosen for upsampling using MRFs. In the first step of the experiments, σ c = 1 was set while β was varied, with β = 0.99 achieving the best performance. For β = 0.99 , σ c was varied in the second step of experiments. With this arrangement, σ c = 17 255 performs the best. A third iteration of experiments was conducted to find the best fitting β value for σ c = 17 255 . While σ c = 17 255 and β = 10 performed best, σ c = 17 255 and β = 0.99 also performed well. Since these parameters are recommended by [46], choosing these parameters ensures comparability with other publications.
Table 3 displays the evaluation results for different up- and downsampling scenarios for each interpolation method. Direct downsampling and upsampling were achieved in one process without any intermediate sampling stages, whereas in other scenarios, the resolution was adjusted in multiple cascading processes. In each scenario, nearest neighbor and bi-linear interpolation outperform cubic interpolation. Moreover, the best result was achieved by iterating the downsampling process by a factor of 2 × 2 with nearest neighbor interpolation.
Table 4 gives an overview of the JBU performance for different upsampling scenarios. All evaluation parameters are the means and standard deviations of multiple objects. Overall, one can see JBU performs best with a nearest neighbor interpolation. Moreover, downsampling the RGB data by a factor of 4 × 4 and later upsampling the data to the depth data resolution results in the best performance in terms of the assessment parameters chosen in this work.
Based on two example objects, all the evaluation values are given in Table 5. For each upsampling method, the best-performing parameter setup is shown. For Markov random fields, n = 60 iterations were used. Δ O f f P o s i t i o n is used in addition to RSME, CORR, and UQI to show the overlap of the RGB image and the depth map in the fused RGB-D image. Due to the cheap computational costs, JBU and interpolation fusion have an advantage over MRFs because of the iterations MRFs have to undergo in order to optimize the results. JBU clearly outperforms MRFs and interpolation in this work. Due to the guided fusion over a convolutional weight, JBU shows good performance in the fusion process. Nearest neighbor outperformes the bi-linear and cubic interpolation and is, therefore, more benefitial for preprocessing.
Naidu et al. [52] also achieved C O R R = 0.99 . Moreover, the authors of [53] report correlation coefficients between 0.95 and 0.99 for various data fusion methods. Zhang [54] achieved a maximum of C O R R = 1 by modifying a multispectral image; however, other methods presented by [54] perfrom worse, with correlation coefficients between C O R R m i n = 0.72 and C O R R m a x = 0.91 . When comparing the evaluation metrics RMSE, UQI, SNR, and CORR of JBU with Zhu et al. [53], Naidu et al. [52], and Zhang [54], we can see that JBU performance is similar and comparable. However, one must note that Zhu et al. [53] and Zheng [54] fused the spectral images, e.g., panchromatic, RGB, or multispectral images, whereas, in our work, the spatial and spectral data in the form of RGB and depth data were fused.

7. Conclusions

This publication compares different upsampling methods for their use in data fusion based on complex objects. Three different processes—classical interpolation methods, MRFs, and JBU—are compared. First, the best parameter for each method was determined before they were compared. By applying JBU, downsampling the RGB data by a factor of 4 × 4, and later upsampling the data to the depth data resolution, as well as using nearest neighbor interpolation, the other methods were outperformed.
Figure 6 shows the RGB-D point cloud of the validation coin. One can see the texture on the coin combined with the height profile, with a total diameter of 23 m m ; the height resolution matches the texture resolution provided by the RGB image.
Figure 7 shows the resulting RGB-D point cloud for the two example stones, which were retrieved using the described JBU setup. By using a 3D laser line scanner and an RGB line camera, high-resolution RGB-D images were created. This enables volume approximation and the detailed location of textual and form-based features. However, there are also flaws seen in Figure 7b. Figure 7b shows noisy depth points due to the rough or reflective topography of the analyzed object, which leads to measurement errors in the depth measurement due to the scattering of the laser beam. Smoothing over this region via the neighborhood would lead to fewer deviations.
Figure 8 shows the results of the different steps of data processing, for example, stone number one. Figure 8a shows the depth map of the cropped depth data, whereas Figure 8b shows the edge correlation between the RGB image and the depth map. After minimizing the artifacts of the depth shadow, a blurry edge at the bottom of Figure 8a can be seen. This noisy region is due to the depth shadows. Figure 8c shows the results of the synthesized depth map of Figure 8a. The synthesizing step uses a threshold defined by the histogram of the depth map in order to minimize these blurry edges. Figure 8d shows an improvement in correlation due to the exclusion of depth shadows. Figure 8e,f depict the result of the JBU data fusion process. Using JBU increases the correlation between the depth data and RGB data but leads to the synthesized RGB image having a few noise residuals, which are visible in the blurry bottom left part of Figure 8e. Feature extraction of the RGB image is used in order to create an edge mask to cut out these blurry residuals and overlap both images. With this step, the correlation of the edges increases even further, as depicted in Figure 8g,h. For the final interpolation of the missing depth map edges with the neighborhood of the missing data, the edge matching of the RGB data and depth data can be achieved.
To conclude this work, data fusion was used for image enhancement by removing artifacts and improving the depth map edge quality by removing the depth shadow and matching the edges of the RGB data and depth data. Moreover, different interpolation methods for RGB data and depth map fusion are compared for multiple assessment parameters. The parameters of each interpolation method were varied to find the best-performing settings. The results show JBU outperforming MRFs and the direct interpolation methods. This work shows the capabilities of RGB-D imaging with temporarily separated imaging systems. The setup-based time delay between both imaging systems was solved using feature extraction and data registration. By imaging RGB data and depth data separately along a conveyor belt system and fusing additional information, the quality of the imaged data was improved.
In future works, other data fusion methods will be compared, including the novel polygon-based approach of triangle-mesh-rasterization projection (TMRP) [41]. Moreover, the spectral dimension in this work is confined to RGB, and this will be expanded by using multispectral and hyperspectral imaging systems in addition to the RGB camera system.

Author Contributions

Conceptualization, writing, visualization, validation, and supervision, L.W.; methodology, experiments, algorithms/software, visualization, and writing, C.G.T.; review, and supervision, K.A.; Conceptualization and hardware support A.G.; review and supervision, G.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Projektträger Jülich, grant numbers FKZ03RU1U151C, FKZ03RU1U152C, and FKZ03RU1U153C, as part of the project RUBIN AMI.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data and software that support the findings of this study are available from the corresponding author, [L.W.], upon request.

Acknowledgments

We thank Christina Junger and Galina Polte for their support, feedback, and discussions during our research. Furthermore, we would like to thank Christina Junger for reviewing this paper. We acknowledge support for the publication costs by the Open Access Publication Fund of the Technische Universität Ilmenau.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wojcik, B.; Zarski, M. The maesurements of surface defect area with an RGB-D camera for BIM-backed bridge inspection. Bull. Pol. Acad. Sci. Tech. Sci. 2021, 69, e137123. [Google Scholar]
  2. Ge, X.; Qin, Y.; Cao, Z.; Gao, Y.; Lian, L.; Bai, J.; Yu, H. A Fine-Grained Method for Detecting Defects of Track Fasteners Using RGB-D Image. In Proceedings of the 6th International Conference on Electrical Engineering and Information Technologies for Rail Transportation (EITRT) 2023, Beijing, China, 19–21 October 2023; pp. 37–44. [Google Scholar]
  3. Fu, L.; Gao, F.; Wu, J.; Li, R.; Karkee, M.; Zhang, Q. Application of consumer RGB-D cameras for fruit detection and localization in field: A critical review. Comput. Electron. Agric. 2020, 17, 105687. [Google Scholar] [CrossRef]
  4. Skoczeń, M.; Ochman, M.; Spyra, K.; Nikodem, M.; Krata, D.; Panek, M.; Pawłowski, A. Obstacle Detection System for Agricultural Mobile Robot Application Using RGB-D Cameras. Sensors 2021, 21, 5292. [Google Scholar] [CrossRef] [PubMed]
  5. Jing, C.; Potgieter, J.; Noble, F.; Wang, R. A comparison and analysis of RGB-D cameras depth perfromance for robotics application. In Proceedings of the 2017 24th International Conference on Mechatronics and Machine Vision in Practice (M2VIP), Auckland, New Zealand, 21–23 November 2017; pp. 1–6. [Google Scholar]
  6. Okafor, N.U.; Potgieter, Y.; Alghorani, F.; Delaney, D.T. Improving Data Quality of Low-cost IoT Sensors in Environmental Monitoring Networks Using Data Fusion and Machine Learning Approach. ICT Express 2020, 6, 220–228. [Google Scholar] [CrossRef]
  7. Nemati, S.; Malhotra, A.; Clifford, C.D. Data Fusion for Improved Respiration Rate Estimation. EURASIP J. Adv. Signal Process. 2021, 2010, 220–228. [Google Scholar] [CrossRef]
  8. Boström, H.; Brohede, M.; Johansson, R.; Karlsson, A.; van Laere, J.; Niklasson, L.; Nilsson, M.; Persson A., S.; Ziemke, T. On the Definition of Information Fusion as a Field of Research; Institutionen för Kommunikation och Information: Skövde, Sweden, 2007; pp. 1–8. [Google Scholar]
  9. Siepmann, J.; Heinze, M.; Kühmstedt, P.; Notni, G. Pixel synchronous measurement of object shape and color. In Proceedings of the SPIE Optical Engineering + Applications, San Diego, CA, USA, 2–6 August 2009; Volume 7432. [Google Scholar]
  10. Qui, D.; Pang, J.; Sun, W.; Yang, C. Deep End-to-End Alignment and Refinement for Time-of-Flight RGB-D Module. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 29 October–1 November 2019; pp. 9994–10003. [Google Scholar]
  11. Eichhardt, I.; Chetverikov, D.; Jankó, Z. Image-guided ToF depth upsampling: A survey. Mach. Vis. Appl. 2017, 28, 267–282. [Google Scholar] [CrossRef]
  12. Hastedt, H.; Luhmann, T. Investigations on a combined RGB/time-of-flight approach for close range applications. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2012, 39, 333–338. [Google Scholar] [CrossRef]
  13. Van den Bergh, M.; Van Gool, L. Combining RGB and ToF Cameras for Real-time 3D Hand Gesture Interaction. In Proceedings of the 2011 IEEE Workshop on Applications of Computer Vision (WACV), Kona, HI, USA, 5–7 January 2011; pp. 66–72. [Google Scholar]
  14. Siemens, S.; Kästner, M.; Riethmeier, E. RGB-D microtopography: A comprehensive dataset for surface analysis and characterization techniques. Data Brief 2023, 48, 109094. [Google Scholar] [CrossRef]
  15. Ming, J.L.-C.; Armenakis, C. Fusion of optical and terrestrial laser scanner data. Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci. 2010, 38, 156–161. [Google Scholar]
  16. Hoegner, L.; Abmayr, T.; Tosic, D.D.; Turzer, S.; Stilla, U. Fusion of 3D Point Clouds with TIR Images for Indoor Scene Reconstruction. Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci. 2018, XLII-1, 189–194. [Google Scholar] [CrossRef]
  17. Gleichauf, J.; Vollet, J.; Pfitzner, C.; Koch, P.; May, S. Sensor Fusion Approach for an Shunting Locomotive. In Informatics in Control, Automation und Robotics (ICINCO 2017); Lecture Notes in Electrical Engineering; Springer: Cham, Switzerland, 2019; Volume 495, pp. 603–624. [Google Scholar]
  18. Ishikawa, R.; Roxas, M.; Sato, Y.; Oishi, T.; Masuda, T.; Ikeuchi, K. A 3D Reconstruction with High Density and Accuracy using Laser Profiler and Camera Fusion System on a Rover. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 620–628. [Google Scholar]
  19. Hach, T.; Steurer, J. A Novel RGB-Z Camera for High-Quality Motion Picture Applications. In Proceedings of the 10th European Conference on Visual Media Production, London, UK, 6–7 November 2013; pp. 1–10. [Google Scholar]
  20. Budzan, S.; Kasprzyk, J. Fusion of 3D laser scanner and depth images for obstacle recognition in mobile applications. Opt. Lasers Eng. 2016, 77, 230–240. [Google Scholar] [CrossRef]
  21. Dahan, M.; Chen, N.; Shamir, A.; Cohen-Or, D. Combining color and depth for enhanced image segmentation and retargeting. Vis. Comput. 2012, 28, 1181–1193. [Google Scholar] [CrossRef]
  22. Vijayanagar, K.; Loghman, M.; Kim, J. Real-Time Refinement of Kinect Depth Maps using Multi-Resolution Anisotropic Diffusion. Mob. Netw. Appl. 2014, 19, 414–425. [Google Scholar] [CrossRef]
  23. Gleichauf, J.; Pfitzner, C.; May, S. Sensor Fusion of a 2D Laser Scanner and a Thermal Camera. In Proceedings of the International Conference on Informatics in Control, Automation and Robotics (ICINCO) 2017, Madrid, Spain, 26–28 July 2017; pp. 398–400. [Google Scholar]
  24. Landmann, M.; Heist, S.; Dietrich, P.; Lutzke, P.; Gebhart, I.; Templin, J.; Kühmstedt, P.; Tünnermann, A.; Notni, G. High-Speed 3D Thermography. Opt. Lasers Eng. 2019, 121, 448–455. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Müller, S.; Stephan, B.; Gross, H.-M.; Notni, G. Point Cloud Hand-Object Segmentation Using Multimodal Imaging with Thermal and Color Data for Safe Robotic Object Handover. Sensors 2021, 21, 5676. [Google Scholar] [CrossRef] [PubMed]
  26. Dabek, P.; Jaroslaw, J.; Radoslaw, Z.; Wodecki, J. An Automatic Procedure for Overheated Idler Detection in Belt Conveyors Using Fusion of Infrared and RGB Images Acquired during UGV Robot Inspection. Energies 2022, 15, 1–20. [Google Scholar] [CrossRef]
  27. Amamara, A.; Aouf, N. Real-time multiview data fusion for object tracking with RGBD sensors. Robotica 2016, 34, 1855–1879. [Google Scholar] [CrossRef]
  28. Petitti, A.; Vulpi, F.; Marani, R.; Milella, A. A self-calibration approach for multi-view RGB-D sensing. In Multimodal Sensing and Artificial Intelligence: Technologies and Applications II; Stella, E., Ed.; International Society for Optics and Photonics SPIE: Bellingham, WA, USA, 2021; Volume 117850C. [Google Scholar]
  29. Anding, K.; Garten, D.; Linß, G.; Pieper, G.; Linß, E. Klassifikation Mineralischer Baurohstoffe mittels Bildverarbeitung und Maschinellem Lernen. In Proceedings of the 16th Workshop “Farbbildverarbeitung” 2010, Ilmenau, Germany, 7–8 October 2010; 2010. Available online: http://germancolorgroup.de/html/Vortr_10_pdf/14_FarbWS2010_GesteinserkennungEND1_8_148-155.pdf (accessed on 18 March 2023).
  30. Anding, K.; Garten, D.; Göpfert, A.; Rückwardt, M.; Reetz, E.; Linß, G. Automatic Petrographic Inspection by using Image Processing and Machine Learning. In Proceedings of the XX IMEKO World Congress, Metrology for Green Growth, Busan, Republic of Korea, 9–14 September 2010. [Google Scholar]
  31. Castanedo, F. A Review of Data Fusion Techniques. World Sci. J. 2013, 2013, 704504. [Google Scholar] [CrossRef]
  32. Elmenreich, W. An Introduction to Sensor Fusion; Vienna University of Technology: Vienna, Austria, 2002; Volume 502, pp. 1–28. [Google Scholar]
  33. Kolar, P.; Benavidez, P.; Jamshidi, M. Survey of Datafusion Techniques for Laser and Vision Based Sensor Integration for Autonomous Navigation. Sensors 2020, 20, 2180. [Google Scholar] [CrossRef]
  34. Park, J.; Kim, H.; Tai, Y.-W.; Brown, M.S.; Kweon, I.S. High-Quality Depth Map Upsampling and Completion for RGB-D Cameras. IEEE Trans. Image Process. 2014, 23, 5559–5572. [Google Scholar] [CrossRef]
  35. Kopf, J.; Cohen, M.; Lischinski, D.; Uyttendaele, M. Joint bilateral upsampling. ACM Trans. Graph. 2007, 26, 96–101. [Google Scholar] [CrossRef]
  36. Ren, Y.; Liu, J.; Yuan, H.; Xiao, Y. Depth Up-Sampling via Pixel-Classifying and Joint Bilateral Filtering. KSII Trans. Internet Inform. Syst. 2018, 12, 3217–3238. [Google Scholar]
  37. Lu, J.; Min, D.; Pahwa, R.S.; Do, M.N. A revisit to MRF-based depth map super-resolution and enhancement. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 985–988. [Google Scholar]
  38. Diebel, J.; Thurn, S. An Application of Markov Random Fields to Range Sensing. In Proceedings of the 18th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 5–8 December 2005; pp. 291–298. [Google Scholar]
  39. Dianyuan, H. Comparison of Commonly Used Image Interpolation Methods. In Proceedings of the 2nd International Conference on Computer Science and Electronics Engineering (ICCSEE 2013), Hangzhou, China, 22–23 March 2013; pp. 1556–1559. [Google Scholar]
  40. Nischwitz, A.; Fischer, M.; Haberäcker, P.; Socher, G. Bildverarbeitung—Band II des Standardwerks Computergrafik und Bildverarbeitung; Springer: Wiesbaden, Germany, 2020. [Google Scholar]
  41. Junger, C.; Buch, B.; Notni, G. Triangle-Mesh-Rasterization-Projection (TMRP): An Algorithm to Project a Point Cloud onto a Consistent, Dense and Accurate 2D Raster Image. Sensors 2023, 23, 7030. [Google Scholar] [CrossRef] [PubMed]
  42. Bleiholder, J.; Naumann, F. Data Fusion. ACM Comput. Surv. 2009, 41, 1–41. [Google Scholar] [CrossRef]
  43. Illmann, R.; Rosenberger, M.; Notni, G. Strategies for Merging Hyperspectral Data of Different Spectral and Spatial Resolution. In Proceedings of the Digital Image Computing: Techniques and Applications (DICTA), Canberra, ACT, Australia, 10–13 December 2018; pp. 1–7. [Google Scholar]
  44. Lahat, D.; Adalý, T.; Jutten, C. Challenges in multimodal data fusion. In Proceedings of the D2014 22nd European Signal Processing Conference (EUSIPCO), Lisbon, Portugal, 1–5 September 2014; pp. 101–105. [Google Scholar]
  45. Riemens, A.; Gangwal, O.P.; Barenbrug, B.; Berretty, R.-P.M. Multistep joint bilateral depth upsampling. In Proceedings of the IS&T/SPIE Electronic Imaging, San Jose, CA, USA, 18–22 January 2009; Volume 7257. [Google Scholar]
  46. Liu, W.; Jia, S.; Li, P.; Chen, X.; Yang, J.; Wu, Q. An MRF-Based Depth Upsampling: Upsample the Depth Map With Its Own Property. IEEE Signal Process. Lett. 2015, 22, 1708–1712. [Google Scholar] [CrossRef]
  47. Garten, D.; Anding, K.; Linß, G.; Brückner, P. Automatische Besatzanalyse mittels Bildverarbeitung und maschinellem Lernen. In Proceedings of the 16th Workshop “Farbbildverarbeitung” 2010, Ilmenau, Germany, 7–8 October 2010; Available online: http://germancolorgroup.de/html/Vortr_10_pdf/16_Anding_Garten_QualiKorn_GFE_11_170-180.pdf (accessed on 18 March 2023).
  48. Microsoft Learn. Azure Kinect DK Hardware Specifications. Available online: https://learn.microsoft.com/en-us/azure/kinect-dk/hardware-specification (accessed on 15 November 2023).
  49. Micro-Epsilon. High-Performance Laser-Scanners. Available online: https://www.micro-epsilon.co.uk/2D_3D/laser-scanner/scanCONTROL-3000/ (accessed on 15 November 2023).
  50. JAI. CV-L107 CL - 3 CCD RGB Line Scan Camera. Available online: https://www.1stvision.com/cameras/models/JAI/CV-L107CL (accessed on 12 September 2023).
  51. Jagalingam, P.; Hegde, A.V. A Review of Quality Metrics for Fused Image. Aquat. Procedia 2015, 4, 133–142. [Google Scholar] [CrossRef]
  52. Naidu, V.P.S.; Raol, J.R. Pixel-level Image Fusion using Wavelets and Principal Component Analysis. Def. Sci. J. 2008, 58, 338–352. [Google Scholar] [CrossRef]
  53. Zhu, X.X.; Bamler, R. A Sparse Image Fusion Algorithm with Application to Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
  54. Zhang, Y. Methods for Image Fuson Quality Assessment—A Review, Comparison And Analysis. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2008, XXXVII, 1101–1109. [Google Scholar]
Figure 1. Setup used for 3D and RGB imaging.
Figure 1. Setup used for 3D and RGB imaging.
Jimaging 10 00073 g001
Figure 2. Example objects utilized: (a) stone-A, (b) stone-B, and the (c) validation coin.
Figure 2. Example objects utilized: (a) stone-A, (b) stone-B, and the (c) validation coin.
Jimaging 10 00073 g002
Figure 3. Results of stone one (object). (a) False color images of depth shadows imaged by the 3D laser line scanner (height legend: yellow > 0 m m , and purple = 0 m m ) and the (b) improved depth map.
Figure 3. Results of stone one (object). (a) False color images of depth shadows imaged by the 3D laser line scanner (height legend: yellow > 0 m m , and purple = 0 m m ) and the (b) improved depth map.
Jimaging 10 00073 g003
Figure 4. Flowchart of the data acquisition process.
Figure 4. Flowchart of the data acquisition process.
Jimaging 10 00073 g004
Figure 5. Flowchart of the data fusion process.
Figure 5. Flowchart of the data fusion process.
Jimaging 10 00073 g005
Figure 6. RGB-D point cloud of the validation coin fused by JBU.
Figure 6. RGB-D point cloud of the validation coin fused by JBU.
Jimaging 10 00073 g006
Figure 7. RGB-D point cloud after data fusion, for example, stones-A (a) and -B (b) fused by JBU.
Figure 7. RGB-D point cloud after data fusion, for example, stones-A (a) and -B (b) fused by JBU.
Jimaging 10 00073 g007
Figure 8. Results of object stone-A. False color image depth maps during the fusion process (a,c,e,g) and the edge correlation of the depth map and RGB image (b,d,f,h). (a,b) show the process result of the cropped depth data, with minimized depth shadows; (c,d) show the process result of the depth data after the synthesis; (e,f) show the process result after JBU data fusion; (g,h) show the process result after the feature extraction process.
Figure 8. Results of object stone-A. False color image depth maps during the fusion process (a,c,e,g) and the edge correlation of the depth map and RGB image (b,d,f,h). (a,b) show the process result of the cropped depth data, with minimized depth shadows; (c,d) show the process result of the depth data after the synthesis; (e,f) show the process result after JBU data fusion; (g,h) show the process result after the feature extraction process.
Jimaging 10 00073 g008
Table 1. Attributes for each imaging system used in the system [49,50].
Table 1. Attributes for each imaging system used in the system [49,50].
AttributeRGB Camera3D Laser Line Scanner
JAI CV-L107 CLScan CONTROL 3000-200
Imaging profileRGB line cameraLasertriangulation
No. of active pixels 3   ×   2048 1   ×   1024
Pixel size14 μ m × 14 μ m 26 μ m × 26 μ m
Table 2. Mean and standard deviation of MRFs with different parameters and cubic interpolatiosn for 60 iterations.
Table 2. Mean and standard deviation of MRFs with different parameters and cubic interpolatiosn for 60 iterations.
ParameterRMSESNRCORRUQI
σ c = 1 β = 0 53.545 ± 9.16 16.99 ± 4.05 0.92 ± 0.04 0.91 ± 0.03
β = 0.25 51.75 ± 8.69 17.65 ± 4.13 0.93 ± 0.034 0.91 ± 0.03
β = 0.5 51.01 ± 8.42 17.92 ± 4.11 0.93 ± 0.03 0.91 ± 0.03
β = 0.75 50.53 ± 8.19 18.01 ± 4.08 0.93 ± 0.03 0.91 ± 0.03
β = 1 137.00 ± 20.63 nannan 0.67 ± 0.1
β = 10 63.45 ± 38.25 17.45 ± 3.35 0.93 ± 0.03 0.86 ± 0.14
β = 2 51.12 ± 7.51 15.80 ± 6.79 0.82 ± 0.29 0.91 ± 0.02
β = 1 86.94 ± 27.20 10.20 ± 3.74 0.79 ± 0.13 0.83 ± 0.08
β = 10 57.72 ± 9.66 15.49 ± 2.78 0.91 ± 0.03 0.90 ± 0.03
β = 0.99 50.29 ± 8.12 18.18 ± 4.07 0.93 ± 0.03 0.91 ± 0.03
β = 0.99 σ c = 0.5 48.62 ± 4.91 18.68 ± 3.63 0.94 ± 0.03 0.92 ± 0.02
σ c = ± 17 255 51.05 ± 8.12 17.89 ± 4.05 0.93 ± 0.03 0.91 ± 0.03
σ c = ± 1 50.23 ± 8.11 18.18 ± 4.07 0.93 ± 0.03 0.91 ± 0.03
σ c = ± 10 90.22 ± 41.43 12.26 ± 7.29 0.80 ± 0.14 0.76 ± 0.17
σ c = 0.25 51.20 ± 8.09 17.83 ± 4.04 0.93 ± 0.03 0.91 ± 0.03
σ c = ± 40 138.60 ± 36.33 6.22 ± 2.35 0.66 ± 0.12 0.61 ± 0.17
σ c = 17 255 β = 0 53.54 ± 9.16 16.99 ± 4.05 0.92 ± 0.04 0.91 ± 0.03
β = 0.25 52.13 ± 8.72 17.50 ± 4.09 0.92 ± 0.04 0.91 ± 0.03
β = 0.5 51.59 ± 8.44 17.70 ± 4.07 0.93 ± 0.03 0.91 ± 0.03
β = 0.75 51.19 ± 8.26 17.84 ± 4.05 0.93 ± 0.03 0.91 ± 0.03
β = 1 137.00 ± 20.63 nannan 0.67 ± 0.10
β = 10 50.86 ± 8.07 17.95 ± 4.05 0.93 ± 0.03 0.91 ± 0.03
β = 2 50.90 ± 8.13 17.94 ± 4.05 0.93 ± 0.03 0.91 ± 0.03
β = 1 137.00 ± 20.63 nannan 0.67 ± 0.10
β = 10 51.05 ± 8.11 17.87 ± 4.13 0.93 ± 0.03 0.91 ± 0.03
β = 0.99 51.05 ± 8.20 17.89 ± 4.05 0.93 ± 0.03 0.91 ± 0.03
Table 3. Mean and standard deviation of different single-interpolation methods and parameters.
Table 3. Mean and standard deviation of different single-interpolation methods and parameters.
DS * and US * Scenario and Interpolation MethodRMSESNRCORRUQI
Direct DS*nearest neighbor 46.81 ± 8.39 18.98 ± 3.86 0.94 ± 0.02 0.91 ± 0.031
bi-linear 52.78 ± 9.79 16.77 ± 4.27 0.92 ± 0.04 0.89 ± 0.04
cubic 53.19 ± 9.29 16.60 ± 4.00 0.92 ± 0.04 0.89 ± 0.04
Direct DS *-factor 2 × 2nearest neighbor 47.33 ± 9.97 18.71 ± 3.94 0.94 ± 0.03 0.89 ± 0.05
bi-linear 52.08 ± 9.23 16.37 ± 4.00 0.92 ± 0.04 0.88 ± 0.05
cubic 59.99 ± 7.36 12.32 ± 2.96 0.89 ± 0.03 0.85 ± 0.04
DS *-factor 2 × 2nearest neighbor 45.03 ± 8.66 19.53 ± 3.93 0.94 ± 0.02 0.90 ± 0.04
bi-linear 50.98 ± 10.88 16.86 ± 4.58 0.92 ± 0.04 0.87 ± 0.06
cubic 54.98 ± 8.02 13.89 ± 2.99 0.90 ± 0.03 0.87 ± 0.04
DS *-factor 4 × 4nearest neighbor 47.85 ± 11.86 19.00 ± 4.92 0.93 ± 0.04 0.88 ± 0.06
bi-linear 49.63 ± 9.80 17.27 ± 4.01 0.92 ± 0.04 0.84 ± 0.09
cubic 52.93 ± 8.64 14.57 ± 2.58 0.91 ± 0.02 0.86 ± 0.06
Direct DS *-factor 4 × 4nearest neighbor 51.20 ± 12.04 17.78 ± 4.38 0.93 ± 0.04 0.86 ± 0.07
bi-linear 49.41 ± 10.35 17.37 ± 3.85 0.93 ± 0.03 0.86 ± 0.07
cubic 57.73 ± 8.83 13.02 ± 2.95 0.89 ± 0.03 0.82 ± 0.05
Direct US *nearest neighbor 55.32 ± 9.73 15.98 ± 4.13 0.91 ± 0.04 0.91 ± 0.03
bi-linear 53.45 ± 8.64 15.81 ± 3.68 0.92 ± 0.03 0.91 ± 0.03
cubic 64.94 ± 7.54 10.92 ± 2.80 0.87 ± 0.04 0.89 ± 0.02
US *-factor 2 × 2nearest neighbor 55.32 ± 9.73 15.98 ± 4.13 nan 0.93 ± 0.02
bi-linear 52.50 ± 8.64 15.82 ± 3.68 nan 0.93 ± 0.02
cubic 69.42 ± 7.67 8.78 ± 2.84 nan 0.90 ± 0.02
* DS: Downsampling to the resolution of the depth data; US: Upsampling to the resolution of the RGB data.
Table 4. Mean and standard deviation of JBU with different interpolation methods and parameters.
Table 4. Mean and standard deviation of JBU with different interpolation methods and parameters.
VariantRMSESNRCORRUQI
Direct US *nearest neighbor 40.93 ± 9.07 21.49 ± 4.89 0.95 ± 0.02 0.94 ± 0.02
bi-linear 71.59 ± 8.81 12.43 ± 3.10 0.86 ± 0.05 0.76 ± 0.06
cubic 69.48 ± 7.83 12.91 ± 3.10 0.87 ± 0.05 0.69 ± 0.08
DS * and US *-factor 2 × 2nearest neighbor 38.35 ± 10.60 23.03 ± 5.37 0.96 ± 0.03 0.93 ± 0.05
bi-linear 83.26 ± 9.81 8.96 ± 3.87 0.75 ± 0.12 0.63 ± 0.12
cubic 84.59 ± 10.52 8.72 ± 3.74 0.74 ± 0.12 0.55 ± 0.16
DS * and US *-factor 4 × 4nearest neighbor 25.22 ± 3.44 32.80 ± 9.62 0.99 ± 0.02 0.97 ± 0.05
bi-linear 118.61 ± 26.90 10.28 ± 6.62 0.74 ± 0.17 0.55 ± 0.29
cubic 124.63 ± 22.01 9.25 ± 5.42 0.73 ± 0.15 0.53 ± 0.27
Procedure similar to that in [21]nearest neighbor 33.17 ± 9.19 25.68 ± 6.05 0.97 ± 0.02 0.96 ± 0.02
bi-linear 89.72 ± 8.27 9.05 ± 2.24 0.77 ± 0.07 0.77 ± 0.05
cubic 85.15 ± 8.31 9.74 ± 2.40 0.79 ± 0.06 0.69 ± 0.06
US *-factor 4 × 4 similar as in [21]nearest neighbor 44.18 ± 9.26 20.15 ± 4.62 nan 0.96 ± 0.02
bi-linear 62.32 ± 8.84 14.46 ± 3.55 nan 0.91 ± 0.03
cubic 57.22 ± 9.22 15.86 ± 3.83 nan 0.90 ± 0.03
DS *: Downsampling to the resolution of the depth data; US *: Upsampling to the resolution of the depth data.
Table 5. Overview of the presented upsampling method performances for two example stones.
Table 5. Overview of the presented upsampling method performances for two example stones.
Example StoneRMSESNRCORRUQI Δ Off position
Stone-A
Interpolation NN 41.00 19.67 0.95 0.92 4.69 %
MRF (n * = 60) 43.23 19.55 0.95 0.93 2.86 %
JBU 11.40 42.46 1.00 0.99 0.25 %
Stone-B
Interpolation NN 61.29 14.75 0.91 0.84 8.83 %
MRF (n * = 60) 53.26 18.28 0.94 0.90 3.99 %
JBU 11.64 44.60 1.00 0.99 0.24 %
* n: number of iterations; NN: nearest neighbor.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wunsch, L.; Tenorio, C.G.; Anding, K.; Golomoz, A.; Notni, G. Data Fusion of RGB and Depth Data with Image Enhancement. J. Imaging 2024, 10, 73. https://doi.org/10.3390/jimaging10030073

AMA Style

Wunsch L, Tenorio CG, Anding K, Golomoz A, Notni G. Data Fusion of RGB and Depth Data with Image Enhancement. Journal of Imaging. 2024; 10(3):73. https://doi.org/10.3390/jimaging10030073

Chicago/Turabian Style

Wunsch, Lennard, Christian Görner Tenorio, Katharina Anding, Andrei Golomoz, and Gunther Notni. 2024. "Data Fusion of RGB and Depth Data with Image Enhancement" Journal of Imaging 10, no. 3: 73. https://doi.org/10.3390/jimaging10030073

APA Style

Wunsch, L., Tenorio, C. G., Anding, K., Golomoz, A., & Notni, G. (2024). Data Fusion of RGB and Depth Data with Image Enhancement. Journal of Imaging, 10(3), 73. https://doi.org/10.3390/jimaging10030073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop