You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

18 April 2025

Enhancing Three-Dimensional Reconstruction Through Intelligent Colormap Selection

,
and
1
Department of Management Science and Technology, Democritus University of Thrace, 65404 Kavala, Greece
2
Center for Research and Technology Hellas, Information Technologies Institute, 57001 Thessaloniki, Greece
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Underwater Vision Sensing System: 2nd Edition

Abstract

Photogrammetry is the process of creating three-dimensional (3D) reconstructions from two-dimensional images. In this paper, a novel method is introduced that significantly enhances 3D reconstruction by improving image quality through a combination of RGB stretching, Contrast Limited Adaptive Histogram Equalization (CLAHE), and colormaps. This approach effectively increases the number of keypoints and matches between images, resulting in more accurate and detailed 3D reconstructions. Additionally, a heuristic methodology is proposed to identify the optimal colormaps for each dataset based on keypoint matches between image pairs. This approach saves valuable time by recommending the most effective colormaps, eliminating the need to test each individually. A new dataset comprising two real-world image collections from underwater cultural heritage sites is presented to validate the algorithm, characterized by challenging environments such as low visibility and varying lighting conditions. The limitations of existing techniques are addressed by this method, providing a robust solution for enhancing image quality in demanding underwater scenarios. Experimental results show that image enhancement can lead to a 7.91% and 11.4% improvement, and the enhancement fusion with colormaps improves reconstruction accuracy by 10.82% and 64.11%. These findings render the proposed methodology a promising tool for 3D reconstruction in challenging environments, like underwater.

1. Introduction

Structure-from-Motion (SfM) is a photogrammetric technique that reconstructs 3D structures from a series of 2D images taken from different viewpoints. It is crucial in fields such as computer vision, where understanding the geometry of an environment is essential for tasks like navigation and object recognition. It plays a crucial role in various fields, including archaeology, where it helps preserve cultural heritage [1], human digitization [2], and virtual reality experiences [3]. Modern techniques leverage advancements in Structure-from-Motion (SfM) and Multi-View Stereo (MVS) to achieve high accuracy and detail. In [4], a comprehensive analysis and refinement of the traditional SfM techniques was presented, and a methodology named COLMAP was introduced. The methodology addressed common challenges in SfM, such as handling large-scale datasets and improving robustness in diverse and complex scenes. Another popular methodology is OpenSfM [5], a solution designed with scalability and modularity, offering flexible integration with other services, making it popular for projects that prioritize extensibility and adaptability. SFM methodologies often face challenges when processing degraded images. To address these limitations, approaches such as [6,7] focus on improving image illumination and enhancing visual quality, ensuring that the images retain more meaningful information for further analysis.
In this paper, a novel methodology to enhance 3D reconstruction by fusing pre-processing techniques such as RGB stretching and CLAHE with colormaps is proposed. By utilizing colormaps, more features can be identified, leading to better image stitching and more accurate 3D reconstructions. The proposed enhancement methodology showed a rise of 7.91% and 11.4%, while colormaps increased the features by 10.82% and 64.11% for the first and second experimental datasets, respectively. Additionally, a heuristic approach for selecting the optimal colormaps for each dataset is introduced. Although various methodologies exist for enhancing 3D reconstruction, none combine colormaps with color enhancement to address the issue of light absorption, a common challenge in underwater environments. The main innovations presented in this paper include:
  • A novel methodology that combines colormaps and color enhancement methods to improve 3D reconstruction;
  • A heuristic approach identifying the optimal colormaps for each dataset;
  • A fusion of preprocessing techniques that enhance 3D reconstruction;
  • An evaluation of 3D reconstruction and the effects of colormaps and enhancement techniques using real-world datasets.
Initially, the dataset images are enhanced using a fusion of CLAHE and RGB stretching. These enhanced images are then processed by a developed methodology that selects two random images, applies 22 colormaps, and counts the number of feature matches between them. The colormap with the fewest matches is eliminated, and the process continues with two new random images and the remaining colormaps until the optimal colormap is identified. Once the best colormap for the dataset is determined, the enhanced images are transformed using this colormap and processed with OpenSfM for 3D reconstruction. Importantly, the final reconstruction step utilizes the original, unprocessed images, ensuring that while image matches are maintained, the resulting 3D model retains its original colors, which is crucial.
In contrast to conventional SfM methodologies, which are robust but struggle with low-contrast or light-absorbing environments, this approach introduces a unique fusion of RGB stretching and CLAHE with an array of colormaps. This fusion enhances subtle image features, typically lost in challenging conditions, such as underwater scenes. By systematically evaluating 22 different colormaps and applying a heuristic selection process, the method adapts the enhancement to each specific dataset, ensuring the optimal balance between feature richness and computational efficiency. Moreover, processing the final 3D reconstruction using the original images preserves the authentic color information, which is critical for applications in cultural heritage, human digitization, and virtual reality. These enhancements collectively address key limitations of current techniques and offer a flexible, scalable, and more accurate solution for 3D reconstruction in diverse imaging conditions.
The remainder of this paper is as follows: Section 2 provides a concise overview of related work. Section 3 outlines the details of the proposed 3D reconstruction framework. Section 4 discusses the methodology evaluation, including the relevant metrics used. Finally, Section 5 presents the conclusions of the study, along with its limitations and suggestions for future improvements.

3. Enhanced 3D Reconstruction Through Intelligent Colormap Selection

3.1. Image Enhancement

The first step involved enhancing the image colors. After evaluating four different enhancement methods and their combinations, it was found that the fusion of RGB stretching [17] together with CLAHE [18] significantly improved feature matching and surpassed the others. RGB stretching was applied by expanding the histograms of the red, green, and blue channels, thereby increasing image contrast and distributing pixel intensity levels. This technique is particularly beneficial for underwater scenarios, where sunlight penetration differs greatly from terrestrial environments. For instance, red light is absorbed within the first 5 m, orange at 10 m, yellow at 20 m, green at 30 m, and blue at 60 m. By applying RGB stretching, these color absorption challenges are addressed, ensuring better image clarity and color representation. The formula used for RGB stretching is presented below:
P o u t = P i n   L m i n V m a x V m i n L m a x L m i n + V m i n
where Pin and Pout are the input and output pixels, respectively. Lmax and Lmin represent the maximum and minimum intensity level values for the input image, and Vmin and Vmax denote the intensity values for the output images.
The other method used to enhance the image contrast is the CLAHE. It is a method employed in image processing to enhance the local contrast and improve the image quality. Unlike traditional histogram equalization methods, CLAHE divides the image into smaller regions named tiles and then performs histogram equalization on each one. Dividing the image into tiles allows for adaptive adjustments based on local contrast characteristics. The image noise is reduced by limiting the contrast enhancement in each tile. The approach ensures that bright and dark areas receive enhancement and have a better visual appearance. It is particularly effective in scenarios of poor lighting conditions, which is the case in underwater environments. It is also effective in images containing a wide range of intensity levels. It is a valuable tool in various fields, including medical imaging and underwater photography.
The fusion of the two techniques starts by stretching the images’ RGB channels and then applying the CLAHE method. The result is an image with better quality that allows the detection of more features. An indicative example of the aforementioned color enhancement fusion is visualized in Figure 1. The figure features an image from a real-world underwater dataset. The image depicts the amphorae of Peristera, located at a depth of 30 m near Skopelos. After the fusion of CLAHE and RGB stretching, the application of colormaps follows, and its formula is presented below:
  I c o l o r m a p   ( x , y ) = C I e n h a n c e d   ( x , y )
where Icolormap (x,y) is the image coming from the colormap application after the fusion of CLAHE and RGB stretching. x and y represent the width and height values, while Ienhanced is the enhanced image, and C(.) is the colormap function that maps grayscale values to colors.
Figure 1. The color enhancement visualization.
It is important to note that first, the image turns to grayscale, and then the colormap is applied. This method ensures that the colormap enhances the intensity variations meaningfully. Colormaps map grayscale intensity values to colors using predefined schemes. For instance, the HSV colormap does not perform a color space conversion but instead assigns colors to intensity values based on the HSV color space. The result is a smooth transition through hues, providing a visually appealing representation of the grayscale data.

3.2. Efficient Colormap Selection Using Heuristics

After enhancing the dataset images with preprocessing techniques, a heuristic algorithm was employed to find the optimal colormaps. This systematic method iteratively evaluates and refines the set of candidate colormaps based on their performance. By progressively narrowing down the colormap set and eliminating the worst-performing ones at each iteration, the heuristic effectively guides the search toward an optimal solution, which is the best colormap for the dataset in this case. Heuristic techniques are used in optimization problems where finding the optimal solution is impractical or impossible within a reasonable time. Instead, heuristics aim to efficiently find good solutions, even if they may not guarantee the best solution.
The heuristic methodology begins by randomly selecting an image from the dataset and then choosing its sequential image. This approach reduces the risk of failing to find matches, especially when the dataset comes from sequential video frames. If they are not sequential, it will still function by identifying fewer matches. For the initial pair of images, 22 colormaps [31] were applied, and the number of matches for each colormap was counted. The colormaps used were autumn, bone, jet, winter, rainbow, ocean, summer, spring, cool, hsv, pink, hot, parula, magma, inferno, plasma, viridis, cividis, twilight, twilight_shifted, turbo, and deepgreen. These colormaps were chosen to provide a varied yet concise selection that balances the number of possibilities with the need for effective presentation. This collection ensured ease of use and simplicity. While reducing the number of examined colormaps may impact stitching accuracy, it also lowers computational costs. Performance and efficiency have to be balanced in the limited options.
Initially, the colormap with the worst performance, based on image matches, was removed. The remaining colormaps were stored in a list. Next, a new image was selected randomly, along with its sequential image. These images were processed with the remaining colormaps, and the results were added to the existing results to update the overall classification. The colormap with the worst performance was removed again. This process was repeated until the best colormap was identified. The process is presented in Algorithm 1. The initial evaluation was based on the formula below:
Algorithm 1: Heuristic Approach for Finding the Optimal Colormaps
Data: Images, cmaps_list, cmaps_to_reduce
01:
    Set total_results to an empty list
02:
    for n in range(len(cmaps_list), 0, -cmaps_to_reduce):
03:
            Randomly select an image and its sequential pair from the dataset
04:
            Enhance the images using a fusion of RGB stretching and CLAHE
05:
             Apply each colormap from cmaps_list to the images and calculate the number of keypoint matches, storing results in colormaps_df
06:
            Append the results from colormaps_df to total_results
07:
            Sort total_results to identify the best colormaps
08:
            Remove the worst-performing colormap from cmaps_list
09:
            Write intermediate results to a text file
10:
    return The best colormaps from the sorted total_results
C 1 ,   C 2 , , C n = T o p n { E ( I , C i ) C i C }
where the Topn (.) function selects the top n colormaps based on the evaluation metric Ε, which is the match number between two consecutive images. C represents the colormaps, and Ci denotes the i-th colormap from the available colormaps set. I denotes the input image, and is E the evaluation metric, which quantifies the quality of the resulting image after applying the colormap C to image I. Then the iterative selection follows the formula below:
C b e s t = a r g m a x C C 1 ,   C 2 , , C 21 E ( I , C )
where Cbest denotes the best colormap and argmax is a function that returns the top n colormaps that were given as input. For each iteration, the top colormaps follow the (2) and continue the iterative process until the top-asked colormaps are found.
The flowchart of the heuristic algorithm, shown in Figure 2, begins with an input image dataset. It then randomly selects an image and its consecutive counterpart, applying various colormaps and counting the matches between them for each colormap. The evaluation results are stored in a list, which is then sorted. The algorithm iteratively removes the worst-performing colormap until the best colormap is identified. Finally, the entire image dataset is processed using this optimal colormap.
Figure 2. Flow chart of the proposed heuristic methodology.

3.3. Enhancing 3D Reconstruction with Colormaps

Once the best colormap was identified in terms of keypoint matches between frames, it transformed the images, which were then given as input to the 3D reconstruction. The methodology for 3D reconstruction was based on OpenSfM and began with extracting metadata from the images. This metadata included image dimensions and, if available, GPS coordinates, capture time, projection type, and focal ratio. The next step was detecting image features, which were improved using colormaps and the proposed color-enhancing method. After feature detection, matches between images were identified. The methodology first determined the list of image pairs to analyze and then performed matching detection for each pair to find corresponding feature points. This process can be time-consuming due to the large number of possible image pairs, but the presence of GPS coordinates or capture time can significantly reduce the time required.
Once matches between image pairs were identified, the matches were linked to build feature track points. A track is a collection of feature points from different images identified as corresponding to the same physical point. The reconstruction process then began, aiming to determine the 3D positions of the tracks along with the camera positions, which represent the motion. Next, triangular meshes were added to the reconstruction to enable smooth transitions between viewers. Corrected versions of the images were created, removing lens distortions and preparing the data for depth map computation. In the next step, a detailed 3D point cloud was generated from the undistorted images and reconstructions. The final result of this methodology was the point cloud. This process utilized enhanced images only for keypoint detection to improve matching, while the unprocessed images were used to create the final point cloud. The procedure is illustrated in Figure 3.
Figure 3. The architectural framework of the SfM methodology illustrating the key computational steps and processes involved.
After enhancing the image, the process involved applying a specific colormap to the dataset images. This step is crucial as it enables the keypoint detection methodology to identify more keypoints, leading to better matches between images. After applying the colormap matches, calculations started, and reconstruction followed. It is important to note that the reconstruction should be based on the original images, not the colormap-applied ones. The colormap is only used to increase the keypoint matches. Reconstructing the initial images increases the number of matches and enhances the quality of the 3D reconstruction. This improvement occurs because the dataset remains the same, but the images are better stitched together. The whole process of heuristic finding the most appropriate colormaps, together with the reconstruction, is presented in Figure 4.
Figure 4. The architecture of the proposed methodology.
While colormaps are utilized in image matching, they play a crucial role in the proposed method for generating the initial point cloud by significantly enhancing feature extraction and overall reconstruction accuracy. It is important to emphasize that colormaps are used solely for image matching. The overall reconstruction is performed using the initial images, which retain all the original details without incorporating colormaps. This contribution goes beyond preprocessing, addressing key limitations in existing methods, as demonstrated by the experimental results presented in the paper.

3.4. Implementation Details

The Scale-Invariant Feature Transform (SIFT) [32] was used to detect keypoints. In addition to SIFT, other methods such as Speeded Up Robust Features (SURF) [33], Accelerated KAZE (AKAZE) [34], Hessian Affine feature point detector combined with HOG descriptor (HAHOG) [35], and Oriented FAST and Rotated BRIEF (ORB) [36] were also tested. SIFT was chosen due to its balanced combination of accuracy and efficiency. The scale independence of SIFT is highly advantageous for the heuristic algorithm, as it allows images to be downscaled, thereby reducing computational costs and increasing processing speed without compromising the accuracy. However, any other keypoint detection method can be used with the proposed methodology. The CLAHE hyperparameters were tuned with a tile size of (4, 4) and clip 2.
The experiments were implemented with no problems using four threads. OpenSfM is primarily CPU-based, with its core Structure-from-Motion optimization relying on the Ceres Solver, which does not support GPU acceleration. Additionally, the depth map estimation techniques are also CPU-intensive and do not require or utilize GPU hardware for processing. Therefore, OpenSfM does not have GPU acceleration needs and operates fully on the CPU. All experiments were conducted on a system equipped with an Intel i5-10600K CPU, 16 GB of RAM, and an Nvidia RTX 1660 GPU. The equipment used in the experiments was sourced from the United States. Specifically, the system included components manufactured by Intel Corporation and NVIDIA Corporation, both headquartered in Santa Clara, California, USA.

4. Experimental Evaluation

Experiments were conducted to assess the effectiveness of various colormaps and their impact on 3D reconstruction. Initially, a heuristic methodology was employed to validate whether colormaps identified as optimal truly exhibited superior performance. Subsequently, reconstructions were performed using 22 different colormaps. The experimental evaluation utilized real-world datasets to ensure assessment under realistic conditions. This Section presents a detailed account of these experiments along with the corresponding results.

4.1. Dataset

The first dataset (dataset A) consisted of 280 underwater images captured near Skopelos at a depth of 30 m. These images of both datasets depict the amphorae of Peristera, originating from a shipwreck dated to the last quarter of the 5th century BC based on typological characteristics. The site features a dense concentration of amphorae, forming a low hill measuring 25 m in length and 12 m in width, and contains approximately 1000 amphorae. The total number of load amphorae was estimated at 4200, which is the largest in size and capacity of the classical period that has been discovered to date.
The second dataset (dataset B) comprised 200 underwater images collected from a site near the location of the first dataset. This new dataset shared similar depth characteristics with the initial one but featured a large amphora surrounded by several smaller ones. Figure 5 visualizes the dataset, presenting one reconstruction for each dataset along with selected images used in creating the 3D model.
Figure 5. Reconstruction of two different underwater scenes.

4.2. Results

The experiments on dataset A were the first to be conducted. For brevity, the best results from the 22 colormaps are presented in Table 2, showcasing the top-performing ones. For this dataset, the best colormaps were the bone, jet, rainbow, cool, and twilight_shifted. It is important to note that while reconstruction points provide additional information, more points do not necessarily result in better stitching. The non-enhanced images detected an average of 12,094 features per image. In contrast, the enhanced images showed a significant improvement, detecting 13,051 features, an increase of 7.91%. Using the colormap bone, the number of detected features rose to 13,403, representing a 10.82% increase. The OpenSfM methodology processes unprocessed images for 3D reconstruction and reconstructs the 3D model based on them, while OpenSfM-C is an enhanced version that utilizes colormaps to improve feature detection before reconstructing the 3D model from the original images. The specific colormap used in each case is indicated in parentheses.
Table 2. 3D model reconstruction results for dataset A.
To evaluate the impact of colormaps on the reconstruction process, it is essential to compare OpenSfM-C with the standard OpenSfM method without colormaps. This direct comparison allows isolation and assesses the contribution of colormaps in enhancing feature detection and overall reconstruction quality. Comparing OpenSfM-C with entirely different reconstruction methods would primarily highlight differences between methodologies rather than isolating the specific effect of colormap usage. To accurately assess how the proposed approach that utilizes colormaps contributes to the reconstruction process, it is essential to compare OpenSfM-C with the standard OpenSfM method without colormaps. This approach ensures that the evaluation focuses on the impact of colormaps rather than broader variations in reconstruction techniques.
Following the initial experiments, dataset B, the second experimental dataset, demonstrated even more pronounced improvements with colormaps utilization compared to dataset A. The initial non-enhanced images in dataset B had an average of 13,297 features per image. Enhancing these images increased the feature count to 14,813, marking an 11.40% improvement. The rainbow colormap achieved the best result, with 21,823 features, representing a remarkable 64.11% increase. In dataset B, most colormaps significantly increased the number of features, highlighting the effectiveness of colormaps in enhancing reconstruction features, particularly in underwater scenarios. The top dataset B colormaps were the jet, summer, rainbow, hsv, and twilight_shifted. Some of them (jet, rainbow, twilight_shifted) were in both datasets, while HSV and summer were at the top only for dataset B. The top results are shown in Table 3.
Table 3. 3D model reconstruction results for dataset B.
Each dataset underwent 24 reconstructions, with various implementation factors tested. Several keypoint detection methods were evaluated, with SIFT emerging as the fastest while still detecting numerous keypoints. The colormaps’ utilization often increased the detected keypoints number, with each reconstruction taking approximately an hour for 200–280 input images. Additionally, the reconstruction was performed using the initial images, ensuring that the color differences introduced by the colormaps did not affect the final result. The colormaps assisted solely in the stitching process, leading to a 3D model with better-stitched images and enhanced details. Although colormaps slightly extended the processing time due to the higher number of detected keypoints, this increase was minor compared to the significant match improvement. While colormaps inherently enhance data and contribute to improved feature detection, the results also demonstrate an increase in reconstructed points. This indicates that not only are more features being detected, but these additional features contribute to a denser and more informative point cloud. The increase in reconstructed points suggests that the enhancement introduced by colormaps leads to a more complete and detailed representation of the scene, supporting the claimed contribution. A histogram comparison is provided in Figure 6 to visually illustrate the differences between the results for datasets A and B.
Figure 6. Visual representation of results for datasets A and B.
For those seeking to maintain accuracy without extended reconstruction times, sampling techniques can be combined with colormaps to achieve similar results with fewer images. This approach will be explored in future work. In Figure 7, a visualization of different colormaps applied to two distinct datasets is shown. This figure demonstrates how colormaps revealed details that were difficult to see in the original images, enabling SIFT to detect more keypoints than in the initial images. In addition to testing all the colormaps to see if they could enhance the reconstruction process, an approach to identify the best colormap for each dataset was developed.
Figure 7. Visualization of the top colormaps for each dataset.
Several methodologies were explored, but were not integrated in the end. Initially, a method was implemented that involved taking two random continuous images and applying each colormap to them while counting the matches. Although this method quickly identified which colormap might fit best, it relied on only two images and could potentially mislead results since the reconstruction process depends on all dataset images. To address this, a heuristic approach was implemented. This approach first tested all 22 colormaps and then iteratively removed the worst-performing ones until the best colormap was identified. A cumulative results list was maintained to further refine this process, where the match number from each new test was added to the previous results. This method reduced the chances of dismissing a colormap that performed poorly on one image but yielded good results on others.
The heuristic approach results for dataset A are shown in Table 4. Four different experiments were conducted, and each time, a different colormap emerged on top due to the randomness in the selected image pairs from the dataset. The table includes the colormap names that achieved the highest overall scores in terms of matches, with the first row representing the best colormap and the rest following in descending order. It is observed that in the first experiment, which included the top colormaps in the top positions, the total matches were higher than in the other experiments. This indicates that these colormaps achieved the highest results and should be prioritized for testing.
Table 4. Heuristic colormap selection approach results for dataset A.
Notably, the jet colormap consistently ranked the second best for reconstruction, while rainbow and twilight_shifted followed in third and fourth place, respectively. These colormaps appeared in every experiment set, indicating that this methodology can reliably recommend suitable colormaps to users. The hsv colormap, which topped three experiments, did not yield good results in reconstruction. The slight inconsistency in identifying the absolute best colormaps for 3D reconstruction arises because the heuristic approach examines only pairs of images. In contrast, the reconstruction process involves each image matching with all others. This limitation is intentional to reduce computational costs while still providing useful colormap recommendations. Interestingly, the jet, rainbow, and twilight_shifted colormaps also performed well for dataset B reconstruction, further validating their effectiveness in underwater environments featuring similar colors.
The results for dataset B were similar to those for dataset A, with jet and rainbow appearing in all cases. Colormaps jet and rainbow frequently occupied the top positions and proved the best colormaps for 3D reconstruction. The colormap twilight_shifted was included in 75% of results and was also a colormap that was very good in reconstruction. Although twilight was also present in dataset A, it did not rank among the top colormaps for reconstruction in dataset B. Interestingly, a different colormap, named turbo, appeared in dataset B but was not among the best in terms of reconstruction.
In dataset B’s case, the heuristic approach worked much better than dataset A, as it identified the top colormaps in all four experiments. It demonstrates that the heuristic approach can effectively provide valuable suggestions about which colormaps to use. The difference in results between datasets A and B highlights the necessity of applying this approach to each dataset individually to account for unique characteristics. Although the best colormaps are often similar across similar datasets, they differ. Therefore, having a methodology that recommends the best colormaps for each dataset can be extremely helpful. The heuristic methodology required approximately 30 min to identify the best colormap when combined with image downscaling, which did not affect the findings. The colormap selection is not a simple task, as it requires performing the 3D reconstruction with all the colormaps to choose the best one. By using this method, the optimal colormap can be found without losing much time. Reconstructing all 22 colormaps in a dataset with 200 images needs approximately 22 h, which shows how important the heuristic approach is, which needed only 30 min to find the best colormap and 1 h for reconstruction. It is also important to note that heuristic results were different from the first iteration’s, which applied 22 colormaps and sorted them according to their matches. The results are shown in Table 5.
Table 5. Heuristic colormap selection approach results for dataset B.
The computational costs of the image processing pipeline were evaluated based on key stages, including image enhancement, colormap creation, and reconstruction. All times are for high-resolution images (5760 × 3840). The image enhancement process took approximately 10 min and 18 s for 280 images, averaging 2.21 s per image. The time needed for a heuristic to find the best colormap was 12 min and 40 s. The colormap creation stage demonstrated improved efficiency, processing the same number of images in 2 min and 20 s, achieving a speed of 1.99 images per second. Regarding reconstruction, the colormap-based reconstruction required 1 h and 56 min, while unprocessed reconstruction took slightly longer at 2 h and 1 min, highlighting the computational efficiency associated with colormap processing. These results emphasize the trade-off between processing time and the benefits of enhanced visualization, demonstrating the efficiency of colormap creation while acknowledging the substantial time investment required for high-resolution image enhancement and reconstruction.
In terms of hardware usage, image enhancement utilized 24% of the CPU and 231.6 MB of memory. The heuristic finding process was more computationally demanding, utilizing 75.8% of the CPU while maintaining a memory usage of 229.6 MB. Colormap creation had a lower CPU load at 13.5% but required significantly more memory, reaching 974.7 MB. For 3D reconstruction, the process without colormap used 37.5% of the CPU and 6290.9 MB of memory, whereas reconstruction with colormap had a similar CPU usage of 37% but a higher memory consumption of 7128.9 MB. None of these processes utilized the GPU, indicating that all computations were performed on the CPU.

5. Conclusions

This paper introduced a novel method for enhancing 3D reconstruction using colormaps and image enhancement techniques. Additionally, a heuristic methodology was proposed to identify the most effective colormaps for each dataset. This approach provided more image features, leading to better stitching and a more detailed 3D model. The methodology was tested using two real-world datasets, demonstrating significant improvements. It did not add much substantial time to the existing process and proved highly effective. The heuristic methodology showed promising results by consistently identifying the top colormaps. It found three out of the four best colormaps in 100% of the experiments in dataset A. In dataset B, the top two colormaps were identified in all of the experiments, and the third-best colormap was included in the top four best colormaps in half of the experiments. This methodology can work quickly, in less than 30 min, providing users with valuable colormap suggestions.
The results demonstrated that this framework could enhance a state-of-the-art SfM system, leading to more complete and accurate scene reconstructions, particularly in challenging lighting conditions such as underwater environments. The proposed enhancement method resulted in improvements of 7.91% for Dataset A and 11.4% for Dataset B. Additionally, the use of colormaps further increased the number of features per image by 10.82% for dataset A and 64.11% for dataset B. These findings highlight the effectiveness of colormaps, especially in poor lighting and underwater scenarios. The real-world dataset confirmed that this methodology is practical and can significantly improve 3D reconstruction.
One limitation of the method is the potential increase in reconstruction time when utilizing the introduced features. The calculation of matches becomes more challenging due to the higher number of features present in each image, leading to increased complexity in match searching. This issue was addressed by employing sampling techniques to reduce the number of input images and accelerate the matching process, which can improve computational efficiency. In future work, other features aimed at enhancing the accuracy of 3D reconstruction or reducing computational costs to achieve optimal efficiency will be explored and evaluated. This can be achieved using sampling methods together with the proposed method for increasing the matches between images. This approach will allow the results to maintain good detail despite a reduced dataset, thanks to the use of colormaps, and produce results faster due to the sampling.
In future work, the methodology will be extended to cover a wider range of environments. Currently, the focus has been on underwater scenarios, as datasets for such environments are more readily available. However, to further evaluate the generalization capabilities of the method, efforts will be made to incorporate datasets from other challenging environments. These additional scenarios will provide valuable insights into the adaptability of the method to different conditions and help validate its performance in more diverse real-world applications.

Author Contributions

Conceptualization, A.V. and S.K.; methodology, A.V. and S.K.; software, A.V.; validation, D.T. and S.K.; formal analysis, A.V. and S.K.; investigation, D.T.; resources, D.T.; data curation, S.K.; writing—original draft preparation, A.V.; writing—review and editing, A.V. and S.K; visualization, A.V.; supervision, D.T. and S.K.; project administration D.T. and S.K.; funding acquisition, D.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Acknowledgments

This research has been financed by the European Regional Development Fund of the European Union and Greek national funds through the Action “Investment Plans of Innovation” in the framework of the Operational Program “Central Macedonia 2014–2020” (project code: i-ScubaDiving, KMP6-0082362). This support is gratefully acknowledged.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dimara, A.; Psarros, D.; Vrochidis, A.; Papaioannou, A.; Krinidis, S.; Anagnostopoulos, C.N. Mixed Reality’s and 3D Modelling Role in Cultural Heritage Preservation as a Pathway to Sustainable Heritage and Management. In IFIP International Conference on Artificial Intelligence Applications and Innovations; Springer Nature: Cham, Switzerland, 2024; pp. 405–418. [Google Scholar] [CrossRef]
  2. Alldieck, T.; Zanfir, M.; Sminchisescu, C. Photorealistic Monocular 3D Reconstruction of Humans Wearing Clothing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1506–1515. [Google Scholar] [CrossRef]
  3. Ferdani, D.; Fanini, B.; Piccioli, M.C.; Carboni, F.; Vigliarolo, P. 3D Reconstruction and Validation of Historical Background for Immersive VR Applications and Games: The Case Study of the Forum of Augustus in Rome. J. Cult. Herit. 2020, 43, 129–143. [Google Scholar] [CrossRef]
  4. Schonberger, J.L.; Frahm, J.M. Structure-from-Motion Revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  5. Snavely, N.; Seitz, S.M.; Szeliski, R. Modeling the World from Internet Photo Collections. Int. J. Comput. Vis. 2008, 80, 189–210. [Google Scholar] [CrossRef]
  6. Liu, Y.; Yan, Z.; Tan, J.; Li, Y. Multi-purpose oriented single nighttime image haze removal based on unified variational retinex model. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 1643–1657. [Google Scholar] [CrossRef]
  7. Aldeeb, N.H.; Hellwich, O. 3D Reconstruction Under Weak Illumination Using Visibility-Enhanced LDR Imagery. In Advances in Computer Vision: Proceedings of the 2019 Computer Vision Conference (CVC); Springer: Cham, Switzerland, 2020; Volume 1, pp. 515–534. [Google Scholar] [CrossRef]
  8. Guo, H.; Peng, S.; Lin, H.; Wang, Q.; Zhang, G.; Bao, H.; Zhou, X. Neural 3D Scene Reconstruction with the Manhattan-World Assumption. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5511–5520. [Google Scholar] [CrossRef]
  9. Williams, F.; Gojcic, Z.; Khamis, S.; Zorin, D.; Bruna, J.; Fidler, S.; Litany, O. Neural Fields as Learnable Kernels for 3D Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18500–18510. [Google Scholar] [CrossRef]
  10. Wang, D.; Cui, X.; Chen, X.; Zou, Z.; Shi, T.; Salcudean, S.; Ward, R. Multi-View 3D Reconstruction with Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 5722–5731. [Google Scholar] [CrossRef]
  11. Sun, J.; Xie, Y.; Chen, L.; Zhou, X.; Bao, H. NeuralRecon: Real-Time Coherent 3D Reconstruction from Monocular Video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15598–15607. [Google Scholar] [CrossRef]
  12. Liu, S.L.; Guo, H.X.; Pan, H.; Wang, P.S.; Tong, X.; Liu, Y. Deep Implicit Moving Least-Squares Functions for 3D Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1788–1797. [Google Scholar] [CrossRef]
  13. Kataria, R.; DeGol, J.; Hoiem, D. Improving Structure from Motion with Reliable Resectioning. In Proceedings of the International Conference on 3D Vision, Fukuoka, Japan, 25–28 November 2020; pp. 41–50. [Google Scholar] [CrossRef]
  14. Chen, S.; Zhang, K.; Shi, Y.; Wang, H.; Zhu, Y.; Song, G.; Zwicker, M. PAniC-3D: Stylized Single-View 3D Reconstruction from Portraits of Anime Characters. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 21068–21077. [Google Scholar] [CrossRef]
  15. Zins, P.; Xu, Y.; Boyer, E.; Wuhrer, S.; Tung, T. Data-Driven 3D Reconstruction of Dressed Humans from Sparse Views. In Proceedings of the International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; pp. 494–504. [Google Scholar] [CrossRef]
  16. DeGol, J.; Bretl, T.; Hoiem, D. Improved Structure from Motion Using Fiducial Marker Matching. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 273–288. [Google Scholar] [CrossRef]
  17. Matsumoto, T.; Hanari, T.; Kawabata, K.; Yashiro, H.; Nakamura, K. Integration of 3D Environment Models Generated from the Sections of the Image Sequence Based on the Consistency of the Estimated Camera Trajectories. IFAC-Pap. 2023, 56, 11281–11286. [Google Scholar] [CrossRef]
  18. Hanari, T.; Kawabata, K.; Nakamura, K. Image Selection Method from Image Sequence to Improve Computational Efficiency of 3D Reconstruction: Analysis of Inter-Image Displacement Based on Optical Flow for Evaluating 3D Reconstruction Performance. In Proceedings of the IEEE/SICE International Symposium on System Integration (SII), Narvik, Norway, 9–12 January 2022; pp. 1041–1045. [Google Scholar] [CrossRef]
  19. DeBortoli, R.; Nicolai, A.; Li, F.; Hollinger, G.A. Real-Time Underwater 3D Reconstruction Using Global Context and Active Labeling. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 6204–6211. [Google Scholar] [CrossRef]
  20. Guerneve, T.; Petillot, Y. Underwater 3D reconstruction using BlueView imaging sonar. In Proceedings of the IEEE OCEANS, Genova, Italy, 18–21 May 2015. [Google Scholar] [CrossRef]
  21. Maccarone, A.; Drummond, K.; McCarthy, A.; Steinlehner, U.K.; Tachella, J.; Garcia, D.A.; Pawlikowska, A.; Lamb, R.A.; Henderson, R.K.; McLaughlin, S.; et al. Submerged single-photon LiDAR imaging sensor used for real-time 3D scene reconstruction in scattering underwater environments. Opt. Express 2023, 31, 16690–16708. [Google Scholar] [CrossRef] [PubMed]
  22. Kulawiak, M.; Łubniewski, Z. 3D object shape reconstruction from underwater multibeam data and overground LiDAR scanning. Pol. Marit. Res. 2018, 2, 47–56. [Google Scholar] [CrossRef]
  23. Vrochidis, A.; Charalampous, P.; Dimitriou, N.; Kladovasilakis, N.; Chatzakis, M.; Georgiadis, G.; Tzovaras, D.; Krinidis, S. Automatic elevator shaft inspection using a multi-sensor measuring system and computer vision techniques. J. Build. Eng. 2024, 82, 108358. [Google Scholar] [CrossRef]
  24. Skarlatos, D.; Agrafiotis, P. Image-based underwater 3D reconstruction for cultural heritage: From image collection to 3D. Critical steps and considerations. In Visual Computing for Cultural Heritage; Springer: Cham, Switzerland, 2020; pp. 141–158. [Google Scholar] [CrossRef]
  25. Eltner, A.; Sofia, G. Structure from motion photogrammetric technique. Dev. Earth Surf. Process. 2020, 23, 1–24. [Google Scholar] [CrossRef]
  26. Wang, X.; Wang, C.; Liu, B.; Zhou, X.; Zhang, L.; Zheng, J.; Bai, X. Multi-view stereo in the deep learning era: A comprehensive review. Displays 2021, 70, 102102. [Google Scholar] [CrossRef]
  27. Calantropio, A.; Chiabrando, F. Underwater cultural heritage documentation using photogrammetry. J. Mar. Sci. Eng. 2024, 12, 413. [Google Scholar] [CrossRef]
  28. Kabadayı, A.; Erdoğan, A. Application of terrestrial photogrammetry method in cultural heritage studies: A case study of Seyfeddin Karasungur. Mersin Photogramm. J. 2022, 4, 62–67. [Google Scholar] [CrossRef]
  29. Xue, Y.; Shi, P.; Jia, F.; Huang, H. 3D reconstruction and automatic leakage defect quantification of metro tunnel based on SfM-Deep learning method. Undergr. Space 2022, 7, 311–323. [Google Scholar] [CrossRef]
  30. Berrett, B.E.; Vernon, C.A.; Beckstrand, H.; Pollei, M.; Markert, K.; Franke, K.W.; Hedengren, J.D. Large-scale reality modeling of a university campus using combined UAV and terrestrial photogrammetry for historical preservation and practical use. Drones 2021, 5, 136. [Google Scholar] [CrossRef]
  31. Thyng, K.M.; Greene, C.A.; Hetland, R.D.; Zimmerle, H.M.; DiMarco, S.F. True colors of oceanography: Guidelines for effective and accurate colormap selection. Oceanography 2016, 29, 9–13. [Google Scholar] [CrossRef]
  32. Zhang, H.; Zheng, R.; Zhang, W.; Shao, J.; Miao, J. An improved SIFT underwater image stitching method. Appl. Sci. 2023, 13, 12251. [Google Scholar] [CrossRef]
  33. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-up robust features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  34. Kalms, L.; Mohamed, K.; Göhringer, D. Accelerated embedded AKAZE feature detection algorithm on FPGA. In Proceedings of the 8th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies, Bochum, Germany, 7–9 June 2017; pp. 1–6. [Google Scholar] [CrossRef]
  35. Meza, J.; Marrugo, A.G.; Sierra, E.; Guerrero, M.; Meneses, J.; Romero, L.A. A structure-from-motion pipeline for topographic reconstructions using unmanned aerial vehicles and open-source software. Commun. Comput. Inf. Sci. 2018, 885, 213–225. [Google Scholar] [CrossRef]
  36. Aglave, P.; Kolkure, V.S. Implementation of high-performance feature extraction method using oriented fast and rotated brief algorithm. Int. J. Res. Eng. Technol. 2015, 4, 394–397. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.