Next Article in Journal
Overcoming Dimensionality Constraints: A Gershgorin Circle Theorem-Based Feature Extraction for Weighted Laplacian Matrices in Computer Vision Applications
Next Article in Special Issue
Residual-Based Multi-Stage Deep Learning Framework for Computer-Aided Alzheimer’s Disease Detection
Previous Article in Journal / Special Issue
The Accuracy of Three-Dimensional Soft Tissue Simulation in Orthognathic Surgery—A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Advances in Real-Time 3D Reconstruction for Medical Endoscopy

by
Alexander Richter
1,2,*,†,
Till Steinmann
2,†,
Jean-Claude Rosenthal
3 and
Stefan J. Rupitsch
2
1
Fraunhofer Institute for High-Speed Dynamics, Ernst–Mach–Institut (EMI), Ernst-Zermelo-Straße 4, 79104 Freiburg, Germany
2
Electrical Instrumentation and Embedded Systems, Albert–Ludwigs–Universität Freiburg, Goerges-Köhler-Allee 106, 79110 Freiburg, Germany
3
Fraunhofer Institute for Telecommunications, Heinrich–Hertz–Institut (HHI), Einsteinufer 37, 10587 Berlin, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(5), 120; https://doi.org/10.3390/jimaging10050120
Submission received: 13 March 2024 / Revised: 23 April 2024 / Accepted: 24 April 2024 / Published: 14 May 2024

Abstract

:
This contribution is intended to provide researchers with a comprehensive overview of the current state-of-the-art concerning real-time 3D reconstruction methods suitable for medical endoscopy. Over the past decade, there have been various technological advancements in computational power and an increased research effort in many computer vision fields such as autonomous driving, robotics, and unmanned aerial vehicles. Some of these advancements can also be adapted to the field of medical endoscopy while coping with challenges such as featureless surfaces, varying lighting conditions, and deformable structures. To provide a comprehensive overview, a logical division of monocular, binocular, trinocular, and multiocular methods is performed and also active and passive methods are distinguished. Within these categories, we consider both flexible and non-flexible endoscopes to cover the state-of-the-art as fully as possible. The relevant error metrics to compare the publications presented here are discussed, and the choice of when to choose a GPU rather than an FPGA for camera-based 3D reconstruction is debated. We elaborate on the good practice of using datasets and provide a direct comparison of the presented work. It is important to note that in addition to medical publications, publications evaluated on the KITTI and Middlebury datasets are also considered to include related methods that may be suited for medical 3D reconstruction.

1. Introduction

In recent decades, Minimally Invasive Surgery (MIS) has demonstrated superior patient recovery and surgical outcomes with smaller incisions, less pain, lower risk of infection, reduced blood loss, shorter hospital stays, and less scarring than traditional surgery. In the context of laparoscopy, which is depicted in Figure 1, substantial progress was made concerning image quality, in vivo imaging techniques such as narrow-band, intra-operative tools in robotics, 3D position tracking, 3D reconstruction of surgical scenes, and many more [1,2]. Despite this progress, a well-experienced surgeon is still needed to carry out minimally-invasive interventions due to a multitude of reasons: a small field of view, a restricted range of motion, a shadowless illumination of the surgical area, and a lack of depth perception, to name a few. But even for trained professionals, it is hard to accurately survey the correct dimensions of anatomical landmarks [3].
During MIS only limited visual cues for depth perception are visible. This increases mental rotation and transformation challenges to surgeons and can contribute to misjudgment, increased cognitive workload, and fatigue [4]. This is underlined by an evaluation of 252 cases of laparoscopic cholecystectomy procedures, where 97% of surgical accidents occurred as a result of visual misperceptions [5]. When compared to conventional MIS imaging, higher-resolution 3D imaging systems have been shown to significantly improve the depth perception of surgeons, which is critical for the precision and safety of surgical procedures [3,6].
By further taking into account the depth information of a scene and creating a 3D reconstruction in real-time, it becomes possible, e.g., to determine the distance between two points more precisely. There are various applications for real-time 3D reconstructions, missing region detection in colonoscopy, registration in sinus surgery, and laparoscopic surgery. Together with Augmented Reality (AR), 3D reconstruction allows for highlighting and viewing objects or organs from different points of view. During a bronchoscopy, colonoscopy, or sinus surgery, a 3D reconstruction could give the physician orientation by utilizing AR to show 3D objects as landmarks, which makes it easier to place stents or remove tissue from the correct location. 3D reconstructions could also provide depth information to cancer- or polyp-detecting algorithms to provide more information and help better detect cancer tissue.
The real-time capability of 3D reconstruction algorithms is essential for surgeries where the physician can be assisted during an operation. In this case, the algorithm has to provide depth information promptly. If the already present video feed during an operation can be used to analyze the scene in real-time and provide depth information without hindering the physician, the optional depth information can only be beneficial. Additional assistance, e.g., by offering navigational aid or removing surgical instruments from the surgeon’s view is also conceivable, although instrument-tissue interaction makes gathering depth information in these areas challenging [7]. Moreover, navigation and instrument segmentation or removal are not in the scope of this contribution.
In principle, it is possible to perform a 3D reconstruction with greater detail in a post-operation video analysis, then however, the generated depth information is not available during the operation. Especially when considering that modern videoendoscopes such as the TIPCAM®3D videoendoscope from KARL STORZ SE & Co. KG, Tuttlingen, Germany, are equipped with 4K resolution at 60 Hz. As of today, due to computational limitations, the real-time constraint does introduce challenges concerning the amount of data that can be processed and is largely dependent on the used hardware. Hence, we focus on publications that claim real-time functionality at the time of publication and those that are likely to reach real-time functionality, with state-of-the-art hardware.
This contribution compares real-time 3D reconstruction methods and the used datasets for evaluation, and further highlights the importance of common evaluation techniques. Common evaluation techniques allow for a comparison between contributions and thereby make advances in the field of real-time 3D reconstruction for medical endoscopy visible. Many of the reviewed contributions use their own evaluation metrics and set of images, which makes it impossible to rank them in the current state-of-the-art. In some datasets, the evaluation parameters for 3D reconstructions are well-defined; for self-sampled images, these parameters are not often published.
Fortunately, new computer vision and AI-based algorithms in the fields of autonomous driving, robotics, and unmanned aerial vehicles have attracted attention for their success in tracking and reconstructing scenes in real time. However, most developed methods rely on a true-to-scale representation of their environment and assume a non-deformable “static world”. In contrast, this assumption does not hold for surgical applications, where organic tissue is inevitably deformed, especially when using flexible endoscopes, where independent and simultaneous movement of the endoscope and tissue is possible. Nevertheless, these advancements could pave the way for groundbreaking achievements in the domain of surgical intra-operative assistance.
As researchers steadily compete to find the best solutions, various approaches for surgical intra-operative assistance have been developed. Some of the methods presented in our work require additional hardware to function. Others solely rely on the visual input of the image sensor and, therefore, do not require further modifications to the existing MIS hardware. To provide a comprehensive overview of the current state-of-the-art for real-time 3D reconstruction methods, a thematic division of the discussed contributions is realized in Figure 2. Here, first, the number of lenses is categorized. Then, passive and active methods are distinguished, and while active methods actively emit a measurand to perform a 3D reconstruction, passive methods rely only on an image sequence provided by the image sensor. Please note that the categorization is performed as provided by the original authors.
While other categorizations such as Endoscope Flexibility, Application Environment, or Operational Complexity are possible; here, we focus mainly on the technologies used for 3D reconstruction. Nevertheless, it is important to be aware of the characteristics of each technology when examining the results presented in this review. For example, flexible endoscopes can have only one lens and a considerably shorter working distance than rigid endoscopes.
Under closer inspection, it becomes apparent that it is challenging to directly compare the sheer endless number of methods in detail. The problems here vary from missing information, e.g., execution times or used hardware, to non-published self-constructed evaluation setups that prevent a comparison against established methods. Therefore, in the following sections, let us contrast the most relevant real-time contributions concerning the state-of-the-art. In this contribution, we concur with the statement of Liu et al. [8] and define real-time for MIS applications as 10 Frames per Second (FPS) or above.

2. Evaluation Tools

This chapter outlines the tools used to compare different 3D reconstruction methods. Methods are compared under the consideration that parameters such as the field of view, and distance to the object can influence the 3D reconstruction. In medical applications, the observed scenes can be challenging as they can contain structureless surfaces, challenging lighting conditions, and liquids, like blood.
First, we will discuss the good practice of using datasets and the associated benefits, such as the ability to evaluate and compare algorithms against each other objectively. For this, both medical and non-medical datasets are addressed.

2.1. Disparity Map vs. Point Cloud

The two primary forms of representing 3D data of stereoscopic scenes are the disparity map and the point cloud [9]. The disparity map is a pixel-based inverse representation of the scene’s depth range from a given perspective [10]. The disparity range in an image can be calculated, for example by Confidently Stable Matching [11], or statistical analysis of the spatial correlation between stereo images [12]. The point cloud is a set of 3D data points that represent the captured scene in 3D space. A conversion between both is possible using the intrinsic and extrinsic camera parameters obtained from a calibration process [13]. To evaluate the accuracy of an algorithm, disparity maps, and point clouds can both be used to compare a result against ground truth data.
Camera-based 3D reconstruction systems commonly exploit the disparity between two images to create a disparity map. However, not all vision systems rely on disparities for depth perception. The time-of-flight technique can obtain depth information directly by measuring the flight time of emitted directional light while measuring occurring phase shifts [14]. Knowing the speed of light and the direction of the emitted light, results in a 3D coordinate, thus directly creating a point cloud.

2.2. Metrics for Comparison

It is essential to differentiate between accuracy and precision, as they are sometimes incorrectly used interchangeably. Accuracy denotes how well the mean value of multiple measurements fits the expected value of the measured quantity. Precision describes how far the measurements are scattered from their mean value on average. For example, when shooting darts at a target, the bullseye is the expected value of the observed throws. The closer the average of all hits is to the bullseye, the higher the accuracy. The closer the hits are to each other, regardless of the location, the higher the precision. This matter is illustrated in Figure 3.
Concerning 3D reconstruction, high precision is often assumed if the algorithm repeatedly produces the same result when evaluating a single image. A better consideration in terms of accuracy, however, is the analysis concerning all pixels of an image. Then, the precision represents the ability to reproducibly deliver results with the same accuracy, and the accuracy represents how true to detail the 3D reconstruction is concerning the ground truth data. Therefore, when a new method is introduced, both the precision and the accuracy should be stated. In the following paragraphs, we aim to introduce the metrics used in the reviewed literature.
One of the easiest methods to evaluate the quality of a 3D reconstruction against ground truth data is to compare the disparity maps, which are described in Section 2.1. Assuming a disparity map produced from an arbitrary stereo reconstruction algorithm and a ground truth disparity map, ideally representing the true disparity map without any errors, then, the mean absolute error (MAE) between the 3D reconstruction and the ground truth data can be calculated according to
MAE = p = 1 n | y ^ p y p | n ,
where y ^ p is the pixel at position p in the resulting disparity map of an arbitrary algorithm, y p is the pixel at position p in the ground truth disparity map, and n is the number of pixels in each disparity map. The MAE is a measure of the average magnitude of errors in a test set, without considering their direction. It describes the equally weighted average over all absolute differences between the expected result and the actual result. This unit of measure is used, for example, when comparing different algorithms against the Stereo Correspondence and Reconstruction of Endoscopic Data (SCARED) dataset by Allan et al. [15].
One of the most significant error metrics is the root mean squared error (RMSE) [16], which is used in the Stereo-Endoscopic Reconstruction Validation dataset based on conebeam CT (SERV-CT) dataset [17] and is defined by
RMSE = p = 1 n ( y ^ p y p ) 2 n .
The RMSE describes the square root of the average of squared differences between the expected result and the actual result. Since the errors are squared, the RMSE gives a higher weight to large errors [18]. Therefore, the RMSE penalizes larger errors more harshly. This becomes clear when comparing five imaginary measurements with an evenly distributed error of two, which results in an MAE and RMSE of 2. However, if the same measurement is performed with four correct values and only one outlier of value ten, we will obtain an MAE of 2 and an RMSE of roughly 4.5. In both cases, the total error is ten, yet the RMSE is higher for the second case due to the larger deviation from the expected value.
To illustrate the differences between the error metrics, Figure 4 demonstrates an exemplary evaluation for each metric. It is important to note that the same evaluation data are used for each error metric.
Figure 4 clearly shows that the error metrics are not interchangeable and the curves show a distinct behavior. Depending on the application, one might choose the MAE metric over the RMSE or vice versa. The MAE is less sensitive to outliers compared to the RMSE [18]. Therefore, to better understand how true to detail a 3D reconstruction is, the MAE might be better suited. However, to determine the accuracy of a model under the consideration of outliers, the RMSE is the better choice.
Assuming that the error distribution is somewhat Gaussian, it is good practice to provide the standard deviation (SD) in addition to the MAE or RMSE. The standard deviation (SD) is defined by
SD = p = 1 n ( y ^ p μ ) 2 n 1 ,
where μ is the mean of y ^ p over all pixels p. The SD is a measure of the amount of variation of a set of values and denotes the precision.
Some datasets introduce error metrics to determine the accuracy of a disparity map or point cloud. The Middlebury dataset (MD) employs a metric, which defines the error, as the percentage of pixels whose absolute disparity error is larger than one pixel. An alternative to MD’s D1-all error is defined by the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) dataset. It calculates the percentage of pixels with an estimation error larger than 3 pixels and larger than 5% of the true disparity [19]. The Bad3 error is similar to the D1-all error metric of the KITTI dataset, as it determines the percentage of pixels with an estimation error larger than 3 pixels. It is used in the SERV-CT dataset to calculate the accuracy of the alignment to the true point cloud.
The runtime of an algorithm is another factor that is especially important for real-time applications. In the field of computer vision, typically either the execution time measured in seconds or the FPS are provided. Depending on the application, the number of times an algorithm must run per second can vary. Many factors influence the FPS of a 3D reconstruction algorithm. Larger depth ranges and larger image sizes, for example, can increase the runtime of an algorithm [20]. In addition, the underlying hardware also has a large impact on the performance, which is discussed further in Section 2.5.
The correct classification of these metrics plays a fundamental role in objective comparability. As many factors influence the performance of an algorithm in terms of quality and speed, the best approach to objectively evaluate and compare new methods is to use common public datasets with consistent metrics. This fact is elaborated in Section 2.3.

2.3. Datasets

Historically, the comparison of methods targeting 3D real-world problems or “3D in the wild” has always been a big challenge in the computer vision community. To overcome this, strong and continuous efforts have been made starting in the late 1990s, introducing the most famous Tsukuba stereo test image dataset with known ground truth values. Nowadays, datasets from various publishers can be easily found on the web. An overview of the datasets used by contributions mentioned in this publication is presented in Figure 5 and Table 1. The significant advantage of using datasets is the ability to objectively and independently compare developed methods against each other by guaranteeing that all contributions use the same data and error metrics. Ideally, the raw data and ground truth data are both provided. In the field of 3D reconstruction that could be a video of a dynamic scene or a still stereo image paired together with a ground truth point cloud.
Some of the less recent datasets use smaller image sizes and are therefore difficult to compare with modern datasets or the high demands of modern operating rooms. However, 3D reconstruction methods that have been published with results on these datasets have the potential to perform well on newer hardware and are therefore also included in this review. The easy access to suitable datasets in different fields can even make it possible to re-evaluate a developed method as shown in the SERV-CT [17]. Sometimes, methods developed for one application will show promising results when evaluated against another dataset for a different application.
For non-medical stereo applications, the Middlebury dataset (MD) and the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) datasets are the most popular. The Middlebury Dataset is a widely recognized benchmark for evaluating computer vision algorithms, providing datasets for stereo vision, optical flow, and 3D reconstruction. It consists of six different datasets, 2001 [21], 2003 [22], 2005 [23,24], 2006 [23,24], 2014 [16], and 2021 [16]. The KITTI dataset by Menze and Geiger [19] is mainly focused on optical flow applications using their autonomous driving platform AnnieWAY [25]. The latest version is from 2015 and updates the 2012 benchmarks for stereo and optical flow applications while adding a benchmark for scene flow applications [19].
For medical applications, authors can validate their developed algorithms either on their own images and videos or one of the following datasets: the Hamlyn Centre Laparoscopic/Endoscopic Video Dataset, the EndoSlam dataset, EndoAbs dataset, the Stereo Correspondence and Reconstruction of Endoscopic Data (SCARED) sub-challenge of the EndoVis challenge, or the SERV-CT dataset.
The Hamlyn Centre Laparoscopic/Endoscopic Video Dataset provides videos for algorithms working on a binocular and monocular setup [26]. The videos provided are contributed by Mountney et al. [27], Stoyanov et al. [1,28], Lerotic et al. [29], Pratt et al. [30], Giannarou et al. [31], and Ye et al. [32,33]. As of today, in the Hamlyn dataset, only the videos provided by Stoyanov et al. [1] and Pratt et al. [30] come with ground truth data. The ground truth data were generated with the stereo matching software Library for Efficient Large-scale Stereo Matching (Libelas) [34]. Contrary to the MD and the KITTI dataset, the Hamlyn Centre Laparoscopic/Endoscopic Video Dataset does not provide evaluation guidelines or a platform for uploads.
EndoSlam is constructed to design and compare 6-Degrees of Freedom (DoF) pose estimation and dense 3D map reconstruction algorithms. It was published in 2020 by Ozyoruk et al. [35] and the dataset contains recordings and ground truth 3D reconstructions from multiple endoscope cameras of porcine animal organs of multiple cadavers. Image sequences are recorded by a hand-manipulated Panda robotic arm with varying frame rates and changing lighting conditions. In total, the dataset consists of 42,700 frames of which 21,428 are recorded via a High-Resolution Endoscope YPC-HD720P at 1280 × 720 pixels, 17,978 are recorded via the Low-Resolution 3 in 1 Endoscope Camera at 640 × 480 pixels, 3055 are recorded via a MiroCam® MC1000-W endoscopic video capsule by IntroMedic (Seoul, Republic of Korea) at 256 × 256 pixels, and 239 are recorded via a Pillcam® COLON2 double endoscope camera capsule by Medtronic (Minneapolis, MN, USA) at 320 × 320 pixels. High-precision ground truth data were generated by two commercially available 3D scanners, Artec 3D Eva by Artec Eva (Senningerberg, Luxembourg) and Einscan Pro 2x by Shining 3D (Hangzhou, China). To simulate the problems of transfer learning and domain adaptation the dataset also contains synthetically generated data from a 3D simulation environment. The dataset is not provided with an evaluation toolkit, but the authors suggest determining the RMSE, as they use it for the evaluation of their self-developed algorithm.
In the work of Rau et al. [36] published in 2023, the 2022 EndoVis sub-challenge SimCol3D, the submitted methods, and their results are addressed. The authors argue that a 3D map of the colon could enhance the identification of unscreened colon tissue and serve as a training platform. In their work, the SimCol3D Endovis challenge, three synthetic colon sub-datasets with a resolution of 475 × 475 pixels are presented together with real video sequences without depth information. The submitted monocular 3D reconstruction methods, their results including RMSE, and published data are discussed in detail. Rau et al. [36] concluded that depth prediction in virtual colonoscopy is robustly solvable, while pose estimation remains an open research question.
The EndoAbs dataset was developed in 2018 by Penza et al. [37]. It focuses on creating a reference point cloud and stereo images of a 3D-printed environment containing abdominal organs. A total of 120 stereo images, with a 640 × 480 pixels resolution, were recorded by a stereo endoscope consisting of two Ultra Mini CMOS analogical color cameras by MISUMI (New Taipei City, Taiwan). Corresponding reference point clouds were created by a VIVID 910 laser scanner, one of which achieves at a distance of 0.6 m an accuracy of x = ± 0.22 , y = ± 0.18 , z = ± 0.07 and a precision of 8 µm. The camera-laser calibration error was determined to be 0.43 mm. For determining the error of a 3D reconstruction, the dataset does not suggest an error metric and does not provide a software toolkit. The phantom abdominal organs were printed from 3D models of the liver, spleen, and kidney and also tried to mimic the stiffness and texture of real organs. Varying lighting conditions and smoke are considered in this dataset to create an environment that closely mimics reality. Doctors rated the dataset a respectable average realism rating of 2.7 out of 5 in a follow-up survey.
The Stereo Correspondence and Reconstruction of Endoscopic Data (SCARED) EndoVis sub-challenge was presented at the International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) in 2019 by Allan et al. [15]. It consists of seven stereo-training datasets and two test datasets captured using the da Vinci Xi surgical robot. Each dataset contains four to five unique scene views, called keyframes, of porcine cadavers, as shown in Figure 6. High-precision ground truth data are provided for the first frame of each keyframe using the Structured Light (SL) approach [16]. The SL depth map is rotated and translated based on robot arm/endoscope movements to generate a dynamic sequence with known ground truth values for most of the frames in a keyframe. In addition, stereo calibration parameters were provided in different qualities to simulate improper endoscope calibration or noisy data. The related SCARED publication lists the ranking of all ten participants and gives some details of submitted methods [15].
The Stereo-Endoscopic Reconstruction Validation dataset based on cone-beam CT (SERV-CT) dataset, similarly to the SCARED dataset, aims to provide a stereo-matching dataset for the challenges encountered during MIS. It was developed by Edwards et al. [17] in 2020 to provide a freely available alternative to SCARED. The SERV-CT consists of two sub-datasets, containing eight calibrated image pairs of one of two porcine cadavers together with ground truth data. The images are recorded with a first-generation da Vinci surgical robot by Intuitive Surgical, at a resolution of 720 × 576 pixels. The CT ground truth data are provided by an O-arm™ Surgical Imaging System by Medtronic. An interventional scanner on the inside of the system generates 3D reconstructions equivalent to CT reconstructed from a rotating X-ray set enclosed within a circular casing. In the second sub-dataset, the authors provide, in addition to the CT data, an RGB surface reconstruction created by the Creaform Go SCAN 20 hand-held structured light scanner. This RGB data allows algorithms to align smooth and featureless surfaces better. To evaluate 3D reconstruction algorithms the SERV-CT provides a toolkit that calculates the RMSE and the Bad3 error. The dataset provides scenes with a range of different tissue types, including smooth surfaces with specular reflection properties, as well as depth variations and occluded surfaces.
A novel 2D-3D registration technique, to register optical mono video sequences with ground truth renderings of a known 3D model, was presented in 2023 by Bobrow et al. [38]. It works by transforming optical images into depth maps with a Generative Adversarial Network (GAN) and aligning edge features with an evolutionary optimizer. The new technique also leverages video information, achieving an average translation error of 0.321 mm and an average rotation error of 0.159° in simulation experiments where error-free ground truth is available. Furthermore, the authors printed a colon with 16 µm resolution on an Objet260 Connex 3 printer by Stratasys (Eden Prairie, MN, USA), molded the print, filmed the resulting mold of the 3D print, and used the proposed 2D-3D registration technique to generate ground truth data. The resulting dataset includes 22 short video sequences registered to generate 10,015 HD video frames of realistic colon phantom models obtained with a clinical colonoscope with ground truth depth, surface normals, optical flow, occlusion, six-degree-of-freedom pose, coverage maps, as well as 3D models. The dataset and registration source code are publicly available.
In 2019, Rau et al. [39] addressed the use of computer-assisted interventions in colonoscopy for early detection and treatment of colorectal cancer. The authors presented a method to generate depth information from monocular endoscopic images by training a conditional GAN dubbed pix2pix. Also, to overcome the lack of labeled training data in endoscopy, they proposed using simulation environments and training the generator and discriminator of the model on unlabeled real video frames to adapt to real colonoscopy environments. The authors reported promising results on synthetic, phantom, and real datasets and claimed that their generative model outperforms discriminative models when predicting depth from colonoscopy images in terms of accuracy and robustness towards changes in domains. With their extended pix2pix model, they reported a mean RMSE of 1.75 mm on synthetic data and a mean RMSE of 16.55 mm ± 0.81 mm on phantom data, which is partially obstructed by markers. In their work, the authors recorded ground truth data from a phantom model via a CT scan and generated synthetic colonoscopy images. Due to the optically overly smooth surfaces of the phantom model, the authors decided to generate another synthetic dataset consisting of about 5000 images using Unity in the same manner as the original training data. The published dataset consists of 16,016 synthetic colonoscopy RGB images with a resolution of 256 × 256 pixels, varying textures, and corresponding ground truth depth.

2.4. Comparability between Contributions

One problem that arises when comparing contributions from different publications and datasets is the usage of various error metrics to determine the accuracy and precision of a method. This problem is exacerbated by many authors who use self-constructed setups for said evaluation and do not publish the data, making replicable testing virtually impossible. In addition to this, there is only a limited number of medical datasets that concentrate on medical endoscopy or laparoscopy. As 3D reconstruction algorithms oftentimes perform well when evaluating against different environments than the ones the algorithm was developed for, we also consider datasets from different fields of research. Nevertheless, it is important to note that each field of research has its own challenges with respect to 3D reconstructions, e.g., deformations and structureless surfaces in the case of medical endoscopy. Additionally, algorithms applied in medical 3D reconstruction need to handle the presence of liquids, such as blood, smoke, and reflections from the endoscope’s light on the tissue. Thus, medical datasets also need to provide such scenes to train or test algorithms on.
The contributions with the best balance between speed and accuracy are compared against each other in Table 2, Table 3, Table 4 and Table 5. The methods have been split into multiple tables, as the contributions are using different datasets for their evaluation and also different lens setups have to be taken into account. For those who use existing datasets, the results can easily be compared against each other as shown in Table 3 and Table 5. All listed contributions in Table 3, have been evaluated on the test images Tsukuba, Venus, Teddy, and Cones of the Middlebury datasets 2001 [21] and 2003 [22]. The listed contributions in Table 5 use the KITTI 2015 dataset and follow a deep learning approach. The medical contributions listed in Table 2 and Table 4 use the RMSE, MAE, and SD to provide the accuracy and precision of their methods. Because some authors use different error metrics and datasets, an objective comparison between their presented methods is not always possible. The same will apply for runtime comparisons when different image sizes are used, different hardware is used, or if there is no mention of the used hardware, to begin with. Nonetheless, while the comparison between the publications is difficult to evaluate, useful information may be extracted. Thus, the information presented in Table 2 and Table 4 may be challenging to compare but is still provided for the sake of completeness.

2.5. GPU vs. FPGA

Most of the contributions presented in this overview, use either a Graphics Processing Unit (GPU) or an Field-Programmable Gate Array (FPGA) for computing a disparity map or point cloud. This section will go into detail about the advantages and disadvantages of both platforms.
The GPU architecture allows only limited hardware access [40]. Therefore, it is possible to run older code on new hardware with little to no change to the code itself, for example, when using NVIDIA’s highly parallelizable computing language CUDA [41]. A study in 2012 estimated that for an average post-doctoral employee, a GPU-based implementation of a dense optical flow, stereo matching, or local image features algorithm would take about two months, while developing the same algorithms on an FPGA will take about 12 months for the same person [42]. Real-time stereo-matching algorithms usually do not have a complex structure and, therefore, do not require many branching conditions or complex calculations. Since GPUs outperform FPGAs in tasks with little to no branching conditions and data dependencies [43], it shows why most real-time stereo matching algorithms use GPUs rather than FPGAs. This can be seen in the results of the work by Humenberger et al. [44], and is discussed in Section 5.1.
FPGAs, on the other hand, have a much lower power consumption [45] than GPUs, which are generally not suitable for power-sensitive systems [40]. An FPGA has a power consumption of 5–85 W depending on the application, while a GPU, like the NVIDIA RTX 3090 (Santa Clara, CA, USA), can draw 350–480 W. On a more complex system with many branching conditions, an FPGA is mostly quicker because of its high processing speed; up to billions of operations per second in parallel. The processing speed of an FPGA is faster than other hardware accelerators [40]. The fast processing speeds are achieved because the FPGA can map functions directly to hardware with no external memory access needed since all weights are stored close to the logic elements [46]. Therefore, the GPU will be the best option for real-time stereo-matching algorithms when the best performance is needed, and power consumption is not a priority. If a power-sensitive system is needed, or the algorithm requires many branching conditions, the FPGA will likely outperform the GPU. Due to the fast improvement of GPUs, these advantages will likely further decrease in the upcoming years. In the penultimate paragraph of Section 5.1, this statement is further supported by analyzing data provided by the contributions [44,47,48,49,50].

3. Monocular: Passive Methods

This section presents methods that rely exclusively on cameras to perform 3D reconstruction tasks. These methods use the visual input of a camera to acquire landmarks and different perspectives of the observed scene. A 3D reconstruction is performed over time, by identifying invariant robust feature points. Due to the monocular setup, matching algorithms can only be applied to images taken at different times, and therefore monocular systems struggle with strong or rapid camera movements, deforming objects as well as structureless scenes.
The following categorizations are carried out according to the designations provided by the individual authors. This is also the case for Structure from Motion (SfM) and Simultaneous Localization and Mapping (SLAM), which could be merged with regard to the real-time criterion. A summary of passive and active real-time monocular methods evaluated against self-developed datasets can be found in Table 2.
Table 2. Comparison of state-of-the-art real-time monocular algorithms evaluated against self-developed datasets. Either the root mean squared error (RMSE) or the mean absolute error (MAE) are determined by the authors to compute the error of the algorithm in millimeters. The methods are listed with the parameters most relevant (see Section 2.2) for real-time monocular methods. The parameters are measurement of accuracy, frame rate, used hardware, and image size. The double line singles out the last publication, as this method, only performs a single-point distance sensing.
Table 2. Comparison of state-of-the-art real-time monocular algorithms evaluated against self-developed datasets. Either the root mean squared error (RMSE) or the mean absolute error (MAE) are determined by the authors to compute the error of the algorithm in millimeters. The methods are listed with the parameters most relevant (see Section 2.2) for real-time monocular methods. The parameters are measurement of accuracy, frame rate, used hardware, and image size. The double line singles out the last publication, as this method, only performs a single-point distance sensing.
Monocular Contribution
(Active/Passive, Self Evaluated)
Error
in mm
FPS
in Hz
Hardware
AMD, Intel, NVIDIA
(Santa Clara, CA, USA)
Image Size
in px
Sfm from tracking [51]RMSE = 1.9332.5 GHz CPU
NVIDIA Quadro FX 570
n.a.
SLAM Dense Surface
Reconstruction [52]
RMSE = 2.541.6Intel Xenon 2.8 GHz
NVIDIA GTX 970
840 × 640
ORBSLAM [53]RMSE = 3–4.11.7Intel i5 3337U 1.8 GHzn.a.
Endo-Depth-
and-Motion [54]
RMSE = 11.023.1AMD Ryzen 9 3900X
NVIDIA RTX 2080 Ti
n.a.
VCSEL single point
laser distance sensing [55]
MAE = 0.045000n.an.a.
3D Scanner
Structured Light [56]
MAE = 0.130n.a. 400 × 400
Infrared coded
Structured Light [57]
MAE = 0.12n.a.n.a. 640 × 480
Multispectral
Structured Light [58]
MAE = 0.64–0.880.02Intel i7 3770 3.9 GHz 1024 × 768

3.1. Structure from Motion (SfM)

The aim of SfM is to reconstruct an accurate 3D representation, of an object or scene, from a series of 2D images. This is performed by analyzing the projected 2D motion field (motion parallax) of a moving object or scene and tracking corresponding pixels from one image to the next. To aid the reconstruction process, the camera’s motion and 3D position can also be approximated. In monocular setups, due to an unknown baseline between images, the absolute scale of the resulting 3D reconstruction can only be roughly estimated.
SfM is a method that can be used in monocular and binocular setups. With that said, to the best of our knowledge, SfM is currently not represented in binocular stereoscopic endoscopy and therefore the focus of this chapter lies on monocular setups.
SfM is typically performed offline; however, in the work of Sun et al. [51], an online method is presented that relies on both the intrinsic camera parameters and the tracking transformations associated with each acquired image. A magnetic motion tracking device is attached to the endoscope to compensate for the inability to annotate reference points during endoscopic procedures. The device enables an estimation of the position and pose of the endoscope in real time. The reconstruction error is determined by a phantom experiment using a gastric model with 28 markers, which resulted in an RMSE of 1.9 mm. Moreover, by utilizing a GPU, the performance of the proposed method reaches about 30 FPS on a system with a 2.5 GHz CPU, 4 GB RAM, and an NVIDIA Quadro FX 570 (Santa Clara, CA, USA) graphics card.
Recasens et al. [54] approach the SfM problem by using the Monodepth2 network architecture and training procedures [59] in combination with a keyframe-based photometric approach for creating a depth map and tracking the camera pose. With their algorithm, Recasens et al. [54] achieve an average RMSE of 11.02 mm for monocular applications. Using an NVIDIA RTX 2080 Ti (Santa Clara, CA, USA), the algorithm runs with approximately 3.1 FPS.
In the work of Yang et al. [60] geometric structural consistency is taken into consideration and a gradient loss to penalize edge fluctuations is introduced. The proposed method is evaluated across different datasets including the EndoSLAM dataset. Here, the method achieves an RMSE of 66 mm at 120 FPS with a resolution of 320 × 320 pixels for a sequence of a stomach.
In comparison, an offline SfM approach is utilized in the work of Vélez et al. [61] to reconstruct the surface of a liver. First, the relative pose of the camera in an endoscopic video sequence is reconstructed, using a keyframe-based 5-Point and Perspective-n-Point (PnP) pose estimation. This approach obtains the pose of the camera by minimizing the reprojection error of known 3D points onto the corresponding points. Thereby, it assumes that the camera is calibrated and therefore the intrinsic parameters are known. Then, an initial 3D surface is obtained via optimal triangulation, as proposed by Hartley and Zisserman [62]. With the aid of a Bundle Adjustment (BA) step, where the parameters of the 3D coordinates, relative motion, and the optical characteristics of the camera are simultaneously refined, the initial reconstruction is further optimized. When compared to the ground truth data, which is acquired with a robotic system and reference patterns, a RMSE of 0.19 mm is achieved.
Similarly, Malti et al. [63] combine both geometric and photometric cues to reconstruct deformable human organs robustly. A visualization of a reconstructed scene is presented in Figure 7. The evaluation of their method features an RMSE less than 0.4 mm on a synthetic deformation model.
The accuracy of the latter approaches shows that when using SfM at an appropriate resolution and working distance, a sub-millimeter resolution can be achieved. Unfortunately, the computational runtime is not documented in [63], so it is not clear whether this method can be used in a daily clinical routine with real-time constraints. To our best knowledge, more recent publications such as the work of Wang et al. [65] in 2017, Turan et al. [66] in 2018, and Widya et al. [67] in 2019, show no further improvements concerning runtime and accuracy. In addition, the authors use a self-constructed dataset which makes an objective comparison challenging.

3.2. Simultaneous Localization and Mapping (SLAM)

SLAM has its origin in the field of robotics, where the static world assumption applies, meaning the world to be mapped consists mainly of rigid non-movable, and non-deformable objects. In principle, the method is independent of the sensor technology, however, oftentimes cameras or Light Detection and Ranging (LiDAR) sensors are used. As the name suggests, SLAM refers to the process of estimating the position and orientation of a sensor, while simultaneously creating a map.
Visual SLAM can only determine the translation and rotation of a moving camera up to scale, resulting in point clouds that are missing a conversion factor to determine their absolute scale. Without data fusion with other sensors, missing absolute scale information can further result in a failure to detect scale drift, which in turn can have an impact on loop detections. Nevertheless, efforts have been made to estimate the baseline between consecutive frames to solve this problem, for instance, the work by Rukhovich et al. [68]. They try to estimate the baseline by training a network on different scenes and different baselines.
Here, additional sensors such as Sound navigation and ranging (Sonar) or LiDAR are often used to perform both the localization and the mapping of a robot [69].
With the aid of SfM techniques, this method can also be used in a passive monocular endoscopic setting to create a map of the environment, hence providing a 3D reconstruction. In the work of Mahmoud et al. [53], a real-time 3D reconstruction method, relying solely on an endoscopic image sequence, is presented. The evaluation is performed against Computed Tomography (CT) scans while fixing the endoscope into place relative to the operating table. This is performed while taking into consideration instrument occlusions and tissue deformations. The method uses oriented FAST and rotated BRIEF SLAM (ORBSLAM) [70] to estimate the endoscope position and create a 3D reconstruction of the environment. The oriented FAST and rotated BRIEF (ORB) feature extractor is an alternative to Scale-invariant feature transform (SIFT) and Speeded up robust features (SURF) [71].
Only 24% of the visible point cloud can be mapped, mainly because the ORB feature detector is not able to detect repeatable feature points on soft organ tissue and the BA considers 11–25% of the map’s points as non-rigid. Of these points, 80% are considered in the RMSE calculation, as the remaining 20% are either outliers or points reconstructed outside the field of acquisition of the CT scanner. A brute force approach is used to match acquired point clouds against the ground truth data. After the matching process, a RMSE of 3 mm is achieved, when taking into account only the position of ORB points. When creating a semi-dense map, as presented in the publication, the RMSE drops to 4.1 mm. Although there is no mention of the resolution of the image, an average frame rate of 2.5 FPS, is reported on a not further specified dataset. To achieve this, the authors use an Intel® Core D 3337U CPU (Santa Clara, CA, USA) running at 1.8 GHz and 6 GB of RAM. The authors attempted a dense SLAM based reconstruction in a follow-up publication [72], but with a refresh rate of under 0.17 FPS, the algorithm does not perform in real-time. The hardware is a system with an Intel® i7 running at 3.4 GHz, 8 GB of memory and NVIDIA GTX 680 (Santa Clara, CA, USA).
A different approach can be found in [52], where a geometry-aware AR framework for depth correct augmentation for intra-operative scenes is presented. The focus of this work is to place virtual objects correctly within the 3D scene in real-time, as can be seen in Figure 8. For this, the camera pose as well as the surfaces of the 3D scene are determined. The developed method is evaluated on the Hamlyn Centre Laparoscopic/Endoscopic Video Dataset [26] and a simulated MIS scene. The evaluation on the Hamlyn Centre Laparoscopic/Endoscopic Video Dataset demonstrates a 3D mesh that shows a good match to the video. This suggests that the method can provide the correct depth information intra-operatively and therefore be advantageous during surgical procedures. However, the authors mention that their method struggles with deformations and instruments, which obstruct large portions of the view. The simulated MIS scene with a realistic human digestive system model is generated with Blender [73]. The RMSE against the ground truth data from the simulation is 1.24 mm for the camera position and 2.54 mm for the surface reconstruction. The 3D surface reconstruction process is implemented without GPU acceleration and reaches 1.7 FPS. The used hardware consists of an Intel® Xeon 2.8 GHz quad-core CPU, 32 GB memory and an NVIDIA GTX 970 graphics card (Santa Clara, CA, USA). A performance increase would likely be possible by upgrading the existing hardware and modifying the algorithm to exploit GPU acceleration.

4. Monocular: Active Methods

In this section, we discuss methods, which resort to utilizing active sensors in addition to the camera’s sensors of the endoscope. The sensors are called active as they actively emit a signal, which can then be detected. The needed modifications of the endoscope can lead to an increase in size and/or power consumption. Such modifications are, for example, a Structured Light (SL) projector or a Time of Flight (ToF) sensor. An SL approach works by projecting a known light pattern onto a surface and the measured distortions in the captured projection pattern are used to determine the depth for each pixel. ToF methods typically utilize either pulsed or modulated infrared light and a light detector module that determines the time delay between emitting and detecting and thereby, by taking into account the speed of light, determines the distance to the surface.
For both procedures, the pose of the camera does not need to be known. In the case of the ToF system, a camera is only needed for color information. In addition to that, the usability and stability of the endoscope can be affected as well. A monocular 3D reconstruction is performed using the additional information provided by the extra sensors. A summary of passive and active real-time monocular methods evaluated against self-developed datasets can be found in Table 2.

4.1. Structured Light (SL)

The SL approach actively projects a spatially varying or color-coded intensity pattern onto a scene. An example of a color-coded pattern on a surface is depicted in Figure 9. The pattern is generated by a projector or light source modulated by a spatial light modulator. A monocular camera is then used to analyze the projected pattern. If the camera detects a planar surface, the pattern observed by the imaging sensor will be similar to that of the projected SL pattern. In the case of a surface with deformations, the camera records a distorted projection pattern. Based on the distortion, the shape of the surface can be reconstructed [74].
Albitar et al. [75] use a monochromatic light and the spatial neighborhood coding strategy, based on the theory of an M-array with three symbols. The authors modeled a pattern consisting of geometrical primitives. To simplify the search for the relevant neighborhood in the projected pattern, one of the primitives is used to store the local orientation of the pattern. The approach decoded 95% of the detected primitives correctly, even in the presence of spatial occlusions. On an image with a size of 568 × 760 pixels, the method achieved 12.5 FPS on a P4 2.67 GHz machine.
Another approach is presented by Jia et al. [57]. To acquire the linear dependency between the object depth and pixel shift, they propose a linear fitting algorithm. Using this dependency, the depth information of an object can be derived. The 3D reconstruction of the observed scene is derived by exploiting the Delaunay triangulation. The Delaunay triangulation takes a set of discrete points, where not all points are colinear, and no four points are cocircular, and applies the triangulation in such a way that the circumcenter of the triangle is defined by three points [76]. The authors claim to achieve real-time performance but do not offer data to back up their statement. At a resolution of 640 × 480 pixels, the method achieves an MAE of 0.12 mm.
According to Schmalz et al. [56], a catadioptric camera can be used to detect a circular multispectral pattern in tubular cavities, as seen in Figure 10. The single-shot SL prototype has a diameter of 3.6 mm and a length of 14 mm. The prototype acquires 3D data at 30 FPS and generates approximately 5000 3D points per second. After using the Iterative Closest Point (ICP) algorithm to merge the acquired point clouds, they are smoothed utilizing the method of Vollmer et al. [77] and a surface reconstruction according to Kazhdan et al. [78] is performed. The approach is evaluated at a working distance of about 1.5 cm after calibration with Zhang’s method [13]. When comparing the 3D reconstruction to a CAD model of a known cavity, the MAE between the reconstructed points and the model is 108 µm.
Another multispectral SL approach, using two camera types for capturing the projected pattern, is presented by Lin et al. [58]. The first camera, a DCU 223C by Thorlabs Ltd. (Ely, UK), has a regular color CCD sensor. The other is a multispectral camera, which contains eight different bandpass filters and eight channels in the output image. In terms of calibration, the geometrical camera calibration is applied, as put forward by Zhang [13]. For the evaluation, ground truth is acquired with a NIKON MCAx24+ (Minato City, Tokyo, Japan) handheld 3D scanner, which has a volume length accuracy of ±0.038 mm. Finally, the acquired 3D point cloud is compared to the ground truth using the ICP algorithm [79]. The resulting RMSE, which is derived based on the results presented in [58], equals 1.1 mm.

4.2. Time of Flight (ToF)

The ToF sensors exploit the phase shift between the emitted modulated light pulse and the one received to calculate the distance to an object, as visualized in Figure 11. For short distances, the hardware needs to be able to resolve the gap to the object in picoseconds because of the speed at which light travels. Due to the low cost, low energy consumption, and the possibility to reach 30 FPS [80], ToF sensors are present in multiple applications such as, computer graphics [81], computer-human interaction [82], robotics [83] and autonomous driving [84].
Based on this research, Groch et al. [14,85] develop an algorithm for 3D reconstruction as well. They use a ToF sensor in combination with the SfM approach to improve the accuracy of their algorithm. However, the authors of both papers do not mention the runtime of their algorithms. While using a ToF approach, a MAE of 4 mm ± 1 mm is achieved by both. In their contribution, Groch et al. [14] compare their result to the result of the stereo matching based Hybrid Recursive Matching (HRM) algorithm by Roehl et al. [86], which achieves an MAE of 2.3 mm ± 0.8 mm. The authors conclude that at the time of the publication, ToF methods are inferior to stereo-matching methods with respect to surface reconstruction.

4.3. Laser Distance Sensing

Laser distance sensing was developed based on the Michelson interferometer [87]. The interferometer exploits the difference in optical path length between a known path and an unknown one, as shown in Figure 12. The difference in optical path length is known as the retardation or optical path difference (OPD) and is induced by a beam splitter. An interferogram is obtained by varying the retardation and recording the signal from the detector. To analyze the frequencies and determine the distance to the surface, a Fourier transform is used [88].
Binocular and Monocular methods tend to fail when using triangulation on homogeneous regions due to a depletion of separable landmarks [89]. Lucesoli et al. [90] propose endoscopic single-point laser distance sensing, to measure the distance between the endoscope and the surface. When using a mechanical stepping motor to implement surface scanning, the acquisition time of the method is greater than one second. The authors achieve a resolution of 20 µm at a measurement distance of 20 mm.
To achieve a long measurement range and high scanning speeds, Moon and Choi [91] use a low-cost Vertical Cavity Surface Emitting Laser (VCSEL). This circumvents the mechanical limitations Lucesoli et al. [90] have. The authors cool the laser down by 21 K, to 3 °C, to increase the electric current, which in turn extends the sweep bandwidth by 1.8 nm. This results in a spatial resolution of 135 µm and a range of 100 mm. Building on the idea, Hariyama et al. [92] use four photodiodes and an Semiconductor Optical Amplifier (SOA) to improve both the measuring range and the accuracy. In combination with a galvanometer, the authors are able to create a 3D reconstruction of a scene. A 10 µm accuracy is achieved with the system, which has a measuring range of two meters.
Vilches et al. [55] in contrast, use a Gradient Index (GRIN) lens and a common-path interferometer to reduce the diameter of the system to 500 µm. The authors achieve a precision of 2.7 µm and an accuracy of 40 µm at a measuring distance of up to 50 mm. The surface distance measurement can be performed at a refresh rate of 5000 FPS in an ex vivo setup.

5. Binocular: Passive Methods

Passive binocular applications use two camera sensors to create a 3D reconstruction of a scene. Stereo-matching is the most common way to create a 3D reconstruction of a scene with cameras. It works by determining the correspondences between the stereo pairs and using the distance between the matched points to create a disparity map, which is an inverse depth map. The accuracy of the resulting disparity map can be influenced by multiple factors. This includes the camera resolution, the stereo baseline, featureless or reflective surfaces, and harsh lighting conditions [93]. Algorithms using the concept of stereo matching typically follow a global or local optimization approach. Global approaches use the entire image to optimize the result based on a well-behaved energy function. Therefore, they are less sensitive to local outliers when compared with local methods but have higher computational costs [94]. Local functions consider only the local neighborhood of pixels to calculate the matching costs and cost aggregation. The winner-takes-all approach, where the disparity with the lowest cost for each pixel is picked, is typically applied to optimize the matching costs.
Again, the following categorizations are carried out according to the descriptions provided by the individual authors. This is also the case for deterministic Stereo Matching and Deep Learning, while Deep Learning methods can utilize Stereo Matching. A summary of the passive real-time binocular methods evaluated against the MD can be found in Section 5.1, along with a summary of self-evaluated methods. In Section 5.2, a summary of Deep Learning methods evaluated against the KITTI dataset can be found.

5.1. Deterministic Stereo Matching

Deterministic stereo matching creates a 3D reconstruction from two images of the same scene. To achieve this, two corresponding points along the epipolar line in the images must be found, as visualized in Figure 13. The epipolar line is where the image and epipolar plane intersect, all corresponding matching points can be found on this line. If the images are rectified, the epipolar line runs horizontally to the image plane, reducing the complexity of the search for corresponding pixels. A summary of the passive real-time binocular methods evaluated against the MD can be found in Table 3. Similarly, the self-evaluated passive real-time binocular methods are summarized in Table 4.
Table 3. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against the Middlebury dataset (MD). The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are Middlebury measurement of accuracy (see Section 2.3, in ascending order), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Table 3. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against the Middlebury dataset (MD). The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are Middlebury measurement of accuracy (see Section 2.3, in ascending order), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Binocular Contribution
(Middlebury Dataset)
Error
in %
FPS
in Hz
Hardware
AMD, Intel, NVIDIA
(Santa Clara, CA, USA)
Inrevium (Shibuya, Tokyo, Japan)
Image Size
in px
Disparity
Range in px
Cross-based
Support Regions [95]
3.9449.7NVIDIA GTX 1070 384 × 280 16
Guided Image
Filtering [96]
5.5517NVIDIA GTX 480 640 × 480 40
Weakly-Textured
Scenes [97]
5.781NVIDIA GTX 8800 512 × 384 48
High-quality
Stereo Vision [49]
6.1731.79Intel (Altera) EP4SGX230
FPGA
1024 × 768 96
Two Pass Adaptive
Support Weights [98]
6.2062NVIDIA GTX 580 320 × 240 32
Hardware Guided
Image Filtering [47]
6.3660Inrevium Kintex-7
FPGA
1280 × 720 64
Line-wise HRM [99]6.6813NVIDIA Tesla C 2070 960 × 540 n.a.
Real-time BFV [100]7.6557NVIDIA GTX 8800 384 × 288 16
Belief Propagation [101]7.6916NVIDIA GTX 7900 320 × 240 16
High-def SM
on FPGA [48]
8.2060Intel (Altera) EP3SL150
FPGA
1024 × 768 64
Embedded Real-time
Systems [44]
9.73573.7NVIDIA GTX 280 320 × 240 15
Table 4. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against self-developed datasets except for the contribution by the Fraunhofer HHI, which evaluated against the SCARED dataset. In the contributions using self-developed datasets, either the root mean squared error (RMSE) or the mean absolute error (MAE) are determined by the authors to compute the error of the algorithm in millimeters. The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are: Measurement of accuracy (in ascending order, RMSE then MAE), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Table 4. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against self-developed datasets except for the contribution by the Fraunhofer HHI, which evaluated against the SCARED dataset. In the contributions using self-developed datasets, either the root mean squared error (RMSE) or the mean absolute error (MAE) are determined by the authors to compute the error of the algorithm in millimeters. The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are: Measurement of accuracy (in ascending order, RMSE then MAE), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Binocular Contribution
(Self Evaluated)
Error
in mm
FPS
in Hz
Hardware
AMD, Intel, NVIDIA
(Santa Clara, CA, USA)
Image Size
in px
Disparity
Range in px
Semi-dense Surface
reconstruction [102]
RMSE = 3.22.64NVIDIA Quadro K5000 1920 × 540 n.a.
Semi-dense Surface
reconstruction [1]
MAE = 1.0615NVIDIA Quadro FX 5800 360 × 288 n.a.
GPU/CPU Surface
reconstruction [103]
MAE = 1.5530Intel i7 930
NVIDIA Tesla C 2070
320 × 240 n.a.
Novel enhancement
to HRM [104]
MAE = 2.0614.5CPUn.a.n.a.
CPU Surface
reconstruction [86]
MAE = 2.620CPU 320 × 240 n.a.
Fraunhofer HHI
stereo pipeline [15]
MAE = 3.44
(SCARED Dataset)
45NVIDIA RTX 3090 1920 × 1080 n.a.
Given the relative positions of the cameras to one another and the parameters of the camera projection, a triangulation can be performed to determine the 3D position for every corresponding point. An example of a 3D reconstruction using deterministic stereo matching can be seen in Figure 14. Liu et al. [105] show that the minimum working distance should be at least 30 times larger than the baseline between the stereo cameras to generate better stereo images. For stereoscopes with a typical baseline of 4 mm this results in a minimum working distance of about 120 mm which is achievable in a medical environment.
Based on our research, deterministic stereo-matching algorithms follow either a correlation-based or a feature-based matching approach to finding corresponding points. Correlation-based methods use image rectification to optimize the matching process and to compensate for calibration and alignment errors. In doing so, stereo information lies on the same scanline, so that the estimating stereo correspondence search becomes a 1D problem [106]. Feature-based methods use a feature descriptor to find prominent features and a nearest neighbor search to find matching points [107]. Due to faster computation times and denser point clouds, correlation-based methods are more common.
Stereo matching typically concentrates on accuracy, run time, or power consumption. Since we focus on real-time applications in this overview, publications considered here run either in real-time or near real-time.
Newer hardware can allow authors to improve the accuracy of their algorithm, without significantly increasing the original runtime. This aspect is evident when looking at the Cross Based Support Regions (CBSR) methods by Zhang et al. [108]. The contributions [95,100] utilize cross-based support regions for creating a disparity map. A support region’s size and shape are determined adaptively regarding local color values and spatial distances. For a stereo image pair, with a resolution of 384 × 288 pixels and a maximum disparity of 16 pixels, [100] achieves 57 FPS on an NVIDIA GTX 8800 (Santa Clara, CA, USA). The contribution of Lee and Hong [95] reaches for the same stereo pair 49.7 FPS on a much newer NVIDIA GTX 1070 (Santa Clara, CA, USA). Regarding the accuracy of the approaches, Lee and Hong [95] provide an error in the MD benchmark of 3.94%, while [100] exhibits an error of 7.65%. Thus, the authors use newer hardware to keep a similar frame rate while nearly doubling the algorithm’s accuracy.
Furthermore, it becomes clear that some authors focus on runtime and others on accuracy. This can be seen in the publications [101,109], where the authors improve the idea of Felzenszwalb and Huttenlocher [110] to achieve accurate stereo matching algorithms. The work presented by them in 2004 paves the way for a linear approach to Belief Propagation (BP) in stereo matching. The contribution made it possible to use BP for real-time stereo-matching algorithms. Yang et al. [101] achieve an error of 7.69% on the MD benchmark. Using an NVIDIA GTX 8800 (Santa Clara, CA, USA) and an image with a resolution of 320 × 240 pixels and a maximum disparity of 16 pixels, they reach 16 FPS. Yang et al. [109] achieve only 0.1 FPS on an image with a resolution of 384 × 288 pixels and a maximum disparity of 16 pixels. The runtime is not real-time, but the accuracy is significantly improved to an error of 4.19% in the MD benchmark. The authors do not mention what hardware they use in this experiment [109].
Yang et al. [97] also exploit the idea of Felzenszwalb and Huttenlocher [110] with an emphasis on the runtime of the algorithm in their work. The idea of the authors is to use plane fitting in combination with BP. When depth values are believed to be incorrect plane fitting can be used to fit these values to a plane. This is used for example to improve the accuracy in weakly textured scenes [111]. The authors utilize BP to correct potential errors, which are caused by the non-robust process of color segmentation [97]. Using an NVIDIA GTX 8800 and an image of size 512 × 384 pixels with a maximum disparity of 48 pixels, the authors achieve a frame rate of 1 FPS. In the MD benchmark, the algorithm scores 5.78%, which puts it into the category of the better scoring algorithms in our overview.
While most methods achieve their best performance on a GPU, some algorithms like Hybrid Recursive Matching (HRM) [112] cannot profit as much from the parallelism on the GPU because of their recursive approach. The publications [86,99,103,104], and the algorithm by Allan et al. [15] use the HRM algorithm as a basis. This method exploits the current stereo image pair and the disparity map from the previous image pair to create a disparity map. An initial disparity guess is the width of the image. After that, the disparity estimations are made based on the disparity map from the previous frame. Therefore, the disparity range is not fixed.
To improve the runtime of the HRM algorithm on the CPU [86], it was partially migrated to the GPU in [103]. Due to the recursive nature of the algorithm, parallelization is only partially possible. The runtime is improved by 10 FPS on the same image size when compared to the publication [86]. Roehl et al. [103] also evaluate the accuracy between a GPU and CPU implementation. The CPU implementation achieves an MAE to the ground truth of 1 mm. Due to the changes made to migrate the algorithm to the GPU, the implementation results in a higher MAE of 1.55 mm.
Rosenthal et al. at the Fraunhofer Institute for Telecommunications, Heinrich–Hertz–Institut, HHI, won the 1st place in the category “lowest mean error” and overall 2nd place at the MICCAI 2019 EndoVis sub-challenge “SCARED” [15] organized by Intuitive Surgical. The method is based only partially on the HRM algorithm and is fully migrated to the GPU based on the work by Waizenegger et al. [113]. The stereoscopic 3D reconstruction pipeline consists of two pre- and post-processing steps to improve the overall performance: (1) Apply contrast enhancement using Contrast-limited adaptive histogram equalization (CLAHE) [114] to accentuate tissue structures, which are badly illuminated due to the given co-axial illumination in endoscopic imaging. (2) Usage of a trained depth histogram heuristic, which is based on typical endoscopic working distances, allowing discarding mismatches violating such depth ranges. Besides these assumptions, the disparity range does not need to be set to a constant value. Due to its recursive nature, it adapts automatically to the scene’s depth structure based on the previous estimation step at time t 1 plus a small overhead to compensate for endoscopic camera movement. This means that the start disparity range will be set to the image width for the first initial guess, 1920 pixels in the case of full HD, but it converges quickly to a meaningful disparity range within 15–20 frames. The algorithm achieves an MAE of 3.44 mm on the SCARED dataset. Depending on the scene structure, the method achieves frame rates of up to 45 FPS at a resolution of 1920 × 1080 pixels on an NVIDIA RTX 3090 (Santa Clara, CA, USA).
Kowalczuk et al. [98] and Hosni et al. [96] build on the concept of Adaptive Support Weights (ASW) by Yoon and Kweon [115] in 2005. The ASW method creates a window around each pixel and sums up all pixels inside this window. Before summation, the pixels in the window are weighted based on their color similarity and spatial gradient, compared with the center pixel. To improve the method by Yoon and Kweon [115], Kowalczuk et al. [98] propose using a two-pass approach. This approach results in an error of 6.20%, when compared to the MD. At a resolution of 320 × 240 pixels and a disparity range of 32 pixels, the authors achieve a frame rate of 62 FPS on an NVIDIA GTX 580 (Santa Clara, CA, USA). The implementation of the ASW algorithm by Hosni et al. [96] accomplishes its accuracy using a guided filter [116], which uses a guiding image to compute the filter output and reduce the errors and outliers created by a cost function. The cost function derives itself by considering the color and spatial gradients. At a resolution of 640 × 480 pixels and a disparity range of 40 pixels, the authors achieve a frame rate of 17 FPS on an NVIDIA GTX 480 (Santa Clara, CA, USA). An error of 5.55% was achieved when benchmarking against the MD. For better power consumption, [47] implemented their version of the ASW algorithm on an FPGA. The implementation refreshes the output at a frame rate of 60 FPS while processing an image with a resolution of 1280 × 720 pixels. In the MD the authors achieve an error of 6.36% with their algorithm. The results demonstrate that the optimized hardware approach by Ttofis and Theocharides [47] will suffer in terms of accuracy when compared to the work accomplished by He et al. [116].
As mentioned in Section 2.5, approaches with a focus on power consumption, like the publications [44,47,48,49,50] usually work with FPGAs or microcomputers. A comparison in power consumption between a CPU based-system with an Intel® Core 2 Duo CPU at 2 GHz, a GPU based-system with an NVIDIA GTX 280 (Santa Clara, CA, USA) and an Intel® Q6600 CPU at 2.4 GHz, and a Digital Signal Processor (DSP) based-system with a TI DSK, is given in the contribution of Humenberger et al. [44]. While processing data, the CPU consumes 57 W, the GPU 205 W, and the DSP uses 5 W. When computing the result of the Teddy stereo pair from the MD, the CPU achieves a frame rate of 12.89 FPS, the GPU reaches 105.4 FPS, and the DSP reaches 7.74 FPS. Zhao et al. [50] use the Semi-global matching (SGM) algorithm to compare an NVIDIA Titan X GPU (Santa Clara, CA, USA), with an NVIDIA Jetson TX2 embedded GPU, and with their own optimized implementation for an FPGA. The GPU is the fastest hardware with 238 FPS but consumes 101 W for the computation. The embedded GPU uses only 11.7 W, but only achieves 29 FPS. The authors manage to run the code while consuming 6.6 W on an FPGA and still reaching 161 FPS. Ttofis and Theocharides [47] obtain the on-chip power consumption for the entire system to 2.8 W, while reaching 60 FPS on an 1280 × 720 pixels and a maximum disparity range of 64 pixels. If power consumption is a concern, the FPGA can reach a high FPS per watt count, and still score an accurate result, as shown in [47,49]. When the highest computational power is needed, such as in an operation scenario, where real-time performance and accuracy are crucial, the GPU will be the best option.
In this state-of-the-art overview, the work of Totz et al. [102] and Stoyanov et al. [1] are some of the few binocular methods that exploit a feature-based matching approach instead of a correlation-based approach. The advantage is that feature-based methods do not require a rectified stereo pair to function. Therefore, this approach turns a 1D search problem along the epipolar line for rectified images into a 2D search problem. While this procedure increases the computational cost significantly, the authors argue that since an existing match is free to move along the image pyramid, which consists of downsamplings of the input image [117], in any direction, the number of successfully matched pixels will be enhanced. The result shown in the contribution by Totz et al. [102] is a sparse point cloud with an RMSE of 3.2 mm when compared to the CT ground truth. On an 1920 × 540 pixels image, a refresh rate of 2.64 FPS will be possible, when using a NVIDIA Quadro K5000 (Santa Clara, CA, USA). The sparse point cloud shown in the contribution of Totz et al. [102] may not be dense enough for a surgeon to see any details. Totz et al. mention a race condition in their algorithm. Hence, using multiple parallel threads could lead to different results on the same scene when running their algorithm multiple times. An improvement in terms of disparity error and speed, though at a lower resolution of 360 × 288 pixels, is presented by Stoyanov et al. [1]. The results here show semi-dense 3D reconstructions that are generated at 15 FPS with CUDA on an NVIDIA Quadro® FX 5800 (Santa Clara, CA, USA). For two different 3D reconstructions of a heart, the evaluation results show a disparity error of 0.89 mm ± 1.13 mm and 1.22 mm ± 1.71 mm, resulting in an average of 1.06 mm.
Song et al. [118] argue that despite the numerous proposed Deep Neural Networks (DNNs) approaches, conventional prior-free approaches remain popular due to the lack of open-source annotated datasets and the limitations of task-specific pre-trained DNNs. According to the authors, there was no successful real-time algorithm for non-GPU environments among the prior-free stereo matching algorithms. Therefore, they present the first CPU-level real-time prior-free stereo matching algorithm for MIS. The algorithm makes use of a patch-based fast disparity searching method, together with a coarse-to-fine Bayesian probability, and spatial Gaussian mixed model to evaluate the patch probability at different scales. Experiments demonstrated the method’s capability to handle ambiguities introduced by the textureless surfaces and photometric inconsistencies. When evaluating on an Intel i5-9400 CPU (Santa Clara, CA, USA), on the SCARED dataset, the method achieves an MAE of 2.158 mm with ~14 FPS at a resolution of 1280 × 720 pixels, and on the SERV-CT dataset, the method achieves an MAE of 3.750 mm with ~15 FPS at a resolution of 720 × 576 pixels. Please note that the full resolution of the SCARED dataset is 1280 × 1024. The code and a small synthetic dataset of a male colon with 310 stereo pairs at 640 × 480 pixels are available online.

5.2. Deep Learning Photogrammetry

Deep learning algorithms define the state-of-the-art for applications such as robotics, autonomous driving, and speech recognition. In medical stereo matching, given a well-defined dataset they produce similar results, in terms of accuracy, compared to the previously mentioned methods, based on the results of Allan et al. [15]. A summary of the passive real-time binocular methods that utilize Deep Learning and have been evaluated against the KITTI dataset can be found in Table 5.
Table 5. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI). The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are KITTI measurement of accuracy (see Section 2.3, in ascending order), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Table 5. Comparison of state-of-the-art real-time stereo matching algorithms evaluated against the Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI). The methods are listed with the parameters most relevant (see Section 2.2) for real-time stereo matching. The parameters are KITTI measurement of accuracy (see Section 2.3, in ascending order), frame rate, used hardware as well as image size and disparity range for the respective stereo pair.
Binocular Contribution
(KITTI Dataset)
Error
in %
FPS
in HZ
HardwareImage Size
in px
Disparity
Range in px
UASNet [119]1.643.32.5 GHz CPU 1242 × 375 ≤150
ACVNet [120]1.6552 × NVIDIA RTX 3090 1242 × 375 ≤150
LeaStereo [121]1.653.3NVIDIA V 100 1242 × 375 ≤150
CVCNet [122]1.7413.52 × NVIDIA RTX 2080 Ti 1242 × 375 ≤150
HITNet [123]1.9850NVIDIA GTX Titan V 1242 × 375 ≤150
HSM [124]2.147NVIDIA GTX Titan X 1242 × 375 ≤150
DeepPruner [125]2.155.54 × NVIDIA GTX Titan X 1242 × 375 ≤150
DispNetC [126]4.0515NVIDIA GTX Titan X 1242 × 375 ≤150
MADNet [127]4.6650NVIDIA GTX 1080 Ti 1242 × 375 ≤150
StereoNet [128]4.8360NVIDIA GTX Titan X 1242 × 375 ≤150
FP-Stereo [50]7.9147Xilinx ZCU102 (FPGA) 1242 × 374 128
CNN L12 [129]n.a.60NVIDIA GTX Titan X 620 × 188 n.a.
Deep learning algorithms used in 3D reconstruction can be classified into three categories (1) non-end-to-end, (2) end-to-end, and (3) unsupervised learning algorithms [130].
(1) Non-end-to-end algorithms do not include post-processing to optimize disparity estimation, such as outlier removal, in the network. Therefore, separate post-processing steps are still required to use these algorithms in practical applications [130]. A simplified structure of a Convolutional Neural Network (CNN) network can be seen in Figure 15. According to Zhou et al. [130], non-end-to-end algorithms can achieve a D1-all score of 1 pixel in the KITTI dataset. The disadvantage of this approach is high computational cost, which results in runtimes up to 67 s on an NVIDIA Titan X (Santa Clara, CA, USA) [131]. Also, due to the architecture and setup of the networks, the typical memory footprint is larger than that for deterministic methods. The limited receptive field and the lack of context are further disadvantages. Due to the long runtimes, these methods are not suited for real-time applications.
(2) End-to-end algorithms integrate the whole stereo pipeline into the network. Therefore, the stereo-matching problem becomes more straightforward when compared to non-end-to-end networks.
Huo et al. [132] present a binocular method that is based on StereoNet [128]. First, an initial disparity map is generated using Semi-Global Block Matching (SGBM). Then, gaps in the disparity map are filled with a self-developed algorithm and a confidence map is generated by comparing the right camera image with a generated right camera image, which is derived from the left camera image and the disparity map. The disparity confidence map is presented to StereoNet as a dataset for training. Thereafter, based on the depth map predicted by StereoNet, the corresponding left image of each depth map is input into the ORBSLAM framework. Color information is obtained from the left camera image, while ORBSLAM is used to estimate the camera pose, and StereoNet is used to acquire depth information. When compared against ground truth data that was generated with the aid of a Einscan-Pro-2x-Plus 3D scanner by Shining 3D (Hangzhou, China) and a phantom as well as a real pig stomach, the proposed method achieves an RMSE of 1.62 mm. The method reaches a framerate of 20 FPS with a resolution of 1280 × 720 pixels, the resulting stitched point cloud can be seen in Figure 16. It consists of 834,650 points, which ensures the real-time performance of the algorithm.
Tonioni et al. [127] published in 2019 the Modularly Adaptive Network (MADNet) algorithm. With its implementation, the network can adapt to new target domains, such as medical applications, which can be different than the one it was originally trained for. Compared to the state-of-the-art, it is one of the fastest networks available and therefore can run in real-time. When applied on the KITTI dataset it gains a D1-all score of 4.66% with a frame rate of 50 FPS using an NVIDIA GTX 1080 Ti GPU. In the medical domain, the dataset was evaluated on the SERV-CT dataset, on which it scored a Bad3 score of 26.58% ± 18.11% for the first sub-dataset and 38.24% ± 30.26% for the second sub-dataset at 35 FPS using an NVIDIA Tesla V100 GPU. As for the RMSE in each sub dataset, it achieved 4.23 mm ± 1.42 mm for the first one and 6.31 mm ± 3.13 mm for the second one. DeepPruner by Duggal et al. [125] uses a differentiable PatchMatch module. This approach allows the authors to mark a disparity as invalid without requiring a full cost volume evaluation, resulting in quicker runtimes. The minimized disparity range propagated to further refine the disparity range. On the KITTI dataset the algorithm scores a D1-all score of 2.15% with a frame rate of 5.5 FPS using four NVIDIA Titan X GPU (Santa Clara, CA, USA). Edwards et al. [17] evaluated the algorithm on the SERV-CT dataset, on which it scored a Bad3 score of 12.50% ± 7.44% for the first sub-dataset and 19.13% ± 16.95% for the second sub-dataset at 16 FPS using an NVIDIA Tesla V100 GPU. As for the RMSE in each sub-dataset, it achieved 2.91 mm ± 1.71 mm for the first one and 3.21 mm ± 1.31 mm for the second one. Yang et al. [124] use an hierarchical deep stereo matching (HSM) network to develop an approach that can estimate depth quickly. To keep memory consumption low for high-resolution stereo images the authors use an incremental approach to search for correspondences. On the KITTI dataset the algorithm achieves a D1-all score of 2.14% with a frame rate of 7 FPS using an NVIDIA Titan X GPU (Santa Clara, CA, USA). In the publication [17] the HSM network is run on the SERV-CT dataset to evaluate it for medical applications. On this benchmark, it performed as one of the best with a Bad3 score of 8.34% ± 8.31% for the first sub-dataset and 5.46% ± 2.96% for the second sub-dataset at 20 FPS using an NVIDIA Tesla V100 GPU. As for the RMSE in each sub-dataset, it achieved 3.18 mm ± 2.03 mm for the first one and 2.12 mm ± 0.54 mm for the second one. The DispNetC network reaches 15 FPS in the KITTI dataset and a D1-all score of 4.05% [126]. To train the network, an artificially created dataset with more than 35,000 labeled training samples is used. It was also run on the SERV-CT dataset and achieved there a Bad3 score of 40.09% ± 26.41% for the first sub-dataset and 47.87% ± 27.58% for the second sub-dataset at 20 FPS using an NVIDIA Tesla V100 GPU [17]. In the SERV-CT dataset they also calculated the RMSE for each sub-dataset, as 4.58 mm ± 0.76 mm for the first one and 7.07 mm ± 4.70 mm for the second one. As there are no sufficiently large laparoscopic datasets available, ref. [17] used a pre-trained version (against KITTI) of the DispNetC for their evaluation. HITNet runs on an NVIDIA Titan V GPU (Santa Clara, CA, USA) at 50 FPS and manages a D1-all score in the KITTI dataset of 0.5 pixels. As of today, HITNet has the highest accuracy of the above 10 FPS achieving networks concerning the Middelburry-V3 dataset, the KITTI 2012 and 2015 dataset, and the ETH3D benchmark for 3D reconstruction [123]. The UASNet network utilizes an uncertainty distribution guided range prediction model and an uncertainty-based disparity sampler module to achieve a D1-all score of 1.64% at a frame rate of 3.3 FPS on a 2.5 GHz CPU in the benchmark [119].
Cheng et al. [121] developed a method exploiting Neural Architecture Search (NAS) that performs very similarly on the KITTI dataset. The authors achieve a D1-all score of 1.65% at a frame rate of 3.3 FPS on an NVIDIA V 100 (Santa Clara, CA, USA). In the contribution by Xu et al. [120], the authors use Attention Concatenation Volume (ACV) for their network to acquire a D1-all score of 1.65% at a frame rate of 5 FPS on two NVIDIA RTX 3090. Zhu et al. [122] apply super-resolution, which is an upsampling of the input image, to improve the results of their cross-view capture network. Using two NVIDIA RTX 2080 Ti, they attain a D1-all score of 1.74% at a frame rate of 13.5 FPS. StereoNet has a D1-all score of 0.9 pixels in the KITTI dataset and achieves a frame rate of 60 FPS on an NVIDIA Titan X GPU (Santa Clara, CA, USA). The authors also mention a sparse availability of supervised training data. Their aim is to accomplish the same results with an unsupervised learning network [128]. The disadvantages of end-to-end networks are that they require large amounts of data, in the tens of thousands of samples of labeled training data. Unfortunately, there are not many data samples for in vivo procedures with precise ground truth data. Furthermore, the networks are not robust against changes in application because the descriptors must be adjusted for each dataset [130]. In the contribution by Mayer et al. [133], the authors demonstrate that for specific applications, a generalized training dataset is sufficient for an algorithm to perform well later in its application. Therefore, for this approach, a specialized training dataset consisting of real-world data may not be needed.
Chen et al. [134] present a framework for Real-time Scene Reconstruction (FRSR) that is designed to perform a 3D reconstruction of a surgical site in real time. The authors highlight the strong independence of their lightweight encoder–decoder network that performs the disparity estimation. The framework was evaluated via a scene from the da Vinci Research Kit (dVRK) endoscope and more importantly, the SCARED dataset and a self-made clinical dataset captured from an oncology hospital which was also published. The published dataset contains 145,694 image pairs with a resolution of 1920 × 1080 pixels, and calibration parameters are also provided for image rectification and triangulation. The stereo images were recorded during a clinical procedure named Radical Prostatectomy with lymphadenectomy using da Vinci Xi surgical system and a 3D HD video recorder HVO-3300MT by SONY (Basingstoke, UK). The explicit details on how the ground truth in this dataset was acquired are not provided. When evaluating against the SCARED dataset, 25 FPS and an MAE of 2.69 mm ± 1.48 mm and RMSE of 5.47 mm ± 1.34 mm are achieved.
(3) In contrast to end-to-end networks, unsupervised deep learning algorithms approach the disparity estimation problem by minimizing photometric warping errors. The tradeoff is a loss in accuracy [130], which results in an RMSE in the KITTI benchmark of 5.104 mm for well-trained networks while achieving a frame rate of 60 FPS on an NVIDIA Titan X GPU (Santa Clara, CA, USA) [129]. According to Zhou et al. [130], the low quality of unsupervised networks can be traced back to the failure to provide a strong impulse to the network to let it converge correctly when an image reconstruction error occurs.
Self-supervised stacked and Siamese encoder-decoder neural networks to compute accurate disparity maps for 3D laparoscopy are proposed by Bardozzo et al. [135]. These networks are evaluated on three different public datasets, namely the Hamlyn dataset, the SCARED dataset, and the phantom cardiac dataset. Additionally, simulated datasets are generated; however, they are available only upon request and exclusively for research collaborations. The presented networks are capable of producing disparities in real-time on standard GPU-equipped desktop computers. The evaluation results showed that these networks outperformed state-of-the-art methods and achieved an RMSE of 0.29 mm ± 0.05 mm and an MAE of 0.22 mm ± 0.05 mm at ~18 FPS against the SCARED dataset. Finally, a fuzzy post-processing strategy is proposed that slightly improves depth estimation by 0.01 mm at the cost of speed, resulting in ~6.5 FPS.
Large networks need large amounts of domain-specific data for training, which is one of the major drawbacks of deep learning. In recent years research has shown that large amounts of synthetic data can improve the performance of learning-based vision algorithms, which has the advantage that the challenge of generating real-world data for unique scenarios can be avoided. Domain adaptation algorithms are utilized to avoid bigger gaps between simulation and real-world data [35].
Medical devices must be thoroughly validated since human lives may depend on them. Therefore, deep learning and other AI-based systems must pass through a particularly time and resource-consuming validation process. The approval process can be a significant barrier to adopting AI systems in clinical use, according to Benjamens et al. [136]. The FDA proposes a “total product lifecycle” regulatory approach for these methods to remedy that problem. This regulatory approach incorporates continuous learning and improvement of a product in its lifecycle while guaranteeing that safeguards stay in place and the product’s effectiveness is not impacted [137].

6. Multi-Ocular: Passive Methods

Trinocular methods use three camera sensors to create a 3D reconstruction of a scene, as illustrated in Figure 17. Here, the 3D reconstruction process is similar to the one of binocular applications. With three cameras, however, the reconstruction is performed between all possible camera pairs and the depth information is merged, respectively.
Based on the algorithm by Heinrichs et al. [138] a prototype with a triple camera setup is presented by Conen et al. [139]. The results of the trinocular setup are compared to a binocular setup running an SGBM algorithm [140] from the OpenCV library. Conen and Luhmann [89] determine the MAE to the ground truth for the trinocular setup to be 0.29 mm with a SD of 1.14 mm, while the trinocular endoscope is positioned 210 mm away from the object. For the binocular setup, the authors record an MAE to the ground truth of 0.21 mm with a SD of 1.48 mm for the same distance setup. However, in comparison to the binocular setup, which has a typical diameter of 10 mm, the trinocular setup in [139] has a diameter of 14 mm. The authors do not mention the runtime of the algorithm, therefore it has to be assumed that it does not run in real time.
Other multi-ocular approaches use more than three camera sensors or lenses to create 3D reconstructions. Please note that all publications discussed in this section are proof of concepts. Therefore, only limited information is available regarding the subject.
Hassanfiroozi et al. [141] implemented an endoscope using liquid crystal lenses and a single image sensor. The lenses are tuneable by applying a voltage and can switch electronically between 2D and 3D modes. In 2D mode, the lenses pass light rays directly through to an image sensor without focusing, whereas in 3D mode, the light rays are focused on different parts of the image sensor. Even though the stereo basis is small, the lenses can capture an image from different viewpoints such that a 3D reconstruction is possible. The authors focused on the proof of concept and did not further pursue the accuracy or the 3D reconstruction. A year later Hassanfiroozi [142] improved the concept using a multi-liquid crystal three-lens design, which allowed focusing the lenses in 2D mode. The topics of camera calibration, accuracy, and 3D reconstruction were not investigated.

7. Conclusions

The growing demand for depth perception in the field of computer vision and technical advancements in recent years have enabled real-time 3D reconstruction methods for medical endoscopy. Some methods require modifications to the existing MIS hardware to function, while others solely rely on the visual input of the camera system. Either way, the fields of navigation, AR, robotics, and many more have greatly benefited from the ongoing research.
Capturing and displaying high-resolution images will always be one step ahead of complex image analysis. As a result, there is currently no abundance of real-time 3D reconstruction methods for high-resolution medical imaging. In addition to this, the comparison between the relevant contributions is challenging. As of today, there is no clear consensus on how to evaluate newly developed methods, and therefore evaluation methods range from open-access datasets to self-constructed ones. In addition to this, authors use different error metrics to evaluate their methods if not specified by a dataset. However, only when publications apply the same dataset and error metric, will an objective and independent comparison between them be possible. Another advantage of using datasets with a specified error metric is the opportunity to re-evaluate a developed method on a dataset from a different field of work. When comparing publications, it becomes clear that some methods being developed for a specific application show promising results when evaluated against another application. One example of this is SLAM, which was initially developed for robot navigation, but also shows promising results in the field of medical endoscopy with minor modifications.
It is good practice to publish the used dataset and hardware together with the runtime, accuracy, and precision of a method to allow for an objective comparison between contributions. If these parameters are not provided, an objective comparison will not be possible.
In our contribution, we aim to provide an objective and comprehensive overview of the current state-of-the-art concerning real-time 3D reconstruction methods for medical applications. The presented methods are first categorized, and subsequently, the most relevant work is compared and grouped into tables, while taking into account different evaluation methods. Furthermore, the evaluation methods used in the compared contributions, including relevant datasets, are outlined. In this, we provide a comprehensive, well-founded starting point, to identify which approach is most feasible for the medical endoscopy application in mind.

Author Contributions

Conceptualization, A.R., T.S. and S.J.R.; Data curation, A.R., T.S. and J.-C.R.; Formal analysis, A.R. and T.S.; Investigation, A.R., T.S. and S.J.R.; Methodology, A.R. and T.S.; Project administration, A.R. and S.J.R.; Software, A.R. and T.S.; Supervision, S.J.R.; Validation, A.R. and T.S.; Visualization, A.R., T.S. and J.-C.R.; Writing—review and editing, A.R., T.S., J.-C.R. and S.J.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by Fraunhofer Institute for High-Speed Dynamics, Ernst–Mach–Institut, EMI (Freiburg, Germany) and Schölly Fiberoptic GmbH (Denzlingen, Germany).

Data Availability Statement

The data that supports the findings of this work is available from the corresponding publishers, but restrictions may apply to the availability of this data. Some of the referenced work was accessed under license for the current study, and so is not publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Stoyanov, D.; Visentini-Scarzanella, M.; Pratt, P.; Yang, G.Z. Real-Time Stereo Reconstruction in Robotically Assisted Minimally Invasive Surgery. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention: MICCAI International Conference on Medical Image Computing and Computer-Assisted Intervention, Beijing, China, 20–24 September 2010; Volume 13, pp. 275–282. [Google Scholar] [CrossRef]
  2. Lin, B.; Johnson, A.; Qian, X.; Sanchez, J.; Sun, Y. Simultaneous Tracking, 3D Reconstruction and Deforming Point Detection for Stereoscope Guided Surgery. In Proceedings of the Augmented Realitcy Environments for Medical Imaging and Computer-Assisted Interventions, Nagoya, Japan, 22–26 September 2013; Liao, H., Linte, C.A., Masamune, K., Peters, T.M., Zheng, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 35–44. [Google Scholar] [CrossRef]
  3. Bogdanova, R.; Boulanger, P.; Zheng, B. Depth Perception of Surgeons in Minimally Invasive Surgery. Surg. Innov. 2016, 23, 515–524. [Google Scholar] [CrossRef] [PubMed]
  4. Breedveld, P.; Wentink, M. Eye-hand coordination in laparoscopy—An overview of experiments and supporting aids. Minim. Invasive Ther. Allied Technol. 2009, 10, 155–162. [Google Scholar] [CrossRef] [PubMed]
  5. Way, L.; Stewart, L.; Gantert, W.; Liu, K.; Lee, C.; Whang, K.; Hunter, J. Causes and Prevention of Laparoscopic Bile Duct Injuries: Analysis of 252 Cases from a Human Factors and Cognitive Psychology Perspective. Ann. Surg. 2003, 237, 460–469. [Google Scholar] [CrossRef] [PubMed]
  6. Skogås, J.G.; Myhre, H.O.; Ødegård, A.; Kiss, G. Imaging for Minimally Invasive Surgery. Surg. Sci. 2016, 7, 333–341. [Google Scholar] [CrossRef]
  7. Erpenbeck, D.; Lemke, N.; Wittenberg, T.; Rosenthal, J.C.; Kauff, P.; Bergen, T. Evaluierung von 3D-Rekonstruktionsverfahren in der Stereo-Laparoskopie. In Proceedings of the 15th Annual Conference of the German Society for Computer- and Robot-Assisted Surgery (CURAC 2016), Bern, Schwitzerland, 29 September–1 October 2016; pp. 58–63, ISBN 978-3-86247-595-7. [Google Scholar]
  8. Liu, X.; Karmarkar, P.; Voit, D.; Frahm, J.; Weiss, C.; Kraitchman, D.; Bottomley, P. Real-Time High-Resolution MRI Endoscopy at up to 10 Frames per Second. BME Front. 2021, 2021, 6185616. [Google Scholar] [CrossRef] [PubMed]
  9. Trucco, E.; Verri, A. Introductory Techniques for 3-D Computer Vision; Prentice Hall PTR: Hoboken, NJ, USA, 1998; ISBN 978-0-13-261108-4. [Google Scholar]
  10. Min, D.; Yea, S.; Arican, Z.; Vetro, A. Disparity search range estimation: Enforcing temporal consistency. In Proceedings of the 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 2366–2369. [Google Scholar] [CrossRef]
  11. Sara, R. Finding the Largest Unambiguous Component of Stereo Matching. In Proceedings of the Computer Vision—ECCV 2002, Copenhagen, Denmark, 28–31 May 2002; pp. 900–914. [Google Scholar] [CrossRef]
  12. Cyganek, B.; Borgosz, J. An Improved Variogram Analysis of the Maximum Expected Disparity in Stereo Images. In Proceedings of the SCIA’03: Proceedings of the 13th Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2003; Volume 2749, pp. 640–645. [Google Scholar] [CrossRef]
  13. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  14. Groch, A.; Hempel, S.; Speidel, S.; Höller, K.; Engelbrecht, R.; Penne, J.; Seitel, A.; Röhl, S.; Yung, K.; Bodenstedt, S.; et al. In-vitro Evaluation von endoskopischer Oberflächenrekonstruktion mittels Time-of-Flight-Kameratechnik. In Proceedings of the Bildverarbeitung für die Medizin 2011: Algorithmen–Systeme–Anwendungen, Lübeck, Germany, 20–22 March 2011; pp. 184–188. [Google Scholar] [CrossRef]
  15. Allan, M.; McLeod, A.J.; Wang, C.C.; Rosenthal, J.C.; Hu, Z.; Gard, N.; Eisert, P.; Fu, K.; Zeffiro, T.; Xia, W.; et al. Stereo Correspondence and Reconstruction of Endoscopic Data Challenge. arXiv 2021, arXiv:2101.01133. [Google Scholar] [CrossRef]
  16. Scharstein, D.; Hirschmüller, H.; Kitajima, Y.; Krathwohl, G.; Nešić, N.; Wang, X.; Westling, P. High-Resolution Stereo Datasets with Subpixel-Accurate Ground Truth. In Proceedings of the 36th Conference on Pattern Recognition, Münster, Germany, 2–5 September 2014; pp. 31–42. [Google Scholar] [CrossRef]
  17. Edwards, P.; Psychogyios, D.; Speidel, S.; Maier-Hein, L.; Stoyanov, D. SERV-CT: A disparity dataset from CT for validation of endoscopic 3D reconstruction. arXiv 2020, arXiv:2012.11779. [Google Scholar] [CrossRef]
  18. Chai, T.; Draxler, R. Root mean square error (RMSE) or mean absolute error (MAE)? Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef]
  19. Menze, M.; Geiger, A. Object scene flow for autonomous vehicles. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3061–3070. [Google Scholar] [CrossRef]
  20. Schauwecker, K. Real-time stereo vision on FPGAs with SceneScan. In Proceedings of the Forum Bildverarbeitung, Karlsruhe, Germany, 29–30 November 2018; p. 339. [Google Scholar]
  21. Scharstein, D.; Szeliski, R. A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms. Int. J. Comput. Vis. 2002, 47, 7–42. [Google Scholar] [CrossRef]
  22. Scharstein, D.; Szeliski, R. High-accuracy stereo depth maps using structured light. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 1. [Google Scholar] [CrossRef]
  23. Scharstein, D.; Pal, C. Learning Conditional Random Fields for Stereo. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 16 July 2007; pp. 1–8. [Google Scholar] [CrossRef]
  24. Hirschmuller, H.; Scharstein, D. Evaluation of Cost Functions for Stereo Matching. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 16 July 2007; pp. 1–8. [Google Scholar] [CrossRef]
  25. Stiller, C.; Kammel, S.; Pitzer, B.; Ziegler, J.; Werling, M.; Gindele, T.; Jagszent, D. Team AnnieWAY’s Autonomous System. In Proceedings of the Robot Vision, Second International Workshop, RobVis 2008, Auckland, New Zealand, 18–20 February 2008; pp. 248–259. [Google Scholar] [CrossRef]
  26. Mountney, P.; Stoyanov, D.; Yang, G.-Z.; Mylonas, G.; Deligianni, F.; Darzi, A.; Lerotic, M.; Chung, A.J.; Clark, J.; Valibeik, S.; et al. Hamlyn Centre Laparoscopic/Endoscopic Video Datasets. 2021. Available online: https://hamlyn.doc.ic.ac.uk/vision/ (accessed on 9 September 2022).
  27. Mountney, P.; Stoyanov, D.; Yang, G.Z. Three-Dimensional Tissue Deformation Recovery and Tracking. IEEE Signal Process. Mag. 2010, 27, 14–24. [Google Scholar] [CrossRef]
  28. Stoyanov, D.; Mylonas, G.P.; Deligianni, F.; Darzi, A.; Yang, G.Z. Soft-Tissue Motion Tracking and Structure Estimation for Robotic Assisted MIS Procedures. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2005, Palm Springs, CA, USA, 26–29 October 2005; Duncan, J.S., Gerig, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2005; pp. 139–146. [Google Scholar] [CrossRef]
  29. Lerotic, M.; Chung, A.; Clark, J.; Valibeik, S.; Yang, G.Z. Dynamic View Expansion for Enhanced Navigation in Natural Orifice Transluminal Endoscopic Surgery. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention: MICCAI International Conference on Medical Image Computing and Computer-Assisted Intervention, New York, NY, USA, 6–10 September 2008; Volume 11, pp. 467–475. [Google Scholar] [CrossRef]
  30. Pratt, P.; Stoyanov, D.; Visentini-Scarzanella, M.; Yang, G.Z. Dynamic Guidance for Robotic Surgery Using Image-Constrained Biomechanical Models. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention: MICCAI International Conference on Medical Image Computing and Computer-Assisted Intervention, Beijing, China, 20–24 September 2010; Volume 13, pp. 77–85. [Google Scholar] [CrossRef]
  31. Giannarou, S.; Visentini-Scarzanella, M.; Yang, G.Z. Probabilistic Tracking of Affine-Invariant Anisotropic Regions. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 130–143. [Google Scholar] [CrossRef]
  32. Ye, M.; Giannarou, S.; Meining, A.; Yang, G.Z. Online Tracking and Retargeting with Applications to Optical Biopsy in Gastrointestinal Endoscopic Examinations. Med. Image Anal. 2015, 30, 144–157. [Google Scholar] [CrossRef] [PubMed]
  33. Ye, M.; Johns, E.; Handa, A.; Zhang, L.; Pratt, P.; Yang, G. Self-Supervised Siamese Learning on Stereo Image Pairs for Depth Estimation in Robotic Surgery. In Proceedings of the Hamlyn Symposium on Medical Robotics, London, UK, 25–28 June 2017; pp. 27–28. [Google Scholar] [CrossRef]
  34. Geiger, A.; Roser, M.; Urtasun, R. Efficient Large-Scale Stereo Matching. In Proceedings of the Computer Vision—ACCV 2010, Queenstown, New Zealand, 8–12 November 2010; pp. 25–38. [Google Scholar] [CrossRef]
  35. Ozyoruk, K.B.; Gokceler, G.I.; Coskun, G.; Incetan, K.; Almalioglu, Y.; Mahmood, F.; Curto, E.; Perdigoto, L.; Oliveira, M.; Sahin, H.; et al. EndoSLAM Dataset and An Unsupervised Monocular Visual Odometry and Depth Estimation Approach for Endoscopic Videos: Endo-SfMLearner. arXiv 2020, arXiv:2006.16670. [Google Scholar] [CrossRef] [PubMed]
  36. Rau, A.; Bano, S.; Jin, Y.; Azagra, P.; Morlana, J.; Sanderson, E.; Matuszewski, B.J.; Lee, J.Y.; Lee, D.J.; Posner, E.; et al. SimCol3D—3D Reconstruction during Colonoscopy Challenge. arXiv 2023, arXiv:2307.11261. [Google Scholar]
  37. Penza, V.; Ciullo, A.; Moccia, S.; Mattos, L.; De Momi, E. EndoAbS Dataset: Endoscopic Abdominal Stereo Image Dataset for Benchmarking 3D Stereo Reconstruction Algorithms. Int. J. Med. Robot. Comput. Assist. Surg. 2018, 14, e1926. [Google Scholar] [CrossRef]
  38. Bobrow, T.L.; Golhar, M.; Vijayan, R.; Akshintala, V.S.; Garcia, J.R.; Durr, N.J. Colonoscopy 3D video dataset with paired depth from 2D-3D registration. Med. Image Anal. 2023, 90, 102956. [Google Scholar] [CrossRef]
  39. Rau, A.; Edwards, P.; Ahmad, O.; Riordan, P.; Janatka, M.; Lovat, L.; Stoyanov, D. Implicit domain adaptation with conditional generative adversarial networks for depth prediction in endoscopy. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1167–1176. [Google Scholar] [CrossRef] [PubMed]
  40. HajiRassouliha, A.; Taberner, A.; Nash, M.; Nielsen, P. Suitability of recent hardware accelerators (DSPs, FPGAs, and GPUs) for computer vision and image processing algorithms. Signal Process. Image Commun. 2018, 68, 101–119. [Google Scholar] [CrossRef]
  41. Ghorpade, J.; Parande, J.; Kulkarni, M.; Bawaskar, A. GPGPU Processing in CUDA Architecture. arXiv 2012, arXiv:1202.4347. [Google Scholar] [CrossRef]
  42. Pauwels, K.; Tomasi, M.; Alonso, J.D.; Ros, E.; Hulle, M. A Comparison of FPGA and GPU for Real-Time Phase-Based Optical Flow, Stereo, and Local Image Features. IEEE Trans. Comput. 2012, 61, 999–1012. [Google Scholar] [CrossRef]
  43. Qasaimeh, M.; Denolf, K.; Lo, J.; Vissers, K.; Zambreno, J.; Jones, P.H. Comparing Energy Efficiency of CPU, GPU and FPGA Implementations for Vision Kernels. In Proceedings of the 2019 IEEE International Conference on Embedded Software and Systems (ICESS), Las Vegas, NV, USA, 2–3 June 2019; pp. 1–8. [Google Scholar] [CrossRef]
  44. Humenberger, M.; Zinner, C.; Weber, M.; Kubinger, W.; Vincze, M. A Fast Stereo Matching Algorithm Suitable for Embedded Real-Time Systems. Comput. Vis. Image Underst. 2010, 114, 1180–1202. [Google Scholar] [CrossRef]
  45. Feng, X.; Jiang, Y.; Yang, X.; Du, M.; Li, X. Computer vision algorithms and hardware implementations: A survey. Integration 2019, 69, 309–320. [Google Scholar] [CrossRef]
  46. Carballo-Hern’andez, W.; Pelcat, M.; Berry, F. Why is FPGA-GPU Heterogeneity the Best Option for Embedded Deep Neural Networks? arXiv 2021, arXiv:2102.01343. [Google Scholar] [CrossRef]
  47. Ttofis, C.; Theocharides, T. High-quality real-time hardware stereo matching based on guided image filtering. In Proceedings of the 2014 Design, Automation & Test in Europe Conference & Exhibition (DATE), Dresden, Germany, 24–28 March 2014; pp. 1–6. [Google Scholar] [CrossRef]
  48. Zhang, L.; Zhang, K.; Chang, T.S.; Lafruit, G.; Kuzmanov, G.K.; Verkest, D. Real-Time High-Definition Stereo Matching on FPGA. In Proceedings of the 19th ACM/SIGDA International Symposium on Field Programmable Gate Arrays (FPGA’11), Monterey, CA, USA, 27 February–1 March 2011; pp. 55–64. [Google Scholar] [CrossRef]
  49. Wang, W.; Yan, J.; Xu, N.; Wang, Y.; Hsu, F.H. Real-time high-quality stereo vision system in FPGA. In Proceedings of the 2013 International Conference on Field-Programmable Technology (FPT), Kyoto, Japan, 9–11 December 2013; pp. 358–361. [Google Scholar] [CrossRef]
  50. Zhao, J.; Liang, T.; Feng, L.; Ding, W.; Sinha, S.; Zhang, W.; Shen, S. FP-Stereo: Hardware-Efficient Stereo Vision for Embedded Applications. arXiv 2020, arXiv:2006.03250. [Google Scholar]
  51. Sun, D.; Liu, J.; Linte, C.A.; Duan, H.; Robb, R.A. Surface Reconstruction from Tracked Endoscopic Video Using the Structure from Motion Approach. In Proceedings of the Augmented Reality Environments for Medical Imaging and Computer-Assisted Interventions, Nagoya, Japan, 22–26 September 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 127–135. [Google Scholar] [CrossRef]
  52. Chen, L.; Tang, W.; John, N.W.; Wan, T.R.; Zhang, J.J. SLAM-based dense surface reconstruction in monocular Minimally Invasive Surgery and its application to Augmented Reality. Comput. Methods Programs Biomed. 2018, 158, 135–146. [Google Scholar] [CrossRef] [PubMed]
  53. Mahmoud, N.; Cirauqui, I.; Hostettler, A.; Doignon, C.; Soler, L.; Marescaux, J.; Montiel, J.M. ORBSLAM-based endoscope tracking and 3D reconstruction. In Computer-Assisted and Robotic Endoscopy; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2017; Volume 10170, pp. 72–83. [Google Scholar] [CrossRef]
  54. Recasens, D.; Lamarca, J.; Fácil, J.M.; Montiel, J.M.M.; Civera, J. Endo-Depth-and-Motion: Localization and Reconstruction in Endoscopic Videos using Depth Networks and Photometric Constraints. arXiv 2021, arXiv:2103.16525. [Google Scholar] [CrossRef]
  55. Vilches, S.; Zappe, H.; Ataman, Ç. Multi-Point Fiber-Optic Distance Sensor for Endoscopic Surgery Monitoring. IEEE Photonics Technol. Lett. 2023, 35, 883–886. [Google Scholar] [CrossRef]
  56. Schmalz, C.; Forster, F.; Schick, A.; Angelopoulou, E. An endoscopic 3D scanner based on structured light. Med. Image Anal. 2012, 16, 1063–1072. [Google Scholar] [CrossRef]
  57. Jia, T.; Zhou, Z.; Gao, H.; Wang, B. Depth Measurement Based on Infrared Coded Structured Light. Hongwai Yu Jiguang Gongcheng/Infrared Laser Eng. 2015, 44, 1628–1632. [Google Scholar] [CrossRef]
  58. Lin, J.; Clancy, N.T.; Elson, D.S. An endoscopic structured light system using multispectral detection. Int. J. Comput. Assist. Surg. 2015, 10, 1941–1950. [Google Scholar] [CrossRef] [PubMed]
  59. Godard, C.; Aodha, O.M.; Brostow, G.J. Digging Into Self-Supervised Monocular Depth Estimation. arXiv 2018, arXiv:1806.01260. [Google Scholar]
  60. Yang, Y.; Shao, S.; Yang, T.; Wang, P.; Yang, Z.; Wu, C.; Liu, H. A geometry-aware deep network for depth estimation in monocular endoscopy. Eng. Appl. Artif. Intell. 2023, 122, 105989. [Google Scholar] [CrossRef]
  61. Vélez, A.F.M.; Marcinczak, J.M.; Grigat, R.R. Structure from motion based approaches to 3D reconstruction in minimal invasive laparoscopy. In Proceedings of the 9th International Conference on Image Analysis and Recognition, ICIAR 2012; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; pp. 296–303. [Google Scholar] [CrossRef]
  62. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef]
  63. Malti, A.; Bartoli, A.; Collins, T. Template-Based Conformal Shape-from-Motion-and-Shading for Laparoscopy. In Proceedings of the Information Processing in Computer-Assisted Interventions; Abolmaesumi, P., Joskowicz, L., Navab, N., Jannin, P., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 1–10. [Google Scholar] [CrossRef]
  64. Malti, A. Variational Formulation of the Template-Based Quasi-Conformal Shape-from-Motion from Laparoscopic Images. Int. J. Adv. Comput. Sci. Appl. 2014, 5, 162–172. [Google Scholar] [CrossRef]
  65. Wang, R.; Price, T.; Zhao, Q.; Frahm, J.M.; Rosenman, J.; Pizer, S. Improving 3D surface reconstruction from endoscopic video via fusion and refined reflectance modeling. In Proceedings of the Medical Imaging 2017: Image Processing, Orlando, FL, USA, 12–14 February 2017; Styner, M.A., Angelini, E.D., Eds.; International Society for Optics and Photonics SPIE: Bellingham, WA, USA, 2017; pp. 80–86. [Google Scholar] [CrossRef]
  66. Turan, M.; Pilavci, Y.Y.; Ganiyusufoglu, I.; Araújo, H.; Konukoglu, E.; Sitti, M. Sparse-then-Dense Alignment based 3D Map Reconstruction Method for Endoscopic Capsule Robots. arXiv 2017, arXiv:1708.09740. [Google Scholar] [CrossRef]
  67. Widya, A.R.; Monno, Y.; Imahori, K.; Okutomi, M.; Suzuki, S.; Gotoda, T.; Miki, K. 3D Reconstruction of Whole Stomach from Endoscope Video Using Structure-from-Motion. In Proceedings of the 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2019, Berlin, Germany, 23–27 July 2019; pp. 3900–3904. [Google Scholar] [CrossRef]
  68. Rukhovich, D.; Mouritzen, D.; Kaestner, R.; Rufli, M.; Velizhev, A. Estimation of Absolute Scale in Monocular SLAM Using Synthetic Data. arXiv 2019, arXiv:1909.00713. [Google Scholar]
  69. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  70. Mur-Artal, R.; Montiel, J.M.M.; Tardós, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. arXiv 2015, arXiv:1502.00956. [Google Scholar] [CrossRef]
  71. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar] [CrossRef]
  72. Mahmoud, N.; Collins, T.; Hostettler, A.; Soler, L.; Doignon, C.; Montiel, J. Live Tracking and Dense Reconstruction for Hand-held Monocular Endoscopy. IEEE Trans. Med Imaging 2018, 24, 3074–3086. [Google Scholar] [CrossRef]
  73. Blender Foundation. Home of the Blender Project—Free and Open 3D Creation Software. Available online: https://www.blender.org/ (accessed on 9 September 2022).
  74. Geng, J. Structured-light 3D surface imaging: A tutorial. Adv. Opt. Photonics 2011, 3, 128–160. [Google Scholar] [CrossRef]
  75. Albitar, C.; Graebling, P.; Doignon, C. Robust Structured Light Coding for 3D Reconstruction. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007; pp. 1–6. [Google Scholar] [CrossRef]
  76. Lee, D.; Schachter, B. Two Algorithms for Constructing a Delaunay Triangulation. Int. J. Parallel Program. 1980, 9, 219–242. [Google Scholar] [CrossRef]
  77. Vollmer, J.; Mencl, R.; Müller, H. Improved Laplacian Smoothing of Noisy Surface Meshes. Comput. Graph. Forum 1999, 18, 131–138. [Google Scholar] [CrossRef]
  78. Kazhdan, M.; Bolitho, M.; Hoppe, H. Poisson Surface Reconstruction. In Proceedings of the Fourth Eurographics Symposium on Geometry Processing, SGP ’06, Cagliari, Italy, 26–28 June 2006; Eurographics Association: Eindhoven, The Netherlands, 2006; pp. 61–70. [Google Scholar] [CrossRef]
  79. Recherche, E.; Automatique, E.; Antipolis, S.; Zhang, Z. Iterative Point Matching for Registration of Free-Form Curves and Surfaces. Int. J. Comput. Vis. 1992, 13, 119–152. [Google Scholar] [CrossRef]
  80. Foix, S.; Alenya, G.; Torras, C. Lock-in Time-of-Flight (ToF) Cameras: A Survey. IEEE Sens. J. 2011, 11, 1917–1926. [Google Scholar] [CrossRef]
  81. Kolb, A.; Barth, E.; Koch, R.; Larsen, R. Time-of-Flight Sensors in Computer Graphics. In Proceedings of the EUROGRAPHICS (State of the Art Report); Eurographics Association: Eindhoven, The Netherlands, 2009. [Google Scholar] [CrossRef]
  82. Kolb, A.; Barth, E.; Koch, R. ToF-sensors: New dimensions for realism and interactivity. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops, Anchorage, AK, USA, 23–28 June 2008; pp. 1–6. [Google Scholar] [CrossRef]
  83. Laković, N.; Brkić, M.; Batinić, B.; Bajić, J.; Rajs, V.; Kulundžić, N. Application of low-cost VL53L0X ToF sensor for robot environment detection. In Proceedings of the 2019 18th International Symposium INFOTEH-JAHORINA (INFOTEH), Jahorina, East Sarajevo, Bosnia and Herzegovina, 23–28 June 2019; pp. 1–4. [Google Scholar] [CrossRef]
  84. De Ponte Müller, F. Survey on Ranging Sensors and Cooperative Techniques for Relative Positioning of Vehicles. Sensors 2017, 17, 271. [Google Scholar] [CrossRef] [PubMed]
  85. Groch, A.; Haase, S.; Wagner, M.; Kilgus, T.; Kenngott, H.; Schlemmer, H.P.; Hornegger, J.; Meinzer, H.P.; Maier-Hein, L. Optimierte endoskopische Time-of-Flight Oberflächenrekonstruktion durch Integration eines Struktur-durch-Bewegung-Ansatzes. In Proceedings of the Bildverarbeitung für die Medizin 2012: Algorithmen–Systeme—Anwendungen, Berlin, Germany, 18–20 March 2012; Springer: Berlin/Heidelberg, Germany, 2012. Chapter V9. pp. 39–44. [Google Scholar] [CrossRef]
  86. Roehl, S.; Bodenstedt, S.; Suwelack, S.; Kenngott, H.; Mueller-Stich, B.; Dillmann, R.; Speidel, S. Real-time surface reconstruction from stereo endoscopic images for intraoperative registration. In Proceedings of the MedicalImaging 2011: Visualization, Image-Guided Procedures, and Modeling, Lake Buena Vista, FL, USA, 13–15 February 2011; Society of Photo-Optical Instrumentation Engineers SPIE: Bellingham, WA, USA, 2011; Volume 7964, pp. 338–350. [Google Scholar] [CrossRef]
  87. Giacomo, P. The Michelson interferometer. Microchim. Acta 1987, 93, 19–31. [Google Scholar] [CrossRef]
  88. Horlick, G. Introduction to Fourier Transform Spectroscopy. Appl. Spectrosc. 1968, 22, 617–626. [Google Scholar] [CrossRef]
  89. Conen, N.; Luhmann, T. Overview of photogrammetric measurement techniques in minimally invasive surgery using endoscopes. ISPRS—Int. Arch. Photogramm. Remote. Spat. Inf. Sci. 2017, XLII-2/W4, 33–40. [Google Scholar] [CrossRef]
  90. Lucesoli, A.; Criante, L.; Simoni, F.; Di Donato, A.; Rozzi, T.; Farabollini, B.; Bonifazi, F. Quantitative endoscopy by FOLCI-based distance sensor. In Proceedings of the IEEE Sensors, Lecce, Italy, 26–29 October 2008; pp. 870–873. [Google Scholar] [CrossRef]
  91. Moon, S.; Choi, E.S. VCSEL-based swept source for low-cost optical coherence tomography. Biomed. Opt. Express 2017, 8, 1110–1121. [Google Scholar] [CrossRef]
  92. Hariyama, T.; Sandborn, P.A.M.; Watanabe, M.; Wu, M.C. High-accuracy range-sensing system based on FMCW using low-cost VCSEL. Opt. Express 2018, 26, 9285–9297. [Google Scholar] [CrossRef]
  93. Praveen, S. Efficient Depth Estimation Using Sparse Stereo-Vision with Other Perception Techniques. In Coding Theory; IntechOpen: London, UK, 2019; Chapter 7. [Google Scholar] [CrossRef]
  94. Liu, Y.; Aggarwal, J. 3.12—Local and Global Stereo Methods. In Handbook of Image and Video Processing (Second Edition), 2nd ed.; Communications, Networking and Multimedia; Bovik, A., Ed.; Academic Press: Burlington, NJ, USA, 2005; pp. 297–308. [Google Scholar] [CrossRef]
  95. Lee, S.; Hong, H. Near-real-time stereo matching method using both cross-based support regions in stereo views. Opt. Eng. 2018, 57, 023103. [Google Scholar] [CrossRef]
  96. Hosni, A.; Bleyer, M.; Rhemann, C.; Gelautz, M.; Rother, C. Real-Time local stereo matching using guided image filtering. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo (ICME), Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar] [CrossRef]
  97. Yang, Q.; Engels, C.; Akbarzadeh, A. Near Real-time Stereo for Weakly-Textured Scenes. In Proceedings of the British Machine Vision Conference, Leeds, UK, 1–4 September 2008; BMVA Press: Durham, UK, 2008; pp. 1–10. [Google Scholar] [CrossRef]
  98. Kowalczuk, J.; Psota, E.T.; Perez, L.C. Real-Time Stereo Matching on CUDA Using an Iterative Refinement Method for Adaptive Support-Weight Correspondences. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 94–104. [Google Scholar] [CrossRef]
  99. Riechert, C.; Zilly, F.; Mueller, M.; Kauff, P. Real-time disparity estimation using line-wise hybrid recursive matching and cross-bilateral median up-sampling. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba Science City, Japan, 11–15 November 2012; pp. 3168–3171, ISBN 978-4-9906441-0-9. [Google Scholar]
  100. Zhang, K.; Lu, J.; Lafruit, G.; Lauwereins, R.; Gool, L.V. Real-time accurate stereo with bitwise fast voting on CUDA. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 794–800. [Google Scholar] [CrossRef]
  101. Yang, Q.; Wang, L.; Yang, R.; Wang, S.; Liao, M.; Nistér, D. Real-time Global Stereo Matching Using Hierarchical Belief Propagation. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 4–7 September 2006; Volume 6, pp. 989–998. Available online: https://bmva-archive.org.uk/bmvc/2006/papers/324.pdf (accessed on 12 March 2024).
  102. Totz, J.; Thompson, S.; Stoyanov, D.; Gurusamy, K.; Davidson, B.; Hawkes, D.; Clarkson, M. Fast Semi-dense Surface Reconstruction from Stereoscopic Video in Laparoscopic Surgery. In Proceedings of the IPCAI, Fukuoka, Japan, 28 June 2014; pp. 206–215. [Google Scholar] [CrossRef]
  103. Roehl, S.; Bodenstedt, S.; Suwelack, S.; Kenngott, H.; Mueller-Stich, B.P.; Dillmann, R.; Speidel, S. Dense GPU-enhanced surface reconstruction from stereo endoscopic images for intraoperative registration. Med Phys. 2012, 39, 1632–1645. [Google Scholar] [CrossRef] [PubMed]
  104. Singh, T.; Alsadoon, A.; Prasad, P.; Alsadoon, O.H.; Venkata, H.S.; Alrubaie, A. A novel enhanced hybrid recursive algorithm: Image processing based augmented reality for gallbladder and uterus visualisation. Egypt. Inform. J. 2020, 21, 105–118. [Google Scholar] [CrossRef]
  105. Liu, Y.; Yang, J.; Chu, R. Objective Evaluation Criteria for Shooting Quality of Stereo Cameras over Short Distance. Radioengineering 2015, 24, 305–313. [Google Scholar] [CrossRef]
  106. Patoommakesorn, K.; Vignat, F.; Villeneuve, F. The 3D Edge Reconstruction from 2D Image by Using Correlation Based Algorithm. In Proceedings of the 2019 IEEE 6th International Conference on Industrial Engineering and Applications (ICIEA), Tokyo, Japan, 12–15 April 2019; pp. 372–376. [Google Scholar] [CrossRef]
  107. Lin Chen, F.R.; Heipke, C. Feature detection and description for image matching: From hand-crafted design to deep learning. Geo-Spat. Inf. Sci. 2021, 24, 58–74. [Google Scholar] [CrossRef]
  108. Zhang, K.; Lu, J.; Lafruit, G. Cross-Based Local Stereo Matching Using Orthogonal Integral Images. IEEE Trans. Circuits Syst. Video Technol. 2009, 19, 1073–1079. [Google Scholar] [CrossRef]
  109. Yang, Q.; Wang, L.; Yang, R.; Stewénius, H.; Nistér, D. Stereo Matching with Color-Weighted Correlation, Hierarchical Belief Propagation, and Occlusion Handling. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 492–504. [Google Scholar] [CrossRef]
  110. Felzenszwalb, P.; Huttenlocher, D. Efficient belief propagation for early vision. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; Volume 1, p. I. [Google Scholar] [CrossRef]
  111. Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  112. Kauff, P.; Brandenburg, N.; Karl, M.; Schreer, O. Fast Hybrid Block- And Pixel-Recursive Disparity Analysis For Real-Time Applications In Immersive Tele-Conference Scenarios. In Proceedings of the 9-th International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision, Plzen, Czech Republic, 5–9 February 2001; pp. 198–205. [Google Scholar]
  113. Waizenegger, W.; Feldmann, I.; Schreer, O.; Kauff, P.; Eisert, P. Real-time 3D body reconstruction for immersive TV. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 360–364. [Google Scholar] [CrossRef]
  114. Zuiderveld, K. Contrast Limited Adaptive Histogram Equalization; Academic Press Professional, Inc.: Cambridge, MA, USA, 1994; pp. 474–485. ISBN 0-12-336156-7. [Google Scholar] [CrossRef]
  115. Yoon, K.J.; Kweon, I.S. Locally adaptive support-weight approach for visual correspondence search. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 924–931. [Google Scholar] [CrossRef]
  116. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  117. Choudhary, B.K.; Kumar, N.; And, S.; Shanker, P. Pyramid method in image processing. J. Inf. Syst. Commun. 2012, 3, 269. [Google Scholar]
  118. Song, J.; Zhu, Q.; Lin, J.; Ghaffari, M. BDIS: Bayesian Dense Inverse Searching Method for Real-Time Stereo Surgical Image Matching. IEEE Trans. Robot. 2022, 39, 1388–1406. [Google Scholar] [CrossRef]
  119. Mao, Y.; Liu, Z.; Li, W.; Dai, Y.; Wang, Q.; Kim, Y.T.; Lee, H.S. UASNet: Uncertainty Adaptive Sampling Network for Deep Stereo Matching. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 6311–6319. [Google Scholar] [CrossRef]
  120. Xu, G.; Cheng, J.; Guo, P.; Yang, X. ACVNet: Attention Concatenation Volume for Accurate and Efficient Stereo Matching. arXiv 2022, arXiv:2203.02146. [Google Scholar]
  121. Cheng, X.; Zhong, Y.; Harandi, M.; Dai, Y.; Chang, X.; Drummond, T.; Li, H.; Ge, Z. Hierarchical Neural Architecture Search for Deep Stereo Matching. arXiv 2020, arXiv:2010.13501. [Google Scholar]
  122. Zhu, X.; Guo, K.; Fang, H.; Chen, L.; Ren, S.; Hu, B. Cross View Capture for Stereo Image Super-Resolution. IEEE Trans. Multimed. 2021, 24, 3074–3086. [Google Scholar] [CrossRef]
  123. Tankovich, V.; Häne, C.; Fanello, S.R.; Zhang, Y.; Izadi, S.; Bouaziz, S. HITNet: Hierarchical Iterative Tile Refinement Network for Real-time Stereo Matching. arXiv 2020, arXiv:2007.12140. [Google Scholar]
  124. Yang, G.; Manela, J.; Happold, M.; Ramanan, D. Hierarchical Deep Stereo Matching on High-Resolution Images. arXiv 2019, arXiv:1912.06704. [Google Scholar] [CrossRef]
  125. Duggal, S.; Wang, S.; Ma, W.C.; Hu, R.; Urtasun, R. DeepPruner: Learning Efficient Stereo Matching via Differentiable PatchMatch. arXiv 2019, arXiv:1909.05845. [Google Scholar]
  126. Mayer, N.; Ilg, E.; Hausser, P.; Fischer, P.; Cremers, D.; Dosovitskiy, A.; Brox, T. A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4040–4048. [Google Scholar] [CrossRef]
  127. Tonioni, A.; Tosi, F.; Poggi, M.; Mattoccia, S.; Di Stefano, L. Real-Time self-adaptive deep stereo. arXiv 2019, arXiv:1810.05424. [Google Scholar]
  128. Khamis, S.; Fanello, S.R.; Rhemann, C.; Kowdle, A.; Valentin, J.P.C.; Izadi, S. StereoNet: Guided Hierarchical Refinement for Real-Time Edge-Aware Depth Prediction. arXiv 2018, arXiv:1807.08865. [Google Scholar]
  129. Garg, R.; G, V.K.B.; Reid, I.D. Unsupervised CNN for Single View Depth Estimation: Geometry to the Rescue. arXiv 2016, arXiv:1603.04992. [Google Scholar]
  130. Zhou, K.; Meng, X.; Cheng, B. Review of Stereo Matching Algorithms Based on Deep Learning. Comput. Intell. Neurosci. 2020, 2020, 8562323. [Google Scholar] [CrossRef] [PubMed]
  131. Žbontar, J.; LeCun, Y. Computing the stereo matching cost with a convolutional neural network. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1592–1599. [Google Scholar] [CrossRef]
  132. Huo, J.; Zhou, C.; Yuan, B.; Yang, Q.; Wang, L. Real-Time Dense Reconstruction with Binocular Endoscopy Based on StereoNet and ORB-SLAM. Sensors 2023, 23, 2074. [Google Scholar] [CrossRef] [PubMed]
  133. Mayer, N.; Ilg, E.; Fischer, P.; Hazirbas, C.; Cremers, D.; Dosovitskiy, A.; Brox, T. What Makes Good Synthetic Training Data for Learning Disparity and Optical Flow Estimation? arXiv 2018, arXiv:1801.06397. [Google Scholar] [CrossRef]
  134. Chen, Z.; Marzullo, A.; Alberti, D.; Lievore, E.; Fontana, M.; De Cobelli, O.; Musi, G.; Ferrigno, G.; De Momi, E. FRSR: Framework for real-time scene reconstruction in robot-assisted minimally invasive surgery. Comput. Biol. Med. 2023, 163, 107121. [Google Scholar] [CrossRef]
  135. Bardozzo, F.; Collins, T.; Forgione, A.; Hostettler, A.; Tagliaferri, R. StaSiS-Net: A stacked and siamese disparity estimation network for depth reconstruction in modern 3D laparoscopy. Med. Image Anal. 2022, 77, 102380. [Google Scholar] [CrossRef] [PubMed]
  136. Benjamens, S.; Dhunnoo, P.; Mesko, B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: An online database. NPJ Digit. Med. 2020, 3, 118. [Google Scholar] [CrossRef]
  137. Food and Drug Administration. Proposed Regulatory Framework for Modifications to AI/ML-Based Software as a Medical Device. 2019. Available online: https://www.regulations.gov/document/FDA-2019-N-1185-0001 (accessed on 9 October 2023).
  138. Heinrichs, M.; Rodehorst, V.; Hellwich, O. Efficient Semi-Global Matching for Trinocular Stereo. In Proceedings of the PIA 2007—Photogrammetric Image Analysis, Munich, Germany, 19–21 September 2007. [Google Scholar]
  139. Conen, N.; Luhmann, T.; Maas, H.G. Development and Evaluation of a Miniature Trinocular Camera System for Surgical Measurement Applications. PFG-Photogramm. Remote Sens. Geoinf. Sci. 2017, 85, 127–138. [Google Scholar] [CrossRef]
  140. Hirschmüller, H. Stereo Processing by Semi-Global Matching and Mutual Information. IEEE Trans. Pattern Anal. Mach. 2008, 30, 328–341. [Google Scholar] [CrossRef]
  141. Hassanfiroozi, A.; Huang, Y.P.; Javidi, B.; Shieh, H.P. Hexagonal liquid crystal lens array for 3D endoscopy. Opt. Express 2015, 23, 971–981. [Google Scholar] [CrossRef]
  142. Hassanfiroozi, A. Dual layer electrode liquid crystal lens for 2D/3D tunable endoscopy imaging system. Opt. Express 2016, 24, 8527–8538. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Exemplary depiction of a laparoscopic MIS, also referred to as keyhole surgery, using an endoscope.
Figure 1. Exemplary depiction of a laparoscopic MIS, also referred to as keyhole surgery, using an endoscope.
Jimaging 10 00120 g001
Figure 2. Overview of the state-of-the-art real-time camera-based acquisition systems for 3D reconstruction that are discussed in this contribution.
Figure 2. Overview of the state-of-the-art real-time camera-based acquisition systems for 3D reconstruction that are discussed in this contribution.
Jimaging 10 00120 g002
Figure 3. Illustration of accuracy vs. precision, where the bullseye represents the true value that is expected, while black dots represent measurements, hence throwing results.
Figure 3. Illustration of accuracy vs. precision, where the bullseye represents the true value that is expected, while black dots represent measurements, hence throwing results.
Jimaging 10 00120 g003
Figure 4. Visual representation of commonly used error metrics when comparing a method’s performance against ground truth data of a dataset. Where DEV is the deviation to the ground truth, and MAE and RMSE as defined in Section 2.2.
Figure 4. Visual representation of commonly used error metrics when comparing a method’s performance against ground truth data of a dataset. Where DEV is the deviation to the ground truth, and MAE and RMSE as defined in Section 2.2.
Jimaging 10 00120 g004
Figure 5. Historic timeline of datasets for 3D reconstruction.
Figure 5. Historic timeline of datasets for 3D reconstruction.
Jimaging 10 00120 g005
Figure 6. Example image from the SCARED dataset, along with the corresponding depth map [15].
Figure 6. Example image from the SCARED dataset, along with the corresponding depth map [15].
Jimaging 10 00120 g006
Figure 7. 3D reconstruction of an in vivo video sequence from a monocular laparoscope, using the quasi-conformal method as presented by Malti et al. [64].
Figure 7. 3D reconstruction of an in vivo video sequence from a monocular laparoscope, using the quasi-conformal method as presented by Malti et al. [64].
Jimaging 10 00120 g007
Figure 8. Intra-operative example of a measurement during the 3D reconstruction of an in vivo video sequence. The side-view of the intra-operative measurement example on the right shows that the measurement line closely follows the surface curvature. Reprinted with permission from Chen et al. [52]. Copyright 2024 Elsevier.
Figure 8. Intra-operative example of a measurement during the 3D reconstruction of an in vivo video sequence. The side-view of the intra-operative measurement example on the right shows that the measurement line closely follows the surface curvature. Reprinted with permission from Chen et al. [52]. Copyright 2024 Elsevier.
Jimaging 10 00120 g008
Figure 9. Porcine large bowel under Structured Light (SL). Reprinted with permission from Lin et al. [58]. Copyright 2024 Springer Nature.
Figure 9. Porcine large bowel under Structured Light (SL). Reprinted with permission from Lin et al. [58]. Copyright 2024 Springer Nature.
Jimaging 10 00120 g009
Figure 10. Presentation of the calibration process of the SL method. The depicted lamb trachea was examined in an experiment. The resulting 3D reconstruction is shown on the right. The missing area at the top is caused by the camera connection cable. Reprinted with permission from Schmalz et al. [56]. Copyright 2024 Elsevier.
Figure 10. Presentation of the calibration process of the SL method. The depicted lamb trachea was examined in an experiment. The resulting 3D reconstruction is shown on the right. The missing area at the top is caused by the camera connection cable. Reprinted with permission from Schmalz et al. [56]. Copyright 2024 Elsevier.
Jimaging 10 00120 g010
Figure 11. The ToF method uses the phase shift between the emitted and received light pulses to calculate the distance to an object. To determine the distance the phase shift is multiplied by the speed of light and then divided by four times pi, multiplied by the modulation frequency of the emitted light pulse.
Figure 11. The ToF method uses the phase shift between the emitted and received light pulses to calculate the distance to an object. To determine the distance the phase shift is multiplied by the speed of light and then divided by four times pi, multiplied by the modulation frequency of the emitted light pulse.
Jimaging 10 00120 g011
Figure 12. Schematic diagram of a laser distance sensing system based on the Michelson interferometer. The green path depicts the known reference path and the red path length is dependent on the distance to the sample object. The resulting interference on the detector is recorded, and a Fourier transform from the resulting interferogram leads to a spectrum that correlates to the measured depth.
Figure 12. Schematic diagram of a laser distance sensing system based on the Michelson interferometer. The green path depicts the known reference path and the red path length is dependent on the distance to the sample object. The resulting interference on the detector is recorded, and a Fourier transform from the resulting interferogram leads to a spectrum that correlates to the measured depth.
Jimaging 10 00120 g012
Figure 13. Stereo matching takes advantage of the information provided by two images of the same scene. To determine the 3D position of an object, the corresponding location is identified in both images by either using a correlation-based or a feature-based matching approach. With a known baseline, the resulting disparity can be used to triangulate the 3D position.
Figure 13. Stereo matching takes advantage of the information provided by two images of the same scene. To determine the 3D position of an object, the corresponding location is identified in both images by either using a correlation-based or a feature-based matching approach. With a known baseline, the resulting disparity can be used to triangulate the 3D position.
Jimaging 10 00120 g013
Figure 14. Example of a 3D point cloud from different perspectives generated by a deterministic stereo matching algorithm using images from the SCARED dataset [15] as input.
Figure 14. Example of a 3D point cloud from different perspectives generated by a deterministic stereo matching algorithm using images from the SCARED dataset [15] as input.
Jimaging 10 00120 g014
Figure 15. Asimplified structure of a deep learning network such as a CNN. The blue input layers accept the pixels of the stereo images as input. Colored in green are the hidden layers that perform a combination of convolutional operations on the information passed from the blue input layers. To derive the disparity map, the yellow output layer weights the information received by the previous layer.
Figure 15. Asimplified structure of a deep learning network such as a CNN. The blue input layers accept the pixels of the stereo images as input. Colored in green are the hidden layers that perform a combination of convolutional operations on the information passed from the blue input layers. To derive the disparity map, the yellow output layer weights the information received by the previous layer.
Jimaging 10 00120 g015
Figure 16. The mosaicked 3D point cloud of a pig stomach obtained by SGBM on the left, and by StereoNet presented by Huo et al. [132] on the right. Red rectangles indicate areas with outliers in the point cloud that affect the final stitching results due to a rough surface.
Figure 16. The mosaicked 3D point cloud of a pig stomach obtained by SGBM on the left, and by StereoNet presented by Huo et al. [132] on the right. Red rectangles indicate areas with outliers in the point cloud that affect the final stitching results due to a rough surface.
Jimaging 10 00120 g016
Figure 17. Schematic of a trinocular endoscope observing organ tissue. The dashed lines represent the line of sight for each camera. Stereo matching is performed between all possible camera pairs in order to derive a 3D reconstruction.
Figure 17. Schematic of a trinocular endoscope observing organ tissue. The dashed lines represent the line of sight for each camera. Stereo matching is performed between all possible camera pairs in order to derive a 3D reconstruction.
Jimaging 10 00120 g017
Table 1. Datasets for 3D reconstruction with ground truth information.
Table 1. Datasets for 3D reconstruction with ground truth information.
DatasetStaticDynamicGround Truth (GT) TypeRemarks
EndoSlamXStructured LightMonoscopic image sequence
SimCol3DXCT dataMonoscopic images & videos
2D-3D RegistrationX3D ModelMonoscopic videos
Depth from ColonX3D ModelMonoscopic image sequence
HamlynXpartly availableMono- & Stereoscopic
TsukubaXManual segmentationThe first dataset with GT
MiddleburyXStructured LightStereo images
KittiXLiDARStereoscopic videos
SCAREDXStructured LightStereoscopic videos
SERV-CTXCT dataStereoscopic images
EndoAbsXLaserStereoscopic & Synthetic 3D Models
Phantom CardiacXCT dataStereoscopic videos
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Richter, A.; Steinmann, T.; Rosenthal, J.-C.; Rupitsch, S.J. Advances in Real-Time 3D Reconstruction for Medical Endoscopy. J. Imaging 2024, 10, 120. https://doi.org/10.3390/jimaging10050120

AMA Style

Richter A, Steinmann T, Rosenthal J-C, Rupitsch SJ. Advances in Real-Time 3D Reconstruction for Medical Endoscopy. Journal of Imaging. 2024; 10(5):120. https://doi.org/10.3390/jimaging10050120

Chicago/Turabian Style

Richter, Alexander, Till Steinmann, Jean-Claude Rosenthal, and Stefan J. Rupitsch. 2024. "Advances in Real-Time 3D Reconstruction for Medical Endoscopy" Journal of Imaging 10, no. 5: 120. https://doi.org/10.3390/jimaging10050120

APA Style

Richter, A., Steinmann, T., Rosenthal, J. -C., & Rupitsch, S. J. (2024). Advances in Real-Time 3D Reconstruction for Medical Endoscopy. Journal of Imaging, 10(5), 120. https://doi.org/10.3390/jimaging10050120

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop