Next Article in Journal
AI Model for Detection of Abdominal Hemorrhage Lesions in Abdominal CT Images
Next Article in Special Issue
High-Definition Transcranial Direct Current Stimulation in the Right Ventrolateral Prefrontal Cortex Lengthens Sustained Attention in Virtual Reality
Previous Article in Journal
A Review of Voice-Based Pain Detection in Adults Using Artificial Intelligence
Previous Article in Special Issue
Virtual Reality and Augmented Reality in Plastic and Craniomaxillofacial Surgery: A Scoping Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The Role of Augmented Reality in the Advancement of Minimally Invasive Surgery Procedures: A Scoping Review

by
Phillipp Brockmeyer
1,*,†,
Bernhard Wiechens
2,† and
Henning Schliephake
1
1
Department of Oral and Maxillofacial Surgery, University Medical Center Goettingen, D-37075 Goettingen, Germany
2
Department of Orthodontics, University Medical Center Goettingen, D-37075 Goettingen, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2023, 10(4), 501; https://doi.org/10.3390/bioengineering10040501
Submission received: 23 February 2023 / Revised: 13 April 2023 / Accepted: 20 April 2023 / Published: 21 April 2023
(This article belongs to the Special Issue VR/AR Applications in Biomedical Imaging)

Abstract

:
The purpose of this review was to analyze the evidence on the role of augmented reality (AR) in the improvement of minimally invasive surgical (MIS) procedures. A scoping literature search of the PubMed and ScienceDirect databases was performed to identify articles published in the last five years that addressed the direct impact of AR technology on MIS procedures or that addressed an area of education or clinical care that could potentially be used for MIS development. A total of 359 studies were screened and 31 articles were reviewed in depth and categorized into three main groups: Navigation, education and training, and user-environment interfaces. A comparison of studies within the different application groups showed that AR technology can be useful in various disciplines to advance the development of MIS. Although AR-guided navigation systems do not yet offer a precision advantage, benefits include improved ergonomics and visualization, as well as reduced surgical time and blood loss. Benefits can also be seen in improved education and training conditions and improved user-environment interfaces that can indirectly influence MIS procedures. However, there are still technical challenges that need to be addressed to demonstrate added value to patient care and should be evaluated in clinical trials with sufficient patient numbers or even in systematic reviews or meta-analyses.

1. Introduction

1.1. Advantages and Disadvantages of Minimally Invasive Surgery (MIS)

Advances in surgery have the potential to significantly improve patient outcomes and reduce hospital length of stay (LOS) [1]. Healthcare providers can achieve cost savings and improve efficiency in the delivery of healthcare services, and increase the capacity to treat more patients, thereby improving access to care and reducing waiting times [2]. To achieve these goals, it has been emphasized to continue to invest in surgical advances and technologies, such as minimally invasive surgery (MIS) techniques [3] and the development of new, innovative surgical tools [4].
The overall definition of MIS includes surgical procedures that reduce the morbidity of conventional surgical trauma [5]. These techniques have been made possible primarily by the development and adoption of endoscopic systems and the continued miniaturization of imaging systems [5]. Insufflation devices can be used for the controlled inflation of body cavities and the creation of surgical workspaces [5]. In the last decade, the development of robotic-assisted surgery has provided another technological advancement for MIS, which is currently gaining importance [6]. The main advantages of MIS reported in the literature include reduced postoperative pain due to the avoidance of extensive surgical trauma [5]. This may also reduce the number of immobility related morbidities such as postoperative atelectasis or venous thrombosis [7]. In addition, MIS can help to improve the visibility of inaccessible areas, reduce the risk of inflammation, and improve recovery time and cosmesis [5]. Consequently, they can significantly reduce the mean LOS of patients [8,9,10].
One of the major disadvantages of MIS is the extended surgical “learning curve” [11]. with complications occurring early during this period [11]. Watson et al. identified the first 20 procedures as the interval with the highest risk of complications [12]. This finding has led to the widespread use of educational courses, simulators, web-based videos, and mentoring programs in surgical education and training [5]. In addition, there are procedure-specific risks such as insufflation complications, port site metastasis after laparoscopic ablative surgery, and hernia formation and bleeding [5]. Dissections through robotic-assisted minimally invasive procedures generally take longer than traditional open surgical approaches [6], which may be associated with lower operative turnover rates and corresponding economic disadvantages [5]. However, this disadvantage is usually offset by lower costs associated with a shorter LOS [5]. Today, there is a great demand for MIS procedures, and there are some potential opportunities to develop them further.

1.2. The Basics of Augmented Reality (AR) and Its Impact on Healthcare

The development of new technologies has come to the forefront, especially during the Corona pandemic [13]. Virtual reality (VR) and augmented reality (AR), also referred to as mixed reality (MR) [14], are among the three-dimensional (3D) technologies that have been used in consumer marketing for many decades and are currently gaining momentum in healthcare [15].
AR is a technology that overlays computer-generated images or videos onto real-world objects and environments [16]. The use of AR has the potential to revolutionize the way surgery is delivered and received [17]. AR has been shown to improve patient outcomes [18], increase the efficiency and accuracy of surgical procedures [19], and enhance medical education and training [20]. AR technologies such as AR glasses [21], AR head-mounted displays [22], and AR smartphones [23] are being developed and deployed in healthcare settings, with some early applications showing promising results [23]. However, AR in healthcare is still in its early stages of development, and there are several technical and regulatory challenges that must be addressed before it can be widely adopted [24].

2. Materials and Methods

To address the question of the role of AR in improving MIS procedures, a scoping literature search was conducted in the PubMed and ScienceDirect databases to identify all relevant studies and reviews available from January 2018 to February 2023. Different areas of application were identified and the potential advantages and disadvantages of AR were evaluated.
This review article was designed and conducted according to the enhancing transparency in reporting the synthesis of qualitative research (ENTREQ) guidelines [25] to reveal and discuss current evidence on the utility of AR in MIS.
The search strategy followed the identification and screening guidelines by the preferred reporting items for systematic reviews and meta-analyses (PRISMA) statements [26]. The following Mesh search headings and keywords were used: “Augmented Reality”, “AR”, “surgery”, “minimal invasive surgery”, and “MIS”. These terms were used in different boolean combinations. We retrieved all eligible studies and evaluated the reference lists of the identified studies and reviews (Figure 1).
Articles were screened for relevance by two authors (PB and BW) based on title and abstract. In case of uncertainty, full-text screening was performed. A total of 31 articles were independently reviewed by both authors. All articles with questionable relevance were reviewed independently by both authors to reach a consensus. In case of disagreement, a third author (HS) was involved.
Articles were included if they were considered relevant to the advancement of MIS through AR technologies if they were relevant to the practice of MIS (i.e., procedures that fall within the scope of MIS), or its subspecialties if the method used to study AR could be applied to MIS (i.e., use of an AR headset or novel features) and if they had potential educational or didactic value for MIS.
Articles were excluded if they were not directly relevant to MIS techniques, did not address AR, or were not published in English. A total of 31 studies relevant to both AR and MIS promotion were identified. Articles were accessed through the University of Goettingen library system.
Due to the small number of randomized clinical trials, a risk of bias analysis was not required and was not performed.

3. Results

A total of 31 articles from the last five years (2018–2023) were included in the review. The role of AR in the improvement of MIS procedures was identified in three main areas: AR-guided navigation, improving education and training conditions, and building improved user-environment interfaces (Table 1). Some articles were found to be related to more than one primary theme.

3.1. AR-Guided Navigation

Of the 31 papers included a total of 16 articles addressed the topic of AR-based navigation. Six studies were experimental or proof-of-concept approaches. One study was a prospective multicenter clinical approach, one study was a retrospective analysis of clinical data, one study was a prospective case-controlled study, one study was a case report, one study was a clinical investigation, and five studies were review articles. Of these, two articles were systematic reviews. Existing studies and their reported outcome data showed a high degree of heterogeneity (Table 1), which inhibited a comprehensive comparison based on quantitative data.
In an experimental approach, Zadeh et al. analyzed the impact of artificial intelligence (AI) concepts using the Uteraug system, an AR-based guidance application for laparoscopic uterine surgery [28]. A dataset of 3800 images from 79 laparoscopic videos was created, including various annotations: uterine segmentation, uterine contours, and left and right tubal junction regions. The dataset was divided into a training and a test dataset, and a neural network was trained for automatic surgical image interpretation. The performance of the neural network on the training dataset was compared with that of experts on the test dataset. A performance plateau was demonstrated at 700 images for uterine segmentation and 2000 images for uterine contouring. The segmentation results for the training and test datasets were 94.6% and 84.9% (the higher the better), and the final contouring error was 19.5% and 47.3% (the lower the better). The authors concluded that this system could automatically interpret surgical images and provide AR performance equivalent to the current manual system by providing additional image datasets [28].
In their review, Xu et al. analyzed current surgical tracking systems, one of the most important technologies for MIS navigation, and evaluated the advantages and disadvantages of different solutions [29]. The problem of information loss is an inherent drawback of monomodal surgical navigation systems. It is characterized by physical limitations, attenuation, signal dropouts, and unstable accuracy of tracking algorithms. The authors conclude that future research trends, such as the integration of AR for visualization during surgical tracking, will tend to ameliorate this problem [29].
Butler and colleagues described the first in vivo use of AR-guided percutaneous MIS placement of pedicle screws in the setting of spine surgery [30]. A total of 164 MIS cases were performed on patients from June 2020 to March 2022 by three senior surgeons at two institutions. AR was used for percutaneous pedicle screw instrument placement with spinal navigation. A total of 606 pedicle screws were placed. The AR system used consisted of a wireless headset with a transparent near-eye display that projected intraoperative 3D images directly onto the surgeon’s retina. The intraoperative CT data were processed in the headset and integrated into the surgeon’s field of view, creating a “see-through” 3D effect in addition to the standard 2D navigation images. MIS pedicle screw placement was performed percutaneously through a single line of sight with navigated instruments. The average time from registration to percutaneous access to final screw placement was 3 min and 54 s per screw. Learning curve analysis showed that operative times were similar in early cases compared to those performed with more experience with the system. The authors noted the efficiency and safety of AR-guided screw placement compared to conventional technologies [30].
Zhu et al. proposed an AR-based neuroendoscopic navigation system to assist surgeons in the localization and removal of intracerebral hematomas [31]. Dual-mode AR navigation was proposed to provide comprehensive guidance from catheter implantation to hematoma removal. The authors developed a series of experiments to validate the accuracy and validity of this system. The average mean square error of registration between medical images and patients was 0.784 mm, and the variance was 0.1426 mm. The degree of pixel mismatch was less than 1% for the different AR modes. For the catheter implantation experiments, the mean distance error was 1.28 mm and the variance was 0.43 mm, while the mean angular error was 1.34° and the variance was 0.45°. Comparison experiments were performed to evaluate the practicality of the system. The authors concluded that this AR guidance system can provide stereo images with depth information fused to the patient to help surgeons locate targets and remove hematomas [31].
Lecointre and colleagues developed and tested an AR-based robotic assistance system that performed a real-time multimodal and temporal fusion of laparoscopic images with preoperative medical images in a porcine model [32]. It enabled targeted in vivo lymph node detection during minimally invasive pelvic lymphadenectomy. A measurement campaign was performed to determine the most accurate tracking system (UR5 cobot versus NDI Polaris). Subsequent procedures on two pigs consisted of identifying artificial target lymph nodes and anatomical landmarks without and with AR support. The AR overlay on the target structures was quantitatively evaluated. Clinical relevance was assessed by a questionnaire completed by experienced and trainee surgeons. The accuracy of the CT overlay was greater than 90% and the overflow rate was less than 6%. The authors describe the system as reliable, safe, and accurate, and plan to validate its use in a clinical trial [32].
Guo et al. evaluated the feasibility, accuracy, and efficacy of AR technology and 3D plate library assisted posterior MIS for scapular fractures [33]. They retrospectively evaluated the records of 21 patients with scapular fractures treated with posterior MIS with reconstruction plates: 9 patients were treated with conventional fixation, while 12 patients were treated with preoperative virtual simulation and intraoperative navigation-assisted fixation with the AR system. Operative time, blood loss, complications, and Hardegger function were compared between the two groups. It was found that the patients who used the AR system had significantly shorter operation time (−28.75 min; p = 0.0007) and lower blood loss (−81.94 mL, p = 0.0052) than the patients in the conventional surgery group. The authors concluded that AR and 3D plate library-assisted posterior MIS is an effective and reliable method for the treatment of scapular fractures, allowing precise preoperative planning and intraoperative navigation. It also saves time and allows for a more individualized treatment plan, resulting in a safer reduction procedure [33].
Felix et al. placed a total of 124 thoracolumbar pedicle screws under AR guidance in seven cadavers [34]. Sixty-five screws were placed in four donors using open spine surgery. Fifty-nine screws were placed in three donors using an MIS approach. AR was used exclusively for pedicle screw navigation in both open and minimally invasive spine surgery. A total of 124 pedicle screws were placed using AR navigation with 96% accuracy (Gertzbein-Robbins Class A and B). The combined angular error was 2.4° and the distance error was 1.9 mm. The authors concluded that AR is a highly accurate, emerging technology for navigation in open and minimally invasive spine surgery using commercially available headset hardware [34].
In their review, Yuk et al. present recent advances and innovations in the use of simulation methods in spine surgery [35]. These include VR, MR, and AR. While VR and MR are primarily used for teaching and surgical preparation, the authors describe an advantage of AR technology primarily in practical neurosurgical spine surgery situations [35].
Chen and colleagues proposed a novel in situ AR navigation system with enhanced arthroscopic information for knee surgery [36]. In an experimental approach, intraoperative anatomical positions were first determined using arthroscopic images and arthroscopic calibration, and then a tissue property-based model deformation method was proposed to update the preoperative 3D knee model with anatomical position information. The updated model was then rendered using a glasses-free 3D display to obtain a global AR view of the surgical field. In addition, virtual arthroscopic images were generated from the updated preoperative model to provide anatomical information about the surgical site. The experimental results showed that the virtual arthroscopic images could reproduce the correct structural information with a mean error of 0.32 mm. Compared to 2D arthroscopic navigation, AR navigation reduced target errors by 2.10 mm in the knee phantom and by 2.70 mm in the in vitro porcine knee experiment. The authors concluded that AR navigation is useful for minimally invasive knee surgery because it can provide global in situ information and detailed anatomical information [36].
In their systematic review, Benmahdjoub et al. examined AR technology and its utility in craniomaxillofacial surgery [37]. From a total of 7067 articles identified by AR and surgical keywords, 39 articles were selected. Based on these articles, a classification of study types, surgery types, devices used, metrics reported, and benefits was performed. The results suggest that AR technology could provide a number of benefits by overcoming the challenges of traditional navigation systems, such as hand-eye coordination and depth perception. However, the authors also point out that it is difficult to summarize the metrics reported in the articles, to obtain statistical values from the current studies, and that user evaluation studies are not yet available [37].
In their review, Hussain and colleagues evaluate the benefits and challenges of AR systems in skull base surgery [38]. The authors suggest that navigation systems incorporating AR offer comparable results to conventional navigation systems in terms of precision but with improved ergonomics and visualization. However, more needs to be conducted to improve the current state, achieve maximum safety and reliability, and reduce system costs [38].
Hussain and colleagues reviewed relevant features of AR navigation in MIS and examined its evolution over time [39]. They discussed key features relevant to surgical advancement, including technique and technology development, accuracy, overall healthcare costs, operating room time savings, and radiation exposure. They conclude that AR technology will make surgery safer and more efficient, and that fluoroscopy may be completely replaced by image guidance [39].
Hu et al. evaluated the clinical application of the augmented reality computer-assisted spine surgery (ARCASS) system for percutaneous vertebroplasty (PVP) in their prospective case-control study [40]. A total of 18 patients undergoing PVP with the ARCASS system were included. The control group consisted of age- and sex-matched patients who underwent standard PVP and met the same selection criteria as the case group. Compared to the control group, the ARCASS group required significantly fewer fluoroscopies (6 vs. 18, p < 0.001) and had a shorter operative time (78 vs. 205 s, p < 0.001) during entry point identification and local anesthesia, which began with the registration of the skin entry point at the lesion site and ended with identification of the bony entry point. In terms of accuracy, the ARCASS group had a significantly higher percentage of “good” entry points in the lateral view (81.8% vs. 30.0%, p = 0.028) and in the anteroposterior view (72.7% vs. 20.0%, p = 0.020) than the control group. The authors concluded that the AR system is clinically feasible for PVP and provides a more precise bone entry point with reduced operative time and unnecessary radiation exposure [40].
Gribaudo et al. presented an approach for developing real-time AR solutions for navigation in robotic surgery [41]. A modular approach was developed to solve the tracking problem in in vivo robotic surgery. The authors point out that by dividing the entire surgical procedure into a series of phases, it is possible to assign the best tracking strategy to each phase and to reuse implemented software mechanisms in phases with similar characteristics [41].
Chauvet and colleagues reported the first two clinical cases of real-time AR technology in a laparoscopic myomectomy, in which uterine muscle fibers were visualized with diffusion tensor imaging (DTI) after MRI tractography to help surgeons decide whether to proceed with incision [42]. The authors note that AR and DTI fiber tractography are feasible in a uterus with fibroids. They allow for fiber orientation and help the surgeon visualize and decide on the starting point for laparoscopic myomectomy. Attention to fiber orientation may improve scar quality and reduce postoperative architectural disorganization of the uterus [42].
Brebant et al. presented a simplified method of supermicrosurgical lymphovenous anastomosis (LVA) in 30 patients with secondary upper extremity lymphedema [43]. A surgical microscope with an integrated near-infrared illumination system and an AR imaging system was used to evaluate lymphatic supermicrosurgery. The authors note that AR is minimally invasive, highly effective, and has a very low complication rate. The practice of AR guidance is limited by surgical/equipment-related factors and its effectiveness is limited by technical limitations [43].

3.2. Improving Education and Training

Of the 31 articles identified, six dealt with education and training. One study was a web-based survey, one study was a randomized controlled trial, one study was a clinical investigation, and three studies were review articles (non-systematic) with insufficiently reported quantitative measures.
In their article, Balla and colleagues investigated the current knowledge and use of AR, VR, and MR technology to improve MIS procedures in surgical training in Italy [44]. A web-based survey was developed. Responses from 217 physicians were analyzed. Participants were surgeons (138, 63.6%) and surgical residents (79, 36.4%). Mean knowledge of the role of VR, AR, and MR in surgery was 4.9 ± 2.4 (range 1–10). Most participants (122, 56.2%) had no prior experience with the proposed technologies. The authors conclude that the level of knowledge and diffusion of these new technologies is still limited in Italy. Further studies are needed to determine the benefits of AR, VR and MR in surgical training [44].
Wild et al. described the iSurgeon system designed for visual guidance in the operating room via telemetry with AR [45]. Novice laparoscopic surgeons (n = 60) were randomized into two groups in a crossover design: Group 1 trained with verbal guidance only and then with additional telestration with AR on the operating room screen, and vice versa for Group 2. Training consisted of basic laparoscopic training followed by a specially designed training course that included a porcine laparoscopic cholecystectomy (LC). The authors suggest that telestration with AR improves training success and safety in MIS [45].
Gholizadeh and colleagues present a literature review on image guidance in liver surgery, with particular emphasis on information on AR techniques [46]. The results show that the use of AR technology can visualize blood vessels and tumor structures in the liver during surgery and enable precise navigation in complicated surgical procedures. AR has been shown to be safe and effective in both minimally invasive and invasive liver surgery. With recent advances and significant efforts by liver surgeons, AR technologies have the potential to dramatically improve hepatobiliary surgery. However, further clinical studies are needed to evaluate AR as a tool to reduce postoperative morbidity and mortality [46].
In their paper, Godzik et al. described how the disruptive technologies of VR and AR are being used in spine surgery and education [47]. According to the authors, initial experiences with VR and AR technologies demonstrate their applicability and ease of implementation. However, further prospective studies in multi-institutional and industry-academic partnerships are needed to clarify the future of VR and AR in spine surgery education and clinical practice [47].
Benčurik and colleagues described the clinical use of near-infrared (NIR) fluorescein angiography, a technique of AR, in the context of anastomotic leak rate (ALR) in rectal resections [48]. Data analysis of patients after MIS for middle and lower rectal adenocarcinoma with total mesorectal excision (TME) using fluorescence angiography (FA) with indocyanine green (ICG) (100 patients) was compared with a historical control group (100 patients). The authors concluded that the introduction of new procedures and the use of new technologies, such as the use of FA in AR mode for rectal resections with TME for cancer, may lead to a reduction in the incidence of anastomotic leakage [48].
In their review, Pratt et al. describe a number of important new imaging technologies, including AR technologies, that are more or less integrated with transoral robotic surgery (TORS) [49]. The authors conclude that image guidance during TORS procedures is an exciting proposition in terms of registration accuracy because the regions of interest (e.g., base of the tongue, or oropharynx) are typically adjacent to and surrounded by rigid anatomy [49].

3.3. Building Improved User-Environment Interfaces

Of the 31 articles identified, nine addressed the user-environment interface. Of these, seven were experimental or proof-of-concept approaches and two were review articles with no specified search strategy and insufficiently reported quantitative measures.
In their work, Thabit et al. developed an AR-based suture visualization system and evaluated the accuracy and applicability of AR-based navigation for surgical guidance during minimally invasive spring-assisted craniectomy [50]. The authors concluded that the developed AR system has good accuracy (mean distance 2.4 mm) and can be used during surgery. In addition, the system can help in preplanning minimally invasive craniosynostosis surgery to accurately locate the cranial sutures instead of manually palpating them as in the past [50].
Stewart et al. hypothesized that an AR headset that provides a 3D intracorporeal view while pointed at the surgical field could shorten the time and improve the accuracy of robotic bedside tasks [51]. Bedside assistants (a physician assistant, a medical student, three surgical residents, and two attending surgeons) performed validated tasks in a simulated abdominal cavity with a docked surgical robot. Tasks were performed using a bedside monitor with 2D or 3D vision or an optical head-mounted AR device with 2D or 3D vision. The authors concluded that high-resolution 3D AR vision reduced time and improved accuracy for more complex tasks [51].
In their report, Rush et al. describe institutional experiences with the use and implementation of some of the current AR products in spine surgery [52]. Suggested benefits of AR include reduced distraction, reduced surgical line disruption, the ability to perform more minimally invasive procedures, reduced radiation exposure to the surgical team, and improved pedicle screw accuracy [52].
Previous AR applications for robotic minimally invasive surgery have mainly focused on overlaying preoperative 3D images with patient anatomy. The article by Forte and colleagues presents alternative interactive AR tools for robotic surgery [53]. They designed, built, and evaluated four voice-activated features: Displaying live video from the operating room, displaying two-dimensional preoperative images, measuring 3D resections, and warning of invisible instruments. This low-cost system was developed on a da Vinci Si and can be integrated with surgical robots equipped with a stereo camera and viewer. Eight experienced surgeons performed lymphadenectomies in the dry lab and reported that the features improved the procedure. In particular, they appreciated the ability to access the patient’s medical record on demand, measure distances intraoperatively, and interact with the features using voice commands. The authors concluded that these alternative AR features and interaction methods had a positive impact [53].
In their review, Wendler et al. presented the use of novel molecular imaging as one of the pillars of precision surgery [54]. These devices include technologies ranging from artificial intelligence and computer-aided visualization (software) to innovative molecular imaging modalities and surgical navigation (hardware) [54].
Li and colleagues addressed the current challenges of AR and MR in MIS navigation [55]. The surface of soft tissue is smooth and watery with sparse texture, specular reflection, and frequent deformation. As a result, we often obtain sparse feature points that lead to erroneous results using conventional imaging techniques. The authors present an accurate and robust description and matching method for dense feature points in endoscopic videos. Experimental results show that the novel approach can overcome the influence of specular highlights and robustly describe contours from image sequences of soft tissue surfaces. Compared to state-of-the-art feature point description and matching methods, the presented analysis framework shows the key advantages of robustness and accuracy in matching dense point-to-point images even when severe soft tissue deformation occurs, and the authors expect that this new approach has great potential for 2D/3D reconstruction in endoscopy [55].
Jia et al. pointed out that real-time AR for MIS without additional tracking devices is a valuable but challenging task, especially in dynamic surgical environments [56]. Numerous different movements between target organs are caused by respiration, cardiac motion, or surgical tools, and often need to be captured by a moving, manually positioned endoscope. Therefore, the authors proposed a 6DoF motion tracking method that takes advantage of the latest 2D target tracking methods as well as nonlinear pose optimization and tracking loss recovery in SLAM technologies and can be embedded in such an AR system. The results show that the proposed method is more robust and accurate compared to ORB-SLAM2 in the presence of motion deflection or motion blur; however, heavy smoke is still an important factor that reduces tracking accuracy [56].
Wang et al. showed that tracking is a critical step in achieving accurate AR during MIS [57]. In addition to visual tracking in traditional medical AR, visual tracking attracts much attention because of its generality. If the 3D model of the target organ can be obtained from preoperative images and under the assumption of model rigidity, tracking is transformed into a problem of computing the 6-degree-of-freedom pose of the 3D model. The authors present a robust tracking algorithm in an endoscopic AR system that combines the advantages of regional and dense cues in a unified framework. In addition, an appearance matching method and an occlusion processing method are proposed to effectively handle occlusions. Experiments using a synthetic dataset and a simulated surgical environment demonstrate the effectiveness and robustness of the proposed method. This work presents a novel tracking strategy for medical AR applications [57].
Chen and colleagues point out that previous research on the use of AR technology in monocular surgical MIS scenes has mainly focused on superimposing information without addressing proper spatial calibration, which could lead to mislocalization of annotations and labels, as well as inaccurate depth information and tumor measurements [58]. The authors present a novel intraoperative dense surface reconstruction system capable of providing geometry information from monocular MIS videos for geometry-aware AR applications such as position measurements and depth information. The authors conclude that the new framework is robust and accurate when dealing with challenging situations such as rapid endoscopy camera movements in monocular MIS scenes. Both camera tracking and surface reconstruction based on a sparse point cloud are effective and work in real-time. This demonstrates the potential of the new algorithm for accurate AR localization and depth augmentation with geometric cues and correct surface measurements in MIS with monocular endoscopes [58].
Table 1. Categorization of scoping literature search results in relation to the different MIS practice areas.
Table 1. Categorization of scoping literature search results in relation to the different MIS practice areas.
CitationResearch TopicMethodologyOutcome ParameterFindingsConclusions
Navigation
Zadeh et al.
[28]
AI system for laparoscopic AR-guided uterine surgeryExperimental
N = 3800 (images)
SemiquantitativeSegmentation scores:
(The higher, the better):
94.6% (training dataset)
84.9% (test dataset)
Contour error on training:
(The lower, the better):
19.5% (training dataset)
47.3% (test dataset)
System is useful for all surgical steps
Xu et al.
[29]
Advantages and disadvantages of various surgical tracking systemsReview
Specified search strategy
N = 174 (included)
Objective evaluationOverview of surgical navigation systems, tracking technologies, and preoperative planning proceduresInformation loss is a major problem
Butler et al.
[30]
In vivo percutaneously inserted pedicle screws with AR guidanceProspective multicenter
clinical trial
N = 164 (patients)
QuantitativeTime from registration/percutaneous approach to screw placement: 3.54 min/screw
Time per screw placement in first 20 cases: 4.1 min
Time per screw placement in last 20 cases: 3.52 min
(No difference, p = 0.48)
Confirmed efficiency/safety of screw placement with the benefits of AR technology
Zhu et al.
[31]
Dual-mode AR-navigated neuroendoscopy for target localization and hematoma removalProof-of-concept
Experimental
N = 24
QuantitativeRoot Mean Square Error (RMSE)
Between medical images and patients: 0.784 mm
Variance: 0.1426 mm
Pixel mismatching degrees:
<1% in different AR modes
Error of distance in catheter implantation experiments: 1.28 mm
Variance: 0.43 mm
Average error angle: 1.34°
Variance 0.45°
High accuracy and feasibility of the system to provide stereo images with depth information fused to the patient
Lecointre et al.
[32]
AR-based robotic assistance system for laparoscopic detection of target lymph nodes (TLN) in pelvic lymphadenectomyProof-of-concept
Animal study
N = 2 (pigs)
Quantitative and semiquantitativeCT overlay accuracy: >90%
Overflow rates: <6%
Significant higher scores:
TLN: AR score 3.9 ± 0.32 vs. direct vision; DV, 2.1 ± 0.74
(p < 0.001)
Ureter: AR score 3.7 ± 0.48 vs. DV 2.5 ± 0.84 (p = 0.003)
Vessels: AR score 3.4 ± 0.51 vs. DV 1.7 ± 0.67 (p < 0.001)
AR approach with rigid registration is a first step in simplifying complex procedures and improving surgical safety
Guo et al.
[33]
AR-guided MIS approach for scapula fracturesRetrospective clinical trial
N = 21 (patients)
QuantitativeVirtual simulation time:
44.42 ± 15.54 min
Time required for pre-operative plate contouring:
16.08 ± 5.09 min
AR-guided MIS:
Shorter operation time (−28.75 min, p = 0.0007)
Less blood loss (−81.94 mL, p = 0.0052)
Similar follow up outcome (p > 0.05)
Effective and reliable method for treating scapula fractures
Felix et al.
[34]
AR-guided (VisAR) implantation of thoracolumbar pedicle screwsExperimental
N = 7 (cadavers)
Quantitative124 pedicle screws in total
Accuracy: 96% (Gertzbein-Robbins grades A and B)
Combined angle of error: 2.4°
Distance error: 1.9 mm
High-precision, emerging technology for navigating open surgery and MIS techniques with off-the-shelf headset hardware
Yuk et al.
[35]
Advances/Applications of AR in spine surgeryReview
Specified search strategy
N = 41 (included)
Objective evaluationNo randomized controlled trials to date to evaluate accuracy, cost-effectiveness, and patient outcomes. VR training is an effective way to teach traditional/new methods of spine surgeryThe use of VR/AR will increase in spine surgery
Chen et al.
[36]
In-situ AR navigation system with enhanced arthroscopic information for MIS knee surgeryExperimental
N = 2
(knee phantom, swine knee)
QuantitativeMean targeting error
Knee phantom:
Traditional 2D arthroscopy navigation: 4.11 ± 0.80 mm
AR navigation: 2.01 ± 0.65 mm (Significant difference, p < 0.01)
In vitro swine knee:
Traditional 2D arthroscopy navigation: 5.67 ± 0.97 mm
AR navigation: 2.97 ± 0.79 mm (Significant difference, p < 0.01)
Suggested AR navigation is helpful in MIS knee surgeries
Benmahdjoub
et al. [37]
AR in craniomaxillofacial surgerySystematic Review
Specified search strategy
N = 7067 (reviewed)
N = 39 (included)
Objective evaluationClassification of study types, surgery types, equipment used, metrics reported, and benefitsDifficult to aggregate metrics.
Difficult to obtain statistical value.
Lack of user evaluation studies
Hussain et al.
[38]
AR technology in cranial base surgerySystematic review
Specified search strategy
N = 210 (reviewed)
N = 45 (included)
Objective evaluationEvaluate the benefits/challenges/solutions of AR systems in cranial base surgeryGrowing interest in AR systems that can lead to safer and more cost-effective procedures, but issues need to be addressed
Hussain et al.
[39]
Navigation in MIS and its evolution over timeReview
Unspecified
search strategy
N = 54 (included)
Objective evaluationOverview of the characteristics of navigation in MIS over time and key features for surgical advancementNew developments will further enhance the value of 3D navigation in MIS
Hu et al.
[40]
Percutaneous Vertebroplasty (PVP) with the ARCASS AR SystemProspective
case-control study
N = 18 (patients)
QuantitativeARCASS group/control group:
Less frequency of fluoroscopy
(6 vs. 18, p < 0.001)
Shorter operation time
(78 s vs. 205 s, p < 0.001)
Higher proportion of ‘good’ entry point on lateral views
(81.8% vs. 30.0%, p = 0.028)
and anteroposterior views (72.7% vs. 20.0%, p = 0.020)
The ARCASS system provides a more precise bone entry point with less surgical time and unnecessary radiation exposure
Gribaudo et al.
[41]
Development of AR-guided robotic surgeryExperimental
N = not specified
Objective evaluationModular approach to the tracking problem. Segmentation of the entire process into several stagesMay be helpful in surgical implementation
Chauvet et al.
[42]
AR and magnetic resonance diffusion tensor imaging (DTI) for uterine fiber visualization and trackingCase series
N = 2 (patients)
Clinical evaluationLocalization of myomas
Visualization and overlay of uterine muscle fibers
Can help surgeons identify and determine the starting point for laparoscopic myomectomies
Brebant et al.
[43]
AR-guided supermicrosurgical lymphovenous anastomosis (LVA)Clinical trial
N = 32 (patients)
PROMs63 LVAs in total
27 upper extremities
5 lower extremities
Mean operation time:
60–150 min
Patency was confirmed by intraoperative AR-ICG
No postoperative complications
AR-ICG enables a robust validation of LVA
Education and training
Balla et al.
[44]
Knowledge and prevalence of AR in surgical training in ItalyWeb-based survey
N = 217 (participants)
QuantitativeParticipants:
University hospital (41%), general hospital (35%), national health system (6%), general surgery (86%), abdominal surgery (72.8%)
Knowledge of technology:
Mean perceived knowledge (4.9 ± 2.4, out of max. 10), no experience (56.2%), primarily used for training (31.3%), didactic (29%) and intraoperatively (12.4%), Never used before (48.4%)
Interest in technology:
Should be used for teaching, training, and clinical use (80.3%), significant contribution in training (84.3%) and didactic (71.9%)
Limits of technology:
Insufficient knowledge (83.9%) and costs (80.6%)
Knowledge and dissemination still limited
Wild et al.
[45]
AR-telestration for laparoscopic MIS trainingRandomized
controlled trial
N = 60 (participants)
Global Operative Assessment of Laparoscopic Skills (GOALS)
Objective Structured assessment of Technical Skills (OSATS)
Subjective workload (NASA-TLX questionnaire)
QuantitativeFaster training time
(AR vs. verbal guidance)
(1163 ± 275 vs. 1658 ± 375 s, p < 0.001)
Reduced error rates
Better laparoscopic cholecystectomy
(GOALS 21 ± 5 vs. 18 ± 4, p < 0.007 and OSATS 67 ± 11 vs. 61 ± 8, p < 0.015)
Less complications
(13.3% vs. 40%, p < 0.020)
Reduced subjective workload and stress
(33.6 ± 12.0 vs. 30.6 ± 12.9, p < 0.022)
AR-telestration improves training success and MIS-safety
Gholizadeh
et al. [46]
Overview of MIS and conventional liver surgery based on AR trainingReview
Specified search strategy
N = 135 (review)
N = 31 (included)
QuantitativeInconsistency between algorithms used and claimed registration accuracy
(mean 5.38 mm, range 0.93–10.3 mm)
Any AR system (manual, semi-automatic, or automatic) requires human input/knowledge. Methods for determining accuracy are inconsistent. Measurements include pixel-based or spatial 3D registration error. Registration accuracy is difficult to determine. Few patients have undergone AR surgery. AR in soft tissue surgery cannot accurately register the virtual model
Further clinical studies are needed to evaluate AR as a tool to reduce postoperative morbidity and mortality
Godzik et al.
[47]
VR and AR interfaces in spine surgery and educationReview and case report
Unspecified search
strategy
N = 38 (included)
Objective evaluationOverview of potential future applications and demonstration of the feasibility of a VR program for neurosurgical spine training using a case studyVR/AR is easy to implement. Further prospective studies through multi-institutional and industry-academic partnerships are needed to solidify the future of VR/AR in spine surgery education and clinical practice
Benčurik et al.
[48]
New procedures and technologies for total mesorectal excision (TME)Clinical trial
N = 200 (patients)
SemiquantitativeIn fifteen patients (15%), resection was postponed due to inadequate perfusion detected by AR. The incidence of anastomotic leakage was lower in the group with AR than in the group without AR
(9% vs. 19%, p = 0.042)
The use of AR in rectal resections with TME for cancer may lead to a reduction in the incidence of anastomotic leakage
Pratt et al.
[49]
Image guidance and AR in transsoral robotic surgeryLiterature Overview and recent appraisals
N = 10 (included)
Objective evaluationPreoperative imaging guidance
Intraoperative fluorescence imaging. Deformable registration using CBCT imaging. Image guided cochlear implantation
Ability to expand the surgical field with navigational cues and visualization of important anatomical structures
User-environment interface
Thabit et al.
[50]
AR with electromagnetic tracking system for MIS craniosynostosisExperimental
N = 120 (sutures on two skull phantoms)
System Usability Scale (SUS)
QuantitativeDistance of the marked sutures from planning reference:
2.4 ± 1.2 mm
Time per suture: 13 ± 5 s
SUS value: 73
Good accuracy
Helpful in pre-planning MIS craniosynostosis surgery
Stewart et al.
[51]
AR system for bedside surgical assistanceProof-of-concept
N = unspecified
Different bedside tasks with da Vinci Xi surgical system on mock abdominal cavity
SemiquantitativeImproved times for ring path task with better resolution:
lower resolution 23 ± 11 s vs. higher resolution 14 ± 4 s
(p = 0.002)
High-resolution AR reduces time and improves accuracy during more complex laparoscopic procedures
Rush III et al.
[52]
Advantages/disadvantages of AR in spine surgeryReview
Unspecified
search strategy
N = 20 (included)
Objective evaluationDifferent AR systems: Augmedics and Holosurgical ARAI navigation systemAccurate anatomical information with minimal to no radiation exposure
Forte et al.
[53]
Voice-activated system for displaying live video on da Vinci Si surgical robotExperimental
N = 8 (surgeons)
Phantom model
Utility and usability questionnaire
Four voice-controlled AR functions:
Viewing live video
Viewing 2D pre-op images
Measuring 3D distances
Warning about out-of-view instruments
Quantitative and semiquantitativeAverage time for surgeons to become familiar with the technology: 8.47 min
Accuracy of voice commands: 100%
Voice command sensitivity: 89.8%
Support for further exploration
Wendler et al.
[54]
Evaluate new technologies at various stages of the surgical workflowReview
Unspecified search strategy
N = 226 (included)
Objective evaluationArtificial intelligence. Computational visualization. Innovative molecular imaging modalities. Surgical navigationIntegrating molecular imaging could be the key to a new level of precision surgery
Li et al.
[55]
Dense feature point description and matching method in endoscopic videoExperimental
N = 3 (video segments)
QuantitativeTrue Positive Matching Result (TPM): 142.33
False Positive Matching Results (FPM): 10
New approach has great potential for 2D/3D reconstruction in endoscopy
Jia et al.
[56]
6DoF method to improve motion tracking in AR systemsExperimental
N = unspecified
Ex-vivo tissue phantoms (kidney) and clinical datasets
Root Mean Squared Error (RMSE)
QuantitativeRMSE: 2.31 mm (without disctraction)
RMSE: 3.43 mm (middle-level distraction)
RMSE: 3.56 mm (high-level distraction)
Robust and long-term tracking in highly dynamic operating environments
Wang et al.
[57]
Robust tracking algorithm in an endoscopic AR systemExperimental
N = unspecified
Experiments with synthetic and simulation datasets
QuantitativeAverage Contour Distance: 1.2398 pixels
Frame Rates: 38.46 fps
The effectiveness and robustness of the method represents a novel tracking strategy for medical AR
Chen et al.
[58]
Robotic algorithm (SLAM) in monocular surgical MIS scenes for reliable endoscopic camera trackingExperimental
Simulated laparoscopic scene image sequences and clinical data
(N = 877)
Root Mean Square Distance (RMSD)
QuantitativeRMSD: 2.54 mm
Other monocular MIS scene reconstruction method
(RMSD: 7.21 mm)
State-of-the-art stereo reconstruction method
(RMSD: 2.04/2.57 mm)
High accuracy of the developed algorithm

4. Discussion

AR technology is gaining increasing interest in the development of minimally invasive surgery today [19]. This is reflected in the number of research articles published in recent years from a variety of medical disciplines (Table 1). Although many studies have been published on this topic, some of them differ fundamentally in their research questions, data collection methods, and design, making it impossible to compare them in meta-analyses. In addition, most studies to date have been experimental and few have provided significant clinical benefits to patient care.
Most of the developments have been made in spine surgery/orthopedic surgery [30,33,34,35,36,40] and for endoscopic/laparoscopic procedures [28,31,32,33], which have not yet fully exploited the potential of this new technology. AR could help identify and visualize critical anatomical structures such as blood vessels and nerves to reduce the risk of injury or to guide the proper placement of surgical instruments [30,31].
To answer the question of how AR technology can help improve MIS procedures, we have identified three main areas of application. Most research today is focused on intraoperative AR-guided navigation. AR technology is also being used to improve education and surgical training and to develop and improve new user-environment interfaces.
While the studies discussed here indicate some potential of AR technology in the context of improving MIS navigation solutions [28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43], current AR-guided navigation systems have not shown improved precision compared to conventional navigation methods, albeit shown improvements in ergonomics and visualization have been reported [38]. However, what has been clinically demonstrated to date is evidence of shorter operative times and less blood loss, suggesting a gentler surgical approach [33,40]. Most of the AR-guided navigation studies presented here are experimental development approaches, some of which have shown promising results [28,31,32,34,36,41]. However, clinical trials with adequately sized numbers of patients or even systematic reviews demonstrating added value for patient care are still rare [30,33,37,38].
Although AR is predicted to revolutionize surgery, there are several challenges that need to be addressed to ensure widespread adoption [55,56,57,58]. A major issue, especially in soft tissue navigation, is the accuracy and reliability of the AR system, particularly in registering patient data and matching virtual and real views [55,56]. These are prone to errors, especially during intraoperative repositioning. It is moreover important for future research to identify factors for evaluating AR accuracy. For this purpose, Root Mean Square Distance (RMSD) or Root Mean Square Error (RMSE) has been proposed by some authors [31,56,58]. In addition, real-time processing, and visualization of a large amount of data for AR navigation is problematic because it is computationally intense and requires advanced hardware and software infrastructure [30,31,32,33,34,35,36,37,38,39,40,45,50]. Nevertheless, some promising developments in robotics, visualization, positioning, haptics, artificial intelligence, and computer vision have been presented that may help to further advance AR technology for clinical use [55,56,57,58].
Several studies have addressed the benefits of AR technology and its ability to enhance surgical education and training by providing students and professionals with an immersive, interactive, and engaging learning experience [20,44,46,47,51]. However, awareness of this technology among clinically active surgeons remains low [44]. Without requiring expensive equipment or putting patients at risk, trainees can practice and refine surgical procedures using virtual models of patient anatomy [45]. They receive real-time feedback on their performance, allowing them to refine their technique and improve patient outcomes in vivo. In addition, AR can be used to train surgeons in new or rare surgical procedures that are difficult to perform in the real world [46]. The benefits of virtual surgical training could, therefore, indeed impact MIS approaches. However, this influence has yet to be demonstrated via standardized studies with a large number of participants.
The interface between users and different surgical environments can be improved by using AR technology to provide relevant and contextual information in real-time [19,21,22,23,51,53]. Surgeons can visualize patient data and imaging scans in 3D, allowing for more accurate and personalized surgical planning [23]. It can enable remote collaboration and consultation between surgeons, facilitating knowledge sharing and improving patient outcomes [51]. However, to date, most studies on the topic of user interface improvement have been experimental or proof-of-concept approaches. So far, there is no evidence of added value for patient care.

5. Conclusions

AR technology is currently being investigated for MIS approaches in various medical disciplines. Most studies to date have been limited to experimental developmental approaches, while the number of clinical trials and systematic reviews is low. Most articles on AR-guided navigation are limited to endoscopic/laparoscopic MIS approaches. Intraoperative tracking is still considered cumbersome and error prone. In addition, processing large amounts of data in real-time is computationally intensive and requires improved hardware and software solutions. However, AR technology has been shown to improve ergonomics and visualization, as well as reduce operation time and blood loss. AR can also be used to improve education and surgical training and to create and improve new user interfaces that may indirectly improve MIS procedures. However, to date, controlled studies with large case numbers and standardized outcome parameters and reporting are lacking to confirm the added value for clinical use.

Author Contributions

Conceptualization, P.B. and H.S.; methodology, P.B.; software, P.B.; validation, P.B., B.W. and H.S.; formal analysis, P.B.; investigation, P.B.; resources, H.S.; data curation, P.B.; writing—original draft preparation, P.B.; writing—review and editing, P.B., B.W. and H.S.; visualization, P.B.; supervision, H.S.; project administration, P.B.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge support by the Open Access Publication Funds of the University of Goettingen.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Golder, H.J.; Papalois, V. Enhanced Recovery after Surgery: History, Key Advancements and Developments in Transplant Surgery. J. Clin. Med. 2021, 10, 1634. [Google Scholar] [CrossRef] [PubMed]
  2. Baek, H.; Cho, M.; Kim, S.; Hwang, H.; Song, M.; Yoo, S. Analysis of length of hospital stay using electronic health records: A statistical and data mining approach. PLoS ONE 2018, 13, e0195901. [Google Scholar] [CrossRef]
  3. Liu, C.; Pan, L.K. Advances in minimally invasive surgery and clinical measurement. Comput. Assist. Surg. 2019, 24, 1–4. [Google Scholar] [CrossRef]
  4. Hirst, A.; Philippou, Y.; Blazeby, J.; Campbell, B.; Campbell, M.; Feinberg, J.; Rovers, M.; Blencowe, N.; Pennell, C.; Quinn, T.; et al. No Surgical Innovation Without Evaluation: Evolution and Further Development of the IDEAL Framework and Recommendations. Ann. Surg. 2019, 269, 211–220. [Google Scholar] [CrossRef] [PubMed]
  5. Jaffray, B. Minimally invasive surgery. Arch. Dis. Child. 2005, 90, 537–542. [Google Scholar] [CrossRef]
  6. Prete, F.P.; Pezzolla, A.; Prete, F.; Testini, M.; Marzaioli, R.; Patriti, A.; Jimenez-Rodriguez, R.M.; Gurrado, A.; Strippoli, G.F.M. Robotic Versus Laparoscopic Minimally Invasive Surgery for Rectal Cancer: A Systematic Review and Meta-analysis of Randomized Controlled Trials. Ann. Surg. 2018, 267, 1034–1046. [Google Scholar] [CrossRef] [PubMed]
  7. Karayiannakis, A.J.; Makri, G.G.; Mantzioka, A.; Karousos, D.; Karatzas, G. Postoperative pulmonary function after laparoscopic and open cholecystectomy. Br. J. Anaesth. 1996, 77, 448–452. [Google Scholar] [CrossRef] [PubMed]
  8. Zureikat, A.H.; Beane, J.D.; Zenati, M.S.; Al Abbas, A.I.; Boone, B.A.; Moser, A.J.; Bartlett, D.L.; Hogg, M.E.; Zeh, H.J., 3rd. 500 Minimally Invasive Robotic Pancreatoduodenectomies: One Decade of Optimizing Performance. Ann. Surg. 2021, 273, 966–972. [Google Scholar] [CrossRef]
  9. de Rooij, T.; Lu, M.Z.; Steen, M.W.; Gerhards, M.F.; Dijkgraaf, M.G.; Busch, O.R.; Lips, D.J.; Festen, S.; Besselink, M.G. Minimally Invasive Versus Open Pancreatoduodenectomy: Systematic Review and Meta-analysis of Comparative Cohort and Registry Studies. Ann. Surg. 2016, 264, 257–267. [Google Scholar] [CrossRef]
  10. Chadi, S.A.; Guidolin, K.; Caycedo-Marulanda, A.; Sharkawy, A.; Spinelli, A.; Quereshy, F.A.; Okrainec, A. Current Evidence for Minimally Invasive Surgery During the COVID-19 Pandemic and Risk Mitigation Strategies: A Narrative Review. Ann. Surg. 2020, 272, e118–e124. [Google Scholar] [CrossRef]
  11. Chen, M.K.; Schropp, K.P.; Lobe, T.E. Complications of minimal-access surgery in children. J. Pediatr. Surg. 1996, 31, 1161–1165. [Google Scholar] [CrossRef] [PubMed]
  12. Watson, D.I.; Baigrie, R.J.; Jamieson, G.G. A learning curve for laparoscopic fundoplication. Definable, avoidable, or a waste of time? Ann. Surg. 1996, 224, 198–203. [Google Scholar] [CrossRef] [PubMed]
  13. Chandra, M.; Kumar, K.; Thakur, P.; Chattopadhyaya, S.; Alam, F.; Kumar, S. Digital technologies, healthcare and COVID-19: Insights from developing and emerging nations. Health Technol. 2022, 12, 547–568. [Google Scholar] [CrossRef]
  14. Milgram, P.; Kishino, F. A Taxonomy of Mixed Reality Visual Displays. IEICE Trans. Inf. Syst. 1994, E77-D, 1321–1329. [Google Scholar]
  15. Wedel, M.; Bigné, E.; Zhang, J. Virtual and augmented reality: Advancing research in consumer marketing. Int. J. Res. Mark. 2020, 37, 443–465. [Google Scholar] [CrossRef]
  16. Chen, Y.; Wang, Q.; Chen, H.; Song, X.; Tang, H.; Tian, M. An overview of augmented reality technology. J. Phys. Conf. Ser. 2019, 1237, 022082. [Google Scholar] [CrossRef]
  17. Munzer, B.W.; Khan, M.M.; Shipman, B.; Mahajan, P. Augmented reality in emergency medicine: A scoping review. J. Med. Internet Res. 2019, 21, e12368. [Google Scholar] [CrossRef] [PubMed]
  18. Naziri, Q.; Mixa, P.J.; Murray, D.P.; Abraham, R.; Zikria, B.A.; Sastry, A.; Patel, P.D. Robotic-Assisted and Computer-Navigated Unicompartmental Knee Arthroplasties: A Systematic Review. Surg. Technol. Int. 2018, 32, 271–278. [Google Scholar] [PubMed]
  19. Vavra, P.; Roman, J.; Zonca, P.; Ihnat, P.; Nemec, M.; Kumar, J.; Habib, N.; El-Gendi, A. Recent Development of Augmented Reality in Surgery: A Review. J. Healthc. Eng. 2017, 2017, 4574172. [Google Scholar] [CrossRef]
  20. Herron, J. Augmented reality in medical education and training. J. Electron. Resour. Med. Libr. 2016, 13, 51–55. [Google Scholar] [CrossRef]
  21. Basoglu, N.A.; Goken, M.; Dabic, M.; Ozdemir Gungor, D.; Daim, T.U. Exploring adoption of augmented reality smart glasses: Applications in the medical industry. Front. Eng. 2018, 5, 167–181. [Google Scholar] [CrossRef]
  22. Gallos, P.; Georgiadis, C.; Liaskos, J.; Mantas, J. Augmented reality glasses and head-mounted display devices in healthcare. In Data, Informatics and Technology: An Inspiration for Improved Healthcare; IOS Press: Amsterdam, The Netherlands, 2018; pp. 82–85. [Google Scholar]
  23. Moro, C.; Štromberga, Z.; Raikos, A.; Stirling, A. The effectiveness of virtual and augmented reality in health sciences and medical anatomy. Anat. Sci. Educ. 2017, 10, 549–559. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, L.; Day, T.W.; Tang, W.; John, N.W. Recent developments and future challenges in medical mixed reality. In Proceedings of the 2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Nantes, France, 9–13 October 2017; pp. 123–135. [Google Scholar]
  25. Tong, A.; Flemming, K.; McInnes, E.; Oliver, S.; Craig, J. Enhancing transparency in reporting the synthesis of qualitative research: ENTREQ. BMC Med. Res. Methodol. 2012, 12, 181. [Google Scholar] [CrossRef] [PubMed]
  26. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. J. Clin. Epidemiol. 2021, 134, 178–189. [Google Scholar] [CrossRef]
  27. Haddaway, N.R.; Page, M.J.; Pritchard, C.C.; McGuinness, L.A. PRISMA2020: An R package and Shiny app for producing PRISMA 2020-compliant flow diagrams, with interactivity for optimised digital transparency and Open Synthesis. Campbell Syst. Rev. 2022, 18, e1230. [Google Scholar] [CrossRef] [PubMed]
  28. Zadeh, S.M.; François, T.; Comptour, A.; Canis, M.; Bourdel, N.; Bartoli, A. SurgAI3.8K: A labelled dataset of gynaecologic organs in laparoscopy, with application to automatic augmented reality surgical guidance. J. Minim. Invasive Gynecol. 2023. [Google Scholar] [CrossRef]
  29. Xu, L.; Zhang, H.; Wang, J.; Li, A.; Song, S.; Ren, H.; Qi, L.; Gu, J.J.; Meng, M.Q.H. Information loss challenges in surgical navigation systems: From information fusion to AI-based approaches. Inf. Fusion 2023, 92, 13–36. [Google Scholar] [CrossRef]
  30. Butler, A.J.; Colman, M.W.; Lynch, J.; Phillips, F.M. Augmented reality in minimally invasive spine surgery: Early efficiency and complications of percutaneous pedicle screw instrumentation. Spine J. 2023, 23, 27–33. [Google Scholar] [CrossRef]
  31. Zhu, T.; Jiang, S.; Yang, Z.; Zhou, Z.; Li, Y.; Ma, S.; Zhuo, J. A neuroendoscopic navigation system based on dual-mode augmented reality for minimally invasive surgical treatment of hypertensive intracerebral hemorrhage. Comput. Biol. Med. 2022, 140, 105091. [Google Scholar] [CrossRef]
  32. Lecointre, L.; Verde, J.; Goffin, L.; Venkatasamy, A.; Seeliger, B.; Lodi, M.; Swanström, L.L.; Akladios, C.; Gallix, B. Robotically assisted augmented reality system for identification of targeted lymph nodes in laparoscopic gynecological surgery: A first step toward the identification of sentinel node: Augmented reality in gynecological surgery. Surg. Endosc. 2022, 36, 9224–9233. [Google Scholar] [CrossRef]
  33. Guo, Q.; Li, X.; Tang, Y.; Huang, Y.; Luo, L. Augmented reality and three-dimensional plate library-assisted posterior minimally invasive surgery for scapula fracture. Int. Orthop. 2022, 46, 875–882. [Google Scholar] [CrossRef] [PubMed]
  34. Felix, B.; Kalatar, S.B.; Moatz, B.; Hofstetter, C.; Karsy, M.; Parr, R.; Gibby, W. Augmented Reality Spine Surgery Navigation: Increasing Pedicle Screw Insertion Accuracy for Both Open and Minimally Invasive Spine Surgeries. Spine 2022, 47, 865–872. [Google Scholar] [CrossRef] [PubMed]
  35. Yuk, F.J.; Maragkos, G.A.; Sato, K.; Steinberger, J. Current innovation in virtual and augmented reality in spine surgery. Ann. Transl. Med. 2021, 9, 94. [Google Scholar] [CrossRef] [PubMed]
  36. Chen, F.; Cui, X.; Han, B.; Liu, J.; Zhang, X.; Liao, H. Augmented reality navigation for minimally invasive knee surgery using enhanced arthroscopy. Comput. Methods Programs Biomed. 2021, 201, 105952. [Google Scholar] [CrossRef] [PubMed]
  37. Benmahdjoub, M.; van Walsum, T.; van Twisk, P.; Wolvius, E.B. Augmented reality in craniomaxillofacial surgery: Added value and proposed recommendations through a systematic review of the literature. Int. J. Oral Maxillofac. Surg. 2021, 50, 969–978. [Google Scholar] [CrossRef]
  38. Hussain, R.; Lalande, A.; Guigou, C.; Bozorg-Grayeli, A. Contribution of Augmented Reality to Minimally Invasive Computer-Assisted Cranial Base Surgery. IEEE J. Biomed. Health Inform. 2020, 24, 2093–2106. [Google Scholar] [CrossRef] [PubMed]
  39. Hussain, I.; Cosar, M.; Kirnaz, S.; Schmidt, F.A.; Wipplinger, C.; Wong, T.; Härtl, R. Evolving Navigation, Robotics, and Augmented Reality in Minimally Invasive Spine Surgery. Glob. Spine J. 2020, 10, 22s–33s. [Google Scholar] [CrossRef]
  40. Hu, M.H.; Chiang, C.C.; Wang, M.L.; Wu, N.Y.; Lee, P.Y. Clinical feasibility of the augmented reality computer-assisted spine surgery system for percutaneous vertebroplasty. Eur. Spine J. 2020, 29, 1590–1596. [Google Scholar] [CrossRef]
  41. Gribaudo, M.; Piazzolla, P.; Porpiglia, F.; Vezzetti, E.; Violante, M.G. 3D augmentation of the surgical video stream: Toward a modular approach. Comput. Methods Programs Biomed. 2020, 191, 105505. [Google Scholar] [CrossRef]
  42. Chauvet, P.; Bourdel, N.; Calvet, L.; Magnin, B.; Teluob, G.; Canis, M.; Bartoli, A. Augmented Reality with Diffusion Tensor Imaging and Tractography during Laparoscopic Myomectomies. J. Minim. Invasive Gynecol. 2020, 27, 973–976. [Google Scholar] [CrossRef]
  43. Brebant, V.; Heine, N.; Lamby, P.; Heidekrueger, P.I.; Forte, A.J.; Prantl, L.; Aung, T. Augmented reality of indocyanine green fluorescence in simplified lymphovenous anastomosis in lymphatic surgery. Clin. Hemorheol. Microcirc. 2019, 73, 125–133. [Google Scholar] [CrossRef] [PubMed]
  44. Balla, A.; Sartori, A.; Botteri, E.; Podda, M.; Ortenzi, M.; Silecchia, G.; Guerrieri, M.; Agresta, F. Augmented reality (AR) in minimally invasive surgery (MIS) training: Where are we now in Italy? The Italian Society of Endoscopic Surgery (SICE) ARMIS survey. Updates Surg. 2023, 75, 85–93. [Google Scholar] [CrossRef] [PubMed]
  45. Wild, C.; Lang, F.; Gerhäuser, A.S.; Schmidt, M.W.; Kowalewski, K.F.; Petersen, J.; Kenngott, H.G.; Müller-Stich, B.P.; Nickel, F. Telestration with augmented reality for visual presentation of intraoperative target structures in minimally invasive surgery: A randomized controlled study. Surg. Endosc. 2022, 36, 7453–7461. [Google Scholar] [CrossRef] [PubMed]
  46. Gholizadeh, M.; Bakhshali, M.A.; Mazlooman, S.R.; Aliakbarian, M.; Gholizadeh, F.; Eslami, S.; Modrzejewski, A. Minimally invasive and invasive liver surgery based on augmented reality training: A review of the literature. J. Robot. Surg. 2022. [Google Scholar] [CrossRef] [PubMed]
  47. Godzik, J.; Farber, S.H.; Urakov, T.; Steinberger, J.; Knipscher, L.J.; Ehredt, R.B.; Tumialán, L.M.; Uribe, J.S. “Disruptive Technology” in Spine Surgery and Education: Virtual and Augmented Reality. Oper. Neurosurg. 2021, 21, S85–S93. [Google Scholar] [CrossRef]
  48. Benčurik, V.; Skrovina, M.; Martínek, L.; Bartoš, J.; Macháčková, M.; Dosoudil, M.; Štěpánová, E.; Přibylová, L.; Briš, R.; Anděl, P.; et al. Fluorescent angiography in colorectal surgery, the influence of augmented reality in the operating room on the anastomotic leakage after low rectal resection. Rozhl. Chir. 2021, 100, 543–551. [Google Scholar] [CrossRef]
  49. Pratt, P.; Arora, A. Transoral Robotic Surgery: Image Guidance and Augmented Reality. ORL J. Otorhinolaryngol. Relat. Spec. 2018, 80, 204–212. [Google Scholar] [CrossRef]
  50. Thabit, A.; Benmahdjoub, M.; van Veelen, M.C.; Niessen, W.J.; Wolvius, E.B.; van Walsum, T. Augmented reality navigation for minimally invasive craniosynostosis surgery: A phantom study. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 1453–1460. [Google Scholar] [CrossRef] [PubMed]
  51. Stewart, C.L.; Fong, A.; Payyavula, G.; DiMaio, S.; Lafaro, K.; Tallmon, K.; Wren, S.; Sorger, J.; Fong, Y. Study on augmented reality for robotic surgery bedside assistants. J. Robot. Surg. 2022, 16, 1019–1026. [Google Scholar] [CrossRef]
  52. Rush, A.J., 3rd; Shepard, N.; Nolte, M.; Siemionow, K.; Phillips, F. Augmented Reality in Spine Surgery: Current State of the Art. Int. J. Spine Surg. 2022, 16, S22–S27. [Google Scholar] [CrossRef]
  53. Forte, M.P.; Gourishetti, R.; Javot, B.; Engler, T.; Gomez, E.D.; Kuchenbecker, K.J. Design of interactive augmented reality functions for robotic surgery and evaluation in dry-lab lymphadenectomy. Int. J. Med. Robot. 2022, 18, e2351. [Google Scholar] [CrossRef] [PubMed]
  54. Wendler, T.; van Leeuwen, F.W.B.; Navab, N.; van Oosterom, M.N. How molecular imaging will enable robotic precision surgery: The role of artificial intelligence, augmented reality, and navigation. Eur. J. Nucl. Med. Mol. Imaging 2021, 48, 4201–4224. [Google Scholar] [CrossRef] [PubMed]
  55. Li, R.; Pan, J.; Yang, Y.; Wei, N.; Yan, B.; Liu, H.; Yang, Y.; Qin, H. Accurate and robust feature description and dense point-wise matching based on feature fusion for endoscopic images. Comput. Med. Imaging Graph. 2021, 94, 102007. [Google Scholar] [CrossRef]
  56. Jia, T.; Taylor, Z.A.; Chen, X. Long term and robust 6DoF motion tracking for highly dynamic stereo endoscopy videos. Comput. Med. Imaging Graph. 2021, 94, 101995. [Google Scholar] [CrossRef] [PubMed]
  57. Wang, R.; Zhang, M.; Meng, X.; Geng, Z.; Wang, F.Y. 3-D Tracking for Augmented Reality Using Combined Region and Dense Cues in Endoscopic Surgery. IEEE J. Biomed. Health Inform. 2018, 22, 1540–1551. [Google Scholar] [CrossRef] [PubMed]
  58. Chen, L.; Tang, W.; John, N.W.; Wan, T.R.; Zhang, J.J. SLAM-based dense surface reconstruction in monocular Minimally Invasive Surgery and its application to Augmented Reality. Comput. Methods Programs Biomed. 2018, 158, 135–146. [Google Scholar] [CrossRef]
Figure 1. Preferred reporting items for systematic reviews and meta-analyses (PRISMA) flowchart of search strategy [27].
Figure 1. Preferred reporting items for systematic reviews and meta-analyses (PRISMA) flowchart of search strategy [27].
Bioengineering 10 00501 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Brockmeyer, P.; Wiechens, B.; Schliephake, H. The Role of Augmented Reality in the Advancement of Minimally Invasive Surgery Procedures: A Scoping Review. Bioengineering 2023, 10, 501. https://doi.org/10.3390/bioengineering10040501

AMA Style

Brockmeyer P, Wiechens B, Schliephake H. The Role of Augmented Reality in the Advancement of Minimally Invasive Surgery Procedures: A Scoping Review. Bioengineering. 2023; 10(4):501. https://doi.org/10.3390/bioengineering10040501

Chicago/Turabian Style

Brockmeyer, Phillipp, Bernhard Wiechens, and Henning Schliephake. 2023. "The Role of Augmented Reality in the Advancement of Minimally Invasive Surgery Procedures: A Scoping Review" Bioengineering 10, no. 4: 501. https://doi.org/10.3390/bioengineering10040501

APA Style

Brockmeyer, P., Wiechens, B., & Schliephake, H. (2023). The Role of Augmented Reality in the Advancement of Minimally Invasive Surgery Procedures: A Scoping Review. Bioengineering, 10(4), 501. https://doi.org/10.3390/bioengineering10040501

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop