Next Article in Journal
Differential Inductive Sensing System for Truly Contactless Measuring of Liquids′ Electromagnetic Properties in Tubing
Next Article in Special Issue
Robot-Assisted Gait Self-Training: Assessing the Level Achieved
Previous Article in Journal
Low Spatial Peak-to-Average Power Ratio Transmission for Improved Energy Efficiency in Massive MIMO Systems
Previous Article in Special Issue
Assist-As-Needed Exoskeleton for Hand Joint Rehabilitation Based on Muscle Effort Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery?

1
Centre Hospitalier Intercommunal de POISSY/SAINT-GERMAIN-EN-LAYE 10, Rue Champ de Gaillard, 78300 Poissy, France
2
Department of Hepato-Pancreato-Biliary Surgery, Pederzoli Hospital, 37019 Peschiera del Garda, Italy
3
Department of Surgical, Oncological and Gastroenterological Sciences, University of Padova, 35122 Padova, Italy
4
Department of General-, Visceral-, Vascular- and Transplantation Surgery, University of Magdeburg, Haus 60a, Leipziger Str. 44, 39120 Magdeburg, Germany
5
INKA–Innovation Laboratory for Image Guided Therapy, Medical Faculty, Otto-von-Guericke University Magdeburg, 39120 Magdeburg, Germany
6
School of Computing, Robert Gordon University, Aberdeen AB10 7JG, UK
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5526; https://doi.org/10.3390/s21165526
Submission received: 6 July 2021 / Revised: 3 August 2021 / Accepted: 11 August 2021 / Published: 17 August 2021
(This article belongs to the Special Issue Robotics in Healthcare: Automation, Sensing and Application)

Abstract

:
Most surgeons are skeptical as to the feasibility of autonomous actions in surgery. Interestingly, many examples of autonomous actions already exist and have been around for years. Since the beginning of this millennium, the field of artificial intelligence (AI) has grown exponentially with the development of machine learning (ML), deep learning (DL), computer vision (CV) and natural language processing (NLP). All of these facets of AI will be fundamental to the development of more autonomous actions in surgery, unfortunately, only a limited number of surgeons have or seek expertise in this rapidly evolving field. As opposed to AI in medicine, AI surgery (AIS) involves autonomous movements. Fortuitously, as the field of robotics in surgery has improved, more surgeons are becoming interested in technology and the potential of autonomous actions in procedures such as interventional radiology, endoscopy and surgery. The lack of haptics, or the sensation of touch, has hindered the wider adoption of robotics by many surgeons; however, now that the true potential of robotics can be comprehended, the embracing of AI by the surgical community is more important than ever before. Although current complete surgical systems are mainly only examples of tele-manipulation, for surgeons to get to more autonomously functioning robots, haptics is perhaps not the most important aspect. If the goal is for robots to ultimately become more and more independent, perhaps research should not focus on the concept of haptics as it is perceived by humans, and the focus should be on haptics as it is perceived by robots/computers. This article will discuss aspects of ML, DL, CV and NLP as they pertain to the modern practice of surgery, with a focus on current AI issues and advances that will enable us to get to more autonomous actions in surgery. Ultimately, there may be a paradigm shift that needs to occur in the surgical community as more surgeons with expertise in AI may be needed to fully unlock the potential of AIS in a safe, efficacious and timely manner.

1. Introduction

Unlike artificial intelligence (AI) in medicine, which hopes to use autonomously functioning algorithms to better analyze patient data in an effort to improve patient outcomes, AI in surgery also involves movement. Unlike strict medicine, surgery is an art that is also dynamic. When we use the term “surgery” we are also referring to endoscopy and interventional techniques and procedures, because interventional disciplines continue to coalesce into the same field, a trend that is seen by the continued increase in hybrid operating rooms that have angiography compatible tables, mobile CT scanners, minimally invasive surgical equipment and endoscopes all in the same room that can be used in tandem. Because of the fact that interventional fields of medicine also rely greatly on the medical management of patients, we believe that AI medicine (AIM) and AI surgery (AIS) could one day be considered two distinct disciplines, with AIM reserved for instances where computer algorithms are used to better diagnose, manage or treat patients without a specific interventional procedure being done.
AIS could be a term for autonomously acting machines that can do interventional gestures/actions. According to the Gartner Hype Cycle, many surgeons believe that we are languishing in the “Trough of Disillusionment”, and the promise of autonomous surgery seems like a “pipe dream” for most modern-day surgeons [1,2]. However, the reality is that instances of autonomous actions in surgery already exist. Unfortunately, the reluctance of many laparoscopic surgeons to give up on haptics, or the sense of touch, is actually hindering progress in AIS because of the refusal to embrace robotic tele-manipulation technology, in effect, they are refusing to let go, something that will be needed if the dream of AIS is ever to come to pass [3]. Unfortunately, the medical community has already been shown to be resistant to any automation of medical tasks even simple computations. It is safe to say that automation of surgical tasks will have an even more profound degree of resistance [4].
Another obstacle to the growth of AIS is the dogmatic belief that for something to have AI, it must have algorithms that enable progressive improvement and learning by an artificially intelligent device [5,6]. This creates a conundrum, as theoretically, machine learning (ML) should be infinite, and because of this one wonders what the ultimate purpose of perpetual learning is in surgery. Should it be carried out in the hopes that ultimately the surgical action will become so perfect that it is no longer necessary? How much more perfectly does a robot need to place a titanium clip on the cystic duct?
It could be argued that technology used to create monopolar and bipolar technology is an example of AIS as it has tissue sensing technology that adjusts the action of cautery based on the resistance and capacitance of the tissue (TissueFect™, ValleyLab, Medtronic, Dublin, Ireland). In particular, it has a closed-loop control that analyzes 434,000 data points per second. Does this technology need to improve on that level of data analysis to be considered AI? Or what about Ligasure technology, which uses a high-speed advanced algorithm to seal vessels with changes to the duration of action dependent on tissue thickness (Ligasure, ValleyLab, Medtronic, Dublin, Ireland). We certainly do not mean to imply that there is no room for improvement in these technologies, but at what point should something be defined as AI? Shouldn’t any autonomous action be acknowledged and celebrated as an example of AIS?
ML, Deep Learning (DP), Computer Vision (CV) and possibly Natural Language Processing (NLP) will ultimately be the best path towards more autonomous actions in surgery, but in the meantime, handheld robotics and complete robotic surgical systems are a necessary step that modern day minimally invasive surgeons will need to embrace to ultimately realize the dream of autonomously functioning robots in surgery. This manuscript will discuss the literature on ML, DL, CV and NLP as it pertains to AIM for the pre-operative diagnosis and post-operative management of surgical patients and autonomous robotics in surgery, to attempt to ascertain the current obstacles and necessary next steps in the evolution towards AI/autonomous surgery.

2. Machine Learning

Machine learning (ML) is a genre of artificial intelligence including algorithms that allow machines to solve problems without specific computer programing. While analyzing big data, machines are enabled to assimilate a large amount of information, applicable for risk stratification, diagnosis, treatment decisions, and survival predictions. Not only can AI models analyze large amounts of data collected over long periods of time, providing predictions for future events on the basis of the statistical weight of past correlations, they can also continuously improve with new data. Through a process called “incremental learning”, trainable neural networks improve over time, surpassing unchanging scoring systems and standardized software. Moreover, the human–machine interaction further improves the performance of ML tools. Indeed, the learning process goes far beyond the textbook data, incorporating real-life scenarios and can improve experts’ opinions.
Most of the studies conducted on ML tools have focused on machine vision, biomarker discovery, and clinical matching for diagnosis, classification and outcome prediction [7]. Several studies have applied different ML tools to surgery and, in particular, to risk assessment, performance evaluation, treatment decision making and outcome prediction. In an effort to better identify high-risk surgical patients from complex data, a ML project trained on Pythia was built by Corey et al. to predict postoperative complication risk [8]. By using surgical patient electronic health record (EHR) data, including patient demographics, smoking status, medications, co-morbidities, surgical details, and variables addressing surgical complexity, the authors created a decision support tool for the identification of high-risk patients. Similarly, Bertsimas et al. applied novel ML techniques to design an interactive calculator for emergency surgery [9]. By using data of the American College of Surgeons National Surgical Quality Improvement Program (ACS-NSQIP) database, the authors designed the POTTER application (Predictive OpTimal Trees in Emergency Surgery Risk), to predict postoperative mortality and morbidity [9]. Differently from the standard predictive models, POTTER accounted for nonlinear interactions among variables; thus, the system reboots after each answer, interactively dictating the subsequent question. A similar tool, predicting eight specific postoperative complications and death within one year after major surgery, was developed and validated by Bihorac et al. [10]. MySurgeryRisk can accurately predict acute kidney injury, sepsis, venous thromboembolism, intensive care unit admission >48 h, mechanical ventilation >48 h, and wound, neurologic, and cardiovascular complications with AUC values ranging between 0.82 and 0.94 and death up to 24 months after surgery with AUC values ranging between 0.77 and 0.83. The following studies built different ML models predicting postoperative outcomes that proved to perform better than traditional risk calculators [11,12].
In an attempt to better define patient outcomes, Hung et al. applied ML to objectively measure surgeon performance in a pilot study [13]. By combining procedure-specific automated performance metrics with ML algorithms the authors were able to objectively assess surgeon’s performance after robotic surgery. These algorithms still include biases and misinformation, and thus multi-institutional data from several high-volume groups are needed to train the tool and to create a robust model that is able to correctly predict different clinical outcomes.
Surgical performance has also been measured to allow surgeons to learn from their experience and refine their skills. Either by using surgical video clips or by applying virtual reality scenarios, deep learning models were trained to estimate performance level and specific surgical skills [14,15]. In the near future, we will be able to personalize training through ML tools such as these.
Other studies have used ML for prediction of staging [16], for treatment decision making [17], and to improve case duration estimations [18], further expanding the applicability of ML. Despite these advantages, ML presents several challenges, such as the need to process a large amount of data before it can be analyzed, the necessity of repetitively training the model, and of refining it according to the various clinical scenarios. Ethical considerations should also be taken into account when applying ML to healthcare, including privacy and security issues, and the risk for medico-legal implications. These issues will be discussed at more length below [7].

3. Natural Language Processing

Natural Language Processing (NLP) is the evolution of the interaction of Artificial Intelligence (AI) and linguistics. Distinct from simple text information retrieval (IR), which indexes and searches a large volume of text based on a statistical technique [19], NLP evolved from basic approaches (word to word), through a complex process of coding words, sentences, meanings and contexts, to its modern structure. Since 1950, NLP and IR converged into what is known today as NLP, namely, a computer-based algorithm that handles and elaborates natural (human) language, making it suitable for computation [20].
When applied to healthcare, where available clinical data are kept in Electronic Health Records (EHRs), the need to decode a narrative text coming from this large amount of unstructured data has become urgent because of the complexity of the human language and the routine employment of metaphors and telegraphic prose. When compared to NLP, manual reviewing of EHRs is time consuming, possibly misleading because of biases, and extremely expensive [21].
The tremendous potential value of a big data analytical system in healthcare can be easily explained: EHRs represent at this time the major source of patient information, but unfortunately, for the most part, data regarding symptoms, risk factors for a specific disease or outcomes after medical or surgical treatment come from unstructured text. The ability to translate this amount of information into a coded algorithm could allow for more precise screening programs and modify medical and/or surgical strategies. A systematic review from Kolech et al. analyzed the available methods, employing NLP to interpret symptoms from EHRs of inpatients and outpatients, finding possible future applications for NLP in the normalization of symptoms to controlled vocabularies, in order to avoid overlapping of different words for the same concept [21]. A notable criticism of the available studies has been that reported signs and symptoms are easily mixed as the same variable, making interpretation confusing. In this review, only 11% of studies focused on cancer patients, in contrast with the fact that, currently, a major area of interest for AI (not only NLP) is oncology, where early detection of cancer-specific symptoms could facilitate early diagnosis and potentially enhance screening techniques.
An obvious and immediate advantage of having reliable and decoded data coming from clinical notes is the positive impact on the quality of retrospective studies. Moreover, NLP analysis of symptoms and signs in cancer patients may allow for the improved definition of prognostic factors other than surgical and oncological parameters [22]. Emotional status and quality of life of patients after cancer surgery or other cancer treatment has also been investigated through NLP [23,24]. Banerjee et al., with the creation of a specific domain vocabulary of side-effects after prostatectomy, were able to evaluate minor outcomes hidden in clinical free text, resulting in better management, which could be a game-changer in a population with a 5-year life expectancy rate approaching 99% [23].
When applied to surgery, NPL has been extensively proposed pre-operatively and looking at different post-operative complications such as surgical site infection (SSI). Bucher et al. developed and validated a model of SSI prediction with NLP algorithm by analyzing EHRs from 21,000 patients entered into the ACS-NSQIP, using only text-based documents regarding history and physical condition, operative, progress and nursing notes, radiology reports and discharge summaries [25]. This predictive model had a sensitivity of 79% and a specificity of 92% on external validation, but its added value was the absolute reliable negative predictive value (NPV), which is a relevant issue for events with a low incidence. Anastomotic leak [26], deep venous thrombosis, myocardial infarction and pulmonary embolism were also frequently investigated and results from a recent meta-analysis [27] demonstrated that performance of NLP methods in detection of postoperative complications is similar, if not superior, to non-NLP methods, with a sensitivity of 0.92 versus 0.58 (p < 0.001), and comparable specificity. Moreover, NLP models seem to be better than non-NPL models for ruling out specific surgical outcomes, owing to an optimal true-negative identification power. Interestingly, the ability of algorithms to self-correct can increase the utility of their predictions as datasets grow to become more representative of a patient population.
These NLP applications are surely beneficial for patient management, providing better understanding of peri-operative data, but it can be a useful tool for surgeons as well, particularly when applied to surgical education. For example, decoding intra-operative dialogues between residents and faculty, combining NLP and CV, can create and implement a dataset of technique and surgical strategy, thus, creating a real-life textbook of surgery. In addition, NLP has been efficiently used [28] to assess Entrustable Professional Activities (EPAs). EPAs describe specific behaviors associated with different training levels of residents and it can potentially enhance the understanding of their training and autonomy in surgical practice.
NLP can be used to validate datasets that are the basis of surgical risk predictive models, but the main limit of their widespread use is the non-homogeneity of NLP models and EHRs data entry forms across institutions and countries. A future improvement would entail expansion of registries from local to national and international levels to set algorithms that can be externally validated on various populations. Surgeons and their low-level confidence with AI represent another limit to developing a system that is theoretically perfect and promising: it is important for them to understand how AI may impact healthcare and to elaborate strategies of safe interaction to implement this nascent technology. Synergy between fields of AI is also essential in expanding its applications. Lastly, ethical issues and privacy rules protecting patients’ sensitive data, can limit the large-scale applicability of NLP over EHRs. Nonetheless, the enormous potential of NLP remains fascinating and the multiple potential benefits of its integration into healthcare must be balanced with risks. Although the technology does not currently exist for NLP to influence autonomous actions in surgery, it must be remembered that communication among team members during surgery is fundamental to the successful performance of surgery. Additionally, devices already exist and are used today that are voice-activated (ViKY, Endocontrol, Grenoble, France), and it is conceivable that NLP could eventually evolve to benefit the action of voice-controlled devices during a procedure [29,30].

4. Deep Learning and Computer Vision

Core Computer Vision (CV) tasks include the development of methods to process and analyze digital images and video streams to localize, recognize and track objects of interest [31]. These tasks are considered cornerstones in most autonomously or semi-autonomously acting machines and applications (e.g., robots and self-driving cars) and indeed are critical in AIS and medical image analysis [32]. A typical vision task such as localizing and recognizing objects of interest (e.g., critical findings in chest X-ray images) [33] can be approached either using traditional vision, or advanced deep-learning-based methods (Figure 1).

4.1. Traditional vs. Deep Models

Traditional CV algorithms were based on extracting a set of low-level or high-level features from images or videos (e.g., points of interest, color intensity, edges, etc.), and then using these features to train a supervised learning model such as a support vector machine (SVM) [34], random forest (RF) [35], or other models, to recognize an object of interest or classify an image. Over the past four decades, various methods have been developed to extract key discriminating features from images, including edge detectors [36], scale-invariant features such as SIFT and SURF [37,38], and others, which have helped push the research boundaries in CV as a whole. Despite this, it can be argued that before the widespread development of DL and deep convolutional neural networks (CNN) [39] in 2012, advances in CV were marginal, and even simple vision tasks were considered inherently challenging. One key problem of traditional vision methods is that they rely almost completely on the quality of the extracted features, and in many scenarios, such as AIS, it is difficult to know what features should be considered and then extracted to perform specific tasks. As can be seen in Figure 1, poor features would essentially lead to poor results, regardless of how complicated the learning models are. Furthermore, most of these traditional methods are domain dependent, and are sensitive to noise, light conditions, and the orientation of the objects in the images.
Unlike the aforementioned traditional methods, deep learning methods learn the underlying representation of the image in an end-to-end manner and without the need to handcraft these features [39]. These methods have revolutionized the application of AI across various domains, and significantly improved performance by orders of magnitude in many areas, such as in gaming and AI [40], NLP [41], health [42], medical image analysis [33], and cyber security [43], among others. As can be seen in the schematic diagram in Figure 2 (below), the key advantage of DL methods is the ability to map a set of raw pixels in an image to a particular outcome (e.g., an object’s identity, abnormality) [39], which proved to be inherently challenging tasks for traditional CV methods.

4.2. Advances in Computer Vision

CNN-based methods, in particular, have made significant progress in recent years in the CV domain [44]. They have been successfully applied in several fields such as hand-written recognition [45], image classification [46,47], and face recognition and biometrics [48], among others. Prior to CNNs, the improvements in image classification, segmentation and object detection were marginal and incremental. However, the introduction of CNNs revolutionized this field. For example, Deep Face [49], a face recognition system first proposed by FaceBook in 2014, achieved an accuracy of 97.35%, beating the then state-of-the-art methods by 27%.
Core CV tasks that are considered a cornerstone for any serious AIS attempt [42] such as shape and object detection, recognition, and tracking have become much less challenging even under different conditions and in much less controlled environments. Object detection and recognition have always been the main challenge for traditional CV methods. However, faster region-based CNN (R-CNN) [50], single shot detectors (SSD) [51], region-based fully convolutional networks (R-FCN) [52], and YOLO (you only look once) [53] are all relatively recent methods that have shown superior performance in the field of object detection, tracking, and classification. These methods and their extensions have significantly advanced this area of research and solved some of the most challenging and inherent vision problems, such as occlusions, light conditions, and orientation, among others, which were considered major challenges, even for a specific vision task in a more controlled environment [54]. Additionally, they provide a unique opportunity to advance research and development in AIS and robotic-assisted surgery.
These newly developed methods for localizing and recognizing objects of interest from images have resulted in significant and unprecedented progress in the field of medical image analysis and understanding [55]. For example, CNN-based methods continue to play a key role in advancing classification tasks in the medical domain and providing tools for supporting decision making in clinical settings. Pomponiu et al. [56] used a pre-trained CNN to learn the underlying representation of images and extract representative features that help detect skin cancer. Then, they fed the extracted features into a k nearest neighbor classifier (kNN) and reported an average accuracy of 93% with similar sensitivity and specificity. Similarly, Esteva et al. [57] trained a CNN model using a dataset of 129,450 clinical images representing various skin diseases. The authors also used a pre-trained model to boost performance and reported results comparable with experts. It should be noted that the use of pre-trained models in such tasks is very common and referred to as Transfer Learning (TL).
The key motivation behind such an approach is the fact that training deep models requires large volumes of data (e.g., tens of thousands of images), which may not always be readily available, and can be very expensive to collect and annotate. Therefore, transfer learning provides an opportunity to take advantage of deep models that have already been trained on millions of images. Typical models that are commonly used include AlexNet [47], VGG-16 [58], and GoogleNet Inception [59], among others. A common theme among these models is that they are trained on a large number of images (more than 1.28 million images), and have a deep architecture that enables them to classify 100′s of different objects. In other words, these models are designed to solve very complex tasks for object detection and classification from large volumes of data, and training requires powerful machines and long hours.
TL proved to be very useful in the application of CNN-based methods to various tasks related to cancer detection and diagnosis. A thorough and recent review clearly shows how CNN-based methods, along with pre-trained models, have advanced research in many areas, such as breast, prostate, lung, skin and other types of cancer [60]. Similarly, Levine et al. [61], in a relatively recent review, showed how DL helped to greatly improve cancer diagnosis, and argued that in particular radiology- and pathology-related tasks, DL-based methods achieved comparable performance to medical experts.
CNN-based methods have also greatly advanced research in the area of object localization. This is an important and core task in medical image analysis, by which the machine not only learns to classify an image as normal or abnormal, but also to localize and highlight the area in the image that exhibits the abnormality, if it exists. Examples include the processing and analysis of magnetic resonance imaging (MRI) to detect brain tumors [62], segmenting CT scan images [63], and localizing the area of interest from chest X-Ray images [33]. Such tasks were inherently challenging for conventional ML methods, which requires the careful design of heuristics. However, with DL it became possible to train an end-to-end solution to solve such complex tasks. For instance, Schwab et al. presented a new approach for localizing and recognizing critical findings in chest X-ray images (CHR), using a multi-instance learning method (MIL) that combines classification and detection [33].
MIL is based on dividing the image into a set of patches (patch-based convolutional neural network), and using the prior knowledge of positive and negative images while the proposed method learns which patches in the positive images are negative [33]. Three different public datasets were used, and competitive results were reported. Similarly, Schlemper et al. [63] presented a method incorporating Attention Gate (AG), a mechanism used to help identify salient image regions, into CNN architecture to learn and highlight arbitrary shapes and structures in medical images, and used it for classification and object localization tasks, where promising results were reported in terms of overall prediction performance and segmentation of 3D CT abdominal images.
This progress and advancement in solving such complex vision tasks can be attributed to three main factors. First, the rapid development of the underlying algorithms based on deep neural networks (DNNs) and deep CNNs [39]. Second, the computing power, and the availability of Graphics Processing Unit (GPU)-based machines and cloud services have made it possible to train very deep models using large volumes of images and unstructured documents. The third reason for this progress in CV is the availability of public medical image datasets, helping to accelerate research and development in this field. Examples include datasets related to musculoskeletal radiographs such as MURA [64], which contains 40,561 images from 14,863 studies representing 11,184 different patients. Another example is the colon cancer screening dataset [49], containing 40 colonoscopy videos and containing almost 40,000 image frames, lung image datasets [65], and others. However, more importantly, today there are various online platforms that provide access to similar datasets and provide online computing services to enable researchers to build, test, and share their results. One of the most important platforms is Kaggle (https://www.kaggle.com/ accessed on 16 August 2021), which hosts various types of medical datasets.

4.3. CV in AIS

The significant progress that occurred in object recognition and localization in 2D and 3D images has been reflected in autonomous surgery across different types of applications, including phase recognition [15,66,67], detection and tracking of objects of interest [68,69,70], and segmentation [71,72]. Phase recognition is an important aspect for the training and education of doctors using videos of various types of surgery. However, despite the availability of these videos, their use in training is still limited, because these videos require some sort of pre-processing, and also the segmentation into different phases for subsequent automated skill assessment and feedback [15].
To address this issue, Twinanda et al. [66] built a CNN-based method to perform phase recognition in laparoscopic surgery directly from raw pixels (image and videos). The authors used a dataset of 80 videos of cholecystectomies performed by 13 surgeons to train the CNN, and promising results were reported in terms of the model’s ability to handle complex scenarios and outperform other traditional tools. Similarly, Yu et al. [67] used five different algorithms including CNN for handling the of videos of cataract surgery and Recurrent Neural Networks (RNN) for handling time series data with labels. The results clearly showed that deep learning techniques (CNN and RNN) provide better options for learning from time series data and video images, and can provide accurate and automated detection of phases in cataract surgery. Khalid et al. [15] used deep convolutional neural networks with embedding representation for phase recognition. The authors used 103 video clips of table-top surgical procedures, performed by eight surgeons, including four to five trials of three surgical actions. Promising results with respect to precision, recall and accuracy were reported in terms of the model’s ability to classify performance level and surgical actions.
Object detection and tracking is another important aspect of AI in surgery that has progressed due to the latest developments in deep learning and deep convolutional neural networks. Sarikaya et al. [69] used a dataset of videos from 10 surgeons and applied a deep convolutional neural network to speed up detection and localization of instruments in robot-assisted surgery (RAS). The authors used multimodal CNNs to capture objects of interest and the temporal aspect of the data (motion cues). Results with 91% precision were reported, along with relatively good performance in terms of computational time (0.1 s per frame).
Tracking of objects of interest across different frames is another key aspect of AIS that has also been advanced due to the latest developments in computer vision. Lee et al. [70] proposed a deep-learning-based method for tracking surgical instruments to evaluate a surgeon’s skills in performing procedures using robotic surgery. The authors used 54 videos to train their models and used mean square root error and the area under the curve for evaluation purposes. The results showed that the proposed method was able to accurately track instruments during robotic surgery. The authors concluded that the results suggest that the current method of surgical skill assessment by surgeons could be replaced by the proposed method.
One particular application of CV that has seen significant progress in recent years due to developments of deep-learning-based methods is image and video segmentation. Accurate segmentation of images and videos is crucial for AIS and robot-assisted surgery. A notable example from the Massachusetts Institute of Technology proposed a deep-learning-based approach for robotic instrument segmentation [72]. The authors proposed an architecture based on deep residual models (U-Net) [73]. The method presented provides pixel-level segmentation, where each pixel in the image/video is labeled as an instrument or background, and the authors used eight videos (each one of 255 frames) to train their models, and reported comparable results to those obtained with state-of-the-art methods. Similarly, in [71], the authors used fully convolutional network and optical tracking for segmentation of computer-assisted surgery videos. Overall results of 80.6% for balanced accuracy were reported in a non-real-time version of the method, dropping to 78.2% balanced accuracy in the real-time version.
Various other applications of CV methods and AI can be seen in AIS, including application of CV and AI for education in surgery [74], to improve efficiency in the operating room [75], during neurosurgery [76], and in other surgical disciplines. For a comprehensive and recent review of the use of computer-vision-based methods in AIS and assisted surgery, the reader is referred to [77].

4.4. Reinforcement Learning

Reinforcement learning (RL) is a branch of ML that uses Markov Decision Processes (MDP), which are based on Markov chains, named after Andrey Markov, a Russian mathematician who used this framework to help predict the outcomes of random events [78,79]. In addition to AI in medicine, it has also been applied to autonomous actions in surgery [80]. As noted, enabling the computer to see and recognize things was, in the past, a great hindrance to advancements in AI. Ultimately, an algorithm using a fusion of kinematic and video data was created based on Markov models that enabled segmentation of distinct surgical movements [81]. Combining these innovations with neural networks of DL has enabled vast improvements in the reaction time and overall speed of these complex gestures, and is known as deep reinforcement learning [81].
Currently, most research in RL and surgery involves simulations [82]. Due to the vast amount of data that must be analyzed, researchers first used 2-dimensional (2D) models to devise researchable RL applications. One such model involves a 2D model for teaching gynecological surgeons how to do a hysterectomy. The researchers theorized that by defining precise goals or rewards (also known as “models”), and reproducible tasks or movements mimicking a hysterectomy, that they could combine RL with surgery and develop worthwhile algorithms that could be analyzed, studied and improved [82].
Subsequent teams developed algorithms that were able to control the robotic arms of the da Vinci system and complete tasks in three dimensions (3D), specifically, to identify a real 3D cylinder with precision and accuracy and move the 3D object [83]. Initial work was performed using a simulation and then transferred to real life. They also used MDPs to accomplish these tasks. Ultimately, these researchers were able to program the robot to identify fake blood and shrapnel and to remove this from real life box trainers. These researchers ultimately made this technology open source and called it dVRL, or da Vinci Reinforcement Learning [83].
Other groups also used RL with the da Vinci robot simulator, but with reference to model-free RL algorithms, or non-MDP algorithms [84]. A simple way to understand model-free RL is to think of it as learning based on “trial and error”, as opposed to other approaches that are based on “rewards” or, as mentioned above, “models”. A group from Johns Hopkins termed their non-model algorithm Hybrid Batch Learning (HBL) and found that they could improve the speed of their robot using this approach. They added a second replay buffer to a Q-learning process that used interpreted data from recorded images to streamline decisions on movements, resulting in reduced times for autonomous robotic actions. Q-learning is a form of model-free RL, and may be the most promising form of RL for autonomous actions in surgery as it most closely resembles the human experience [84]. A Q-learning form of RL was also recently developed using the da Vinci Research Kit (Intuitive Surgical, Sunnyvale, CA, USA) to create a system to perform autonomous surgical knots using the robot [85]. However, as in current surgery, what we gain in speed we often end up paying for in increased complications, which brings up the important question of ethics in AI surgical innovation.
Imitation Learning (IL) can be considered an emerging field from the domain of RL, and is a result of recent advances in deep learning and deep convolutional neural networks. The idea is based on teaching an agent (a learning machine/algorithm) to perform a certain task through demonstrations. The overall aim is to learn a mapping function h(x) that creates a map between observations and actions [86]. IL-based methods have been applied successfully to generic navigation tasks, where the agent learns from visual observations (mostly videos or images) [87]. In a recent study, Kim et al. used reinforcement learning to facilitate accurate manipulation and navigation of surgical tools during eye surgery; in particular, they proposed a method to predict the relative goal position on the retinal surface from the current tool-tip position [88].

4.5. Challenges in AIS

Although CV tasks play a crucial role in AIS, autonomous surgery is way more complex than computer vision. It involves movement, navigation, object recognition, classification, taking actions, and much more. Each of these components is a challenging problem by itself, and this can explain why having fully autonomously acting machines in surgery is still far from a reality. That said, the development in CV tasks has made it possible to provide some form of robot-assisted surgery that can help in decision making and acting processes [89]. From a computer vision perspective, there are several challenges that can be considered to be barriers to the widespread development of AIS and robot-assisted surgery. However, two key challenges include data availability and data annotation.

4.5.1. Data Availability

Although various datasets are available in the public domain (as discussed earlier), a very small number of these datasets are designed for autonomous or semi-autonomous surgery. Such datasets are very important for advancing research in AIS and semi-autonomous surgeries [90]. An example of an attempt to prepare a dedicated dataset for AIS is the work presented in [91], which presented an experimental framework to evaluate the performance of CV methods in analyzing operative videos to identify steps of a laparoscopic sleeve gastrectomy (LSG). The authors collected videos capturing operations from start to finish for 18+ patients, and using DNNs, they reported an overall accuracy of 85.6% in identifying operative steps in laparoscopic sleeve gastrectomy, such as port placement, retraction of the liver, liver biopsy, etc.

4.5.2. Data Annotation

The second challenge, which is also closely related to data, is the data annotation process. The annotation process can be defined as simply the process of allocating and annotating the areas of interest in a video or an image. For example, consider a classification task that aims at learning whether an X-ray contains an abnormality or not. This requires experts to label large volumes of such images as either positive or negative instances. Such tasks, however, become more complex when we are talking about defining and allocating areas of interest in a video that captures a full surgical procedure. This process is expensive, time consuming and largely depends on a human’s expertise. Moreover, the annotation process is performed almost manually, where the users are required to manually allocate these areas of interest, often using open-source tools that capture the location and boundaries of these areas.

4.5.3. Ethics of Technological Advancements in AIS

The ethics of the introduction of new surgical devices is a complex issue that has been divided into four broad principles: beneficence, non-maleficence, autonomy and justice [92]. In general, six groups of people are considered when discussing this issue: the medical regulatory board, surgeons, patients, hospitals and surgical centers, industry and surgical societies [93]. The issue of ethics as it pertains to artificial intelligence surgery or more specifically autonomous actions in surgery is unclear because in general surgical robots that have already been approved for use in humans will ultimately be the vehicle used to deliver the artificially intelligent movements. Technically, the artificial intelligence will not be inside of the patient, but the robotic arms will be. Approval for technology that does not enter into the patient is generally easier to obtain then for new technology that goes inside patients. Current examples of autonomously functioning devices in patients include, for example, automatic staplers (iDrive, Medtronic, Dublin, Ireland). As mentioned, this device has a sensor and changes the speed of stapling and functions autonomously once activated. As autonomous actions become more complex, it begs the question of whether or not approval of more artificially intelligent devices will require a more rigorous approval process, even if the AI technical remains outside of patients.
As opposed to the ethical issues of AI in the collection, manipulation and interpretation of medical data, AIS has the added potential danger of real-time analysis of intra-operative issues and potential for complications [94]. Alternatively, it could be argued that AI may result in fewer complications because of technology devised to minimize complications. Clearly, we are many years away from being able to truly study this; nonetheless, it is clear that more surgeons need to become well-versed with issues of AI so that surgeons can truly partner with engineers and computer scientists in the development of safe autonomous actions.
Currently, four classes of medical devices based on risk have been designated. Class 1 devices have minimal risk to the patient and remain in limited contact with the patient and include surgical instruments; class 2 devices include things like CT scanners, ultrasound probes and contact lenses; class 3 includes hemodialysis and implants; and class 4 includes implantable defibrillators and pacemakers [95]. One wonders if a fifth class should be designated for devices that make real-time clinical decisions using AI during procedures. The consequences of not having surgeons be intimately knowledgeable and involved in the development of new technologies such as surgical robotics and AI have been divided into five ethical categories of experience for surgeons and additionally patient expectations and include: rescue, proximity, ordeal, aftermath and presence [96]. It is in this spirit of humility and acknowledgement of the fundamental role of morality in surgery that this article was written.

5. Discussion

All fully autonomous robots have elements of ML, DL, CV and NLP. Although so-called “Strong” AI does not yet exist in surgical robots, examples of “Weak” AI do exist and permit the independent performance of automatic linear stapled gastrointestinal anastomoses and robotic arm alignment during operating table tilting [2,6]. Another way to define Weak AI in surgical robotics is to consider things like the iDrive and da Vinci/Wolf table as reactive agents only as opposed to deliberative agents. Deliberative agents are the hope for Strong AI and would include information gathering, knowledge improvement, exploration, learning and full autonomy. Reactive agents are also known as “finite automata”, and only carry out simple tasks. Essentially, “Weak” AI performs an automatic action, but without the idealized concept of intelligence that one would expect in AI, specifically, a complex interpretation of a problem with an advanced solution equal or superior to a humans. Nonetheless, these examples of autonomous actions are the small steps that we need on the road towards more complete automation. Some researchers in AI believe that AI must ultimately include computers with desires, beliefs and intentions. Although an automated surgical robot should ultimately have the ability to act with intention, it is hard to see the need for a robot having beliefs or desires to function effectively during an interventional procedure; however, this may not be the case in the future once AI is much more advanced [97].
In accordance with this, an international group of robotic surgical experts recently defined six levels of surgical autonomy: level 0 designates no autonomy, level 1 is defined as the tele-manipulation of a robot, level 2 corresponds to a limited autonomous action; level 3 is defined as conditional autonomy where the surgeon selects among several autonomous actions that the robot can perform; level 4 autonomous actions include automated medical decision making under a doctor’s control; and level 5 indicates full autonomy where no human control is necessary. It would appear that the complete surgical system as it is known today fits level 1, but vessel sealing devices such as the Ligasure or cautery devices would be level 2, and automatic stapling devices that have internal sensors could arguably be described as meeting the criteria of level 3. Interestingly, a team from Germany modified a previously developed robotically controlled laparoscope holder [29] to move autonomously and improve with time, resulting in decreased laparoscopic cholecystectomy times [98]. This device could indicate that we have already entered level 4 in humans [98].
In the clinical world, ML and DL are also currently limited to off-line data interpretation such as in Radiomics and evaluation of surgical performance [15,99]. The power of DL is elucidated by the observation that convolutional neural networks have already been shown to function at the level of average breast radiologists in differentiating radiographic breast findings [100]. CV, also known as robotic vision, also relies on DL. DL is a form of ML that creates models from large and complex data sets; however, ML can also create models from sets of data that are not pre-labeled which is known as Unsupervised Learning and models that use algorithms to interact with the environment in real time, which is known as Reinforcement Learning. Ultimately, more autonomous robotic actions will require superior methods of Unsupervised and Reinforcement Learning [101]. The shear amount of data necessary to interpret the vast amount of data generated from robotics in real time is hard to fathom. The hope is that NLP may ultimately be able to facilitate this interaction with man, robot and computer. However, one has to wonder if there is one fundamental flaw in the way that surgical robots are being designed.
Although robotic autonomy has always been the driving factor in even the earliest surgical robots [102,103], a counter-intuitive push has existed for the development of haptics or the perception of touch [104,105]. Measurement of the force exerted on robotic tips focuses on measurements obtained from the tip of robotic instruments. Designs have ranged from two-degree-of-freedom force readings via torsional springs and flexure hinges to graspers with four-degree-of-freedom readings [106,107]. More sensitive robotic tip sensors have also been developed with polyvinylidene fluoride (PVDF) that have the strength to detect force measurements of graspers, but which also detect pulsations of arteries [108]. The fact that the sensors are at the tip of the instrument leads to a whole host of problematic issues from size constraints of components to limitations involving need for repetitive sterilization.
If the goal of robotic surgery is full automation, and all surgeons who use the complete surgical systems operating today are able to perform surgery without haptics, relying on visual cues alone, one has to wonder if the development of haptics is truly necessary. A CNN-based framework to autonomously control guide wires in vessels during angiography is currently being developed that will use haptics recorded during robotic angiography performed by a clinician to develop algorithms that can enable an autonomously functioning robot to successfully navigate vessels intra-luminally [109]. As with humans, the current models involve two-dimensional image interpretation with fluoroscopy being converted into three-dimensional robotic movements [110,111]. A key aspect of this process is that although the robot will be trained with haptic information, ultimately the procedure will be performed without the need for the clinician to sense and interpret the haptic information. This indicates that although haptics may be fundamental to developing these movement algorithms, ultimately haptics interpreted by humans may no longer be necessary.
Interestingly, from the very beginning of surgical robotics, researchers noticed a problem with “noise” and ultimately had to use filters to obtain interpretable data [102]. What if the robot did not need to feel like humans, and what if the noise is what we should be focusing on? Interestingly, the Versius Complete Surgical System has haptic capabilities, but the detection of resting human tremor makes the haptics useless and potentially bothersome [3]. Symbolically, one of the first tools in modern medicine was sound, and, as highlighted by the development of the stethoscope, it is in this spirit that researchers began to harness the information generated by friction, bumps and perforations by analyzing them with various ML algorithms, finding that differentiation of these events was possible during vascular catheterizations with guidewires [112]. The ML algorithms included artificial neural networks, K-nearest Neighbor and Support Vector Machine. Unlike the sensors of robotic graspers, however, this sensor was able to be placed on the proximal end of the guide wire and not on the distal part in the patient [113,114].
Interestingly, this technology has been shown to have the ability to generate quantifiable information from robotic graspers, while the sensor itself still remains outside of the patient [115,116,117]. The potential of this technology has also been shown by demonstrating that everyday procedures like needle insertions can become smarter and potentially safer, specifically in the case of Veress insertion for pneumoperitoneum access, liver ablation of tumors and arthroscopic insertions [118,119].

6. Conclusions

The complexity of the field of artificial intelligence surgery is highlighted by the finding that only 9.8% of medical devices undergo in-human testing within 10 years, and that involvement with an actual clinician during the development of new devices significantly improved the chances of this [120]. Advancing AIS requires compiling large volumes of videos that captures surgery procedures. This requires hundreds if not thousands of fully annotated videos for each specific type of surgery that can be used and shared within the research community. To meet this requirement, data collection, preparation and annotation must be part of future medical practice. It also requires close and cross-disciplinary collaboration from the AI and medical communities. Autonomous actions in robotic surgery will involve a complex interplay of ML, DL, CV and potentially NLP. Modern surgeons must become versed with the basics of AI to better incorporate this exciting new field into surgical practice. Young academic surgeons should consider gaining experience in this field in the form of Masters or PhD programs, as opposed to more traditional fields of study such as molecular biology, genetics and immunology.

Author Contributions

Conceptualization, A.A.G.; writing—original draft, A.A.G., I.F., G.S. and E.E.; writing—review and editing, A.A.G., I.F., G.S., R.C., A.I., E.C. and E.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Oosterhoff, J.H.; Doornberg, J.N. Machine Learning Consortium Artificial intelligence in orthopaedics: False hope or not? A narrative review along the line of Gartner’s hype cycle. EFORT Open Rev. 2020, 5, 593–603. [Google Scholar] [CrossRef]
  2. Gumbs, A.A.; Perretta, S.; D’Allemagne, B.; Chouillard, E. What is Artificial Intelligence Surgery? Artif. Intell. Surg. 2021, 1, 1–10. [Google Scholar] [CrossRef]
  3. Gumbs, A.A.; De Simone, B.; Chouillard, E. Searching for a better definition of robotic surgery: Is it really different from laparoscopy? Mini Invasive Surg. 2020, 2020, 90. [Google Scholar] [CrossRef]
  4. Randolph, A.G.; Haynes, R.B.; Wyatt, J.C.; Cook, D.J.; Guyatt, G.H. Users’ Guides to the Medical Literature: XVIII. How to use an article evaluating the clinical impact of a computer-based clinical decision support system. JAMA 1999, 282, 67–74. [Google Scholar] [CrossRef] [PubMed]
  5. Kassahun, Y.; Yu, B.; Tibebu, A.T.; Stoyanov, D.; Giannarou, S.; Metzen, J.H.; Poorten, E.V. Surgical robotics beyond enhanced dexterity instrumentation: A survey of machine learning techniques and their role in intelligent and autonomous surgical actions. Int. J. Comput. Assist. Radiol. Surg. 2015, 11, 553–568. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Hashimoto, D.A.; Rosman, G.; Rus, D.; Meireles, O.R. Artificial Intelligence in Surgery: Promises and Perils. Ann. Surg. 2018, 268, 70–76. [Google Scholar] [CrossRef]
  7. Ngiam, K.Y.; Khor, I.W. Big data and machine learning algorithms for health-care delivery. Lancet Oncol. 2019, 20, e262–e273. [Google Scholar] [CrossRef]
  8. Corey, K.M.; Kashyap, S.; Lorenzi, E.; Lagoo-Deenadayalan, S.A.; Heller, K.; Whalen, K.; Balu, S.; Heflin, M.T.; McDonald, S.R.; Swaminathan, M.; et al. Development and validation of machine learning models to identify high-risk surgical patients using automatically curated electronic health record data (Pythia): A retrospective, single-site study. PLoS Med. 2018, 15, e1002701. [Google Scholar] [CrossRef] [Green Version]
  9. Bertsimas, D.; Dunn, J.; Velmahos, G.C.; Kaafarani, H.M.A. Surgical Risk Is Not Linear: Derivation and Validation of a Novel, User-friendly, and Machine-learning-based Predictive OpTimal Trees in Emergency Surgery Risk (POTTER) Calculator. Ann. Surg. 2018, 268, 574–583. [Google Scholar] [CrossRef]
  10. Bihorac, A.; Ozrazgat-Baslanti, T.; Ebadi, A.; Motaei, A.; Madkour, M.; Pardalos, P.M.; Lipori, G.; Hogan, W.R.; Efron, P.A.; Moore, F.; et al. MySurgeryRisk: Development and Validation of a Machine-learning Risk Algorithm for Major Complications and Death After Surgery. Ann. Surg. 2019, 269, 652–662. [Google Scholar] [CrossRef]
  11. Chiew, C.J.; Liu, N.; Wong, T.H.; Sim, Y.E.; Abdullah, H.R. Utilizing Machine Learning Methods for Preoperative Prediction of Postsurgical Mortality and Intensive Care Unit Admission. Ann. Surg. 2020, 272, 1133–1139. [Google Scholar] [CrossRef]
  12. El Hechi, M.W.; Eddine, S.A.N.; Maurer, L.; Kaafarani, H.M. Leveraging interpretable machine learning algorithms to predict postoperative patient outcomes on mobile devices. Surgery 2021, 169, 750–754. [Google Scholar] [CrossRef]
  13. Hung, A.J.; Chen, J.; Gill, I.S. Automated Performance Metrics and Machine Learning Algorithms to Measure Surgeon Performance and Anticipate Clinical Outcomes in Robotic Surgery. JAMA Surg. 2018, 153, 770–771. [Google Scholar] [CrossRef]
  14. Winkler-Schwartz, A.; Yilmaz, R.; Mirchi, N.; Bissonnette, V.; Ledwos, N.; Siyar, S.; Azarnoush, H.; Karlik, B.; Del Maestro, R. Machine Learning Identification of Surgical and Operative Factors Associated With Surgical Expertise in Virtual Reality Simulation. JAMA Netw. Open 2019, 2, e198363. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Khalid, S.; Goldenberg, M.; Grantcharov, T.; Taati, B.; Rudzicz, F. Evaluation of Deep Learning Models for Identifying Surgical Actions and Measuring Performance. JAMA Netw. Open 2020, 3, e201664. [Google Scholar] [CrossRef] [PubMed]
  16. O’Sullivan, B.; Huang, S.H.; De Almeida, J.R.; Hope, A. Alpha Test of Intelligent Machine Learning in Staging Head and Neck Cancer. J. Clin. Oncol. 2020, 38, 1255–1257. [Google Scholar] [CrossRef] [PubMed]
  17. Maubert, A.; Birtwisle, L.; Bernard, J.; Benizri, E.; Bereder, J. Can machine learning predict resecability of a peritoneal carcinomatosis? Surg. Oncol. 2019, 29, 120–125. [Google Scholar] [CrossRef]
  18. Bartek, M.A.; Saxena, R.C.; Solomon, S.; Fong, C.T.; Behara, L.D.; Venigandla, R.; Velagapudi, K.; Lang, J.D.; Nair, B.G. Improving Operating Room Efficiency: Machine Learning Approach to Predict Case-Time Duration. J. Am. Coll. Surg. 2019, 229, 346–354.e3. [Google Scholar] [CrossRef]
  19. Nadkarni, P.M.; Ohno-Machado, L.; Chapman, W. Natural language processing: An introduction. J. Am. Med. Inform. Assoc. 2011, 18, 544–551. [Google Scholar] [CrossRef] [Green Version]
  20. Yim, W.W.; Yetisgen, M.; Harris, W.P.; Kwan, S.W. Natural Language Processing in Oncology: A Review. JAMA Oncol. 2016, 2, 797–804. [Google Scholar] [CrossRef] [PubMed]
  21. Koleck, T.A.; Dreisbach, C.; Bourne, P.E.; Bakken, S. Natural language processing of symptoms documented in free-text narratives of electronic health records: A systematic review. J. Am. Med. Inform. Assoc. 2019, 26, 364–379. [Google Scholar] [CrossRef]
  22. Hughes, K.S.; Zhou, J.; Bao, Y.; Singh, P.; Wang, J.; Yin, K. Natural language processing to facilitate breast cancer research and management. Breast J. 2019, 26, 92–99. [Google Scholar] [CrossRef]
  23. Banerjee, I.; Li, K.; Seneviratne, M.; Ferrari, M.; Seto, T.; Brooks, J.D.; Rubin, D.L.; Hernandez-Boussard, T. Weakly supervised natural language processing for assessing patient-centered outcome following prostate cancer treatment. JAMIA Open 2019, 2, 150–159. [Google Scholar] [CrossRef] [Green Version]
  24. Zunic, A.; Corcoran, P.; Spasic, I. Sentiment Analysis in Health and Well-Being: Systematic Review. JMIR Med. Inform. 2020, 8, e16023. [Google Scholar] [CrossRef] [PubMed]
  25. Bucher, B.T.; Shi, J.; Ferraro, J.P.; Skarda, D.E.; Samore, M.H.; Hurdle, J.F.; Gundlapalli, A.V.; Chapman, W.W.; Finlayson, S.R.G. Portable Automated Surveillance of Surgical Site Infections Using Natural Language Processing: Development and Validation. Ann. Surg. 2020, 272, 629–636. [Google Scholar] [CrossRef]
  26. Soguero-Ruiz, C.; Hindberg, K.; Rojo-Alvarez, J.L.; Skrovseth, S.O.; Godtliebsen, F.; Mortensen, K.; Revhaug, A.; Lindsetmo, R.-O.; Augestad, K.M.; Jenssen, R. Support Vector Feature Selection for Early Detection of Anastomosis Leakage from Bag-of-Words in Electronic Health Records. IEEE J. Biomed. Health Inform. 2014, 20, 1404–1415. [Google Scholar] [CrossRef] [PubMed]
  27. Mellia, J.A.; Basta, M.N.; Toyoda, Y.; Othman, S.; Elfanagely, O.; Morris, M.P.; Torre-Healy, L.; Ungar, L.H.; Fischer, J.P. Natural Language Processing in Surgery: A Systematic Review and Meta-analysis. Ann. Surg. 2021, 273, 900–908. [Google Scholar] [CrossRef]
  28. Stahl, C.C.; Jung, S.A.; Rosser, A.A.; Kraut, A.S.; Schnapp, B.H.; Westergaard, M.; Hamedani, A.G.; Minter, R.M.; Greenberg, J.A. Natural language processing and entrustable professional activity text feedback in surgery: A machine learning model of resident autonomy. Am. J. Surg. 2021, 221, 369–375. [Google Scholar] [CrossRef] [PubMed]
  29. Gumbs, A.A.; Crovari, F.; Vidal, C.; Henri, P.; Gayet, B. Modified Robotic Lightweight Endoscope (ViKY) Validation In Vivo in a Porcine Model. Surg. Innov. 2007, 14, 261–264. [Google Scholar] [CrossRef]
  30. Gumbs, A.A.; Croner, R.; Rodriguez, A.; Zuker, N.; Perrakis, A.; Gayet, B. 200 Consecutive laparoscopic pancreatic resections performed with a robotically controlled laparoscope holder. Surg. Endosc. 2013, 27, 3781–3791. [Google Scholar] [CrossRef]
  31. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Prentice-Hall, Inc.: Hoboken, NJ, USA, 2008. [Google Scholar]
  32. Vyborny, C.J.; Giger, M.L. Computer vision and artificial intelligence in mammography. Am. J. Roentgenol. 1994, 162, 699–708. [Google Scholar] [CrossRef]
  33. Schwab, E.; Goossen, A.; Deshpande, H.; Saalbach, A. Localization of Critical Findings in Chest X-Ray Without Local Annotations Using Multi-Instance Learning. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1879–1882. [Google Scholar]
  34. Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef]
  35. Elyan, E.; Gaber, M.M. A genetic algorithm approach to optimising random forests applied to class engineered data. Inf. Sci. 2017, 384, 220–234. [Google Scholar] [CrossRef]
  36. Ding, L.; Goshtasby, A. On the Canny edge detector. Pattern Recognit. 2001, 34, 721–725. [Google Scholar] [CrossRef]
  37. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  38. Bay, H.; Tuytelaars, T.; Gool, L.V. SURF: Speeded Up Robust Features. Computer Vision—ECCV 2006; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  39. Goodfellow, I.; Benjio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  40. Holcomb, S.D.; Porter, W.K.; Ault, S.V.; Mao, G.; Wang, J. Overview on DeepMind and Its AlphaGo Zero AI. In Proceedings of the 2018 International Conference on Big Data and Education, Honolulu, HI, USA, 9–11 March 2018; pp. 67–71. [Google Scholar]
  41. Yang, Z.; Yang, D.; Dyer, C.; He, X.; Smola, A.; Hovy, E. Hierarchical attention networks for document classification. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 1480–1489. [Google Scholar]
  42. Esteva, A.; Robicquet, A.; Ramsundar, B.; Kuleshov, V.; Depristo, M.; Chou, K.; Cui, C.; Corrado, G.; Thrun, S.; Dean, J. A guide to deep learning in healthcare. Nat. Med. 2019, 25, 24–29. [Google Scholar] [CrossRef] [PubMed]
  43. Shone, N.; Ngoc, T.N.; Phai, V.D.; Shi, Q. A deep learning approach to network in-trusion detection. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 41–50. [Google Scholar] [CrossRef] [Green Version]
  44. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
  45. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  46. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  47. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  48. Park, U.; Jain, A.K. Face matching and retrieval using soft biometrics. IEEE Trans. Inf. Forensics Secur. 2010, 5, 406–415. [Google Scholar] [CrossRef] [Green Version]
  49. Taigman, Y.; Yang, M.; Ranzato, M.; Wolf, L. DeepFace: Closing the Gap to Human-Level Performance in Face Verification. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar]
  50. Ren, S.; He, K.; Girshick, R. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.; Berg, A.C. SSD: Single Shot Multibox Detector; CoRR: Ithaca, NY, USA, 2015. [Google Scholar]
  52. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object Detection via Region-Based Fully Convolutional Networks; CoRR: Ithaca, NY, USA, 2016. [Google Scholar]
  53. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef] [Green Version]
  54. Zhao, W.; Chellappa, R.; Phillips, P.J.; Rosenfeld, A. Face recognition: A literature survey. ACM Comput. Surv. 2003, 35, 399–458. [Google Scholar] [CrossRef]
  55. Shen, D.; Wu, G.; Suk, H.-I. Deep Learning in Medical Image Analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Pomponiu, V.; Nejati, H.; Cheung, N.-M. Deepmole: Deep neural networks for skin mole lesion classification. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2623–2627. [Google Scholar]
  57. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  58. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  59. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2818–2826. [Google Scholar] [CrossRef] [Green Version]
  60. Hu, Z.; Tang, J.; Wang, Z.; Zhang, K.; Zhang, L.; Sun, Q. Deep learning for image-based cancer detection and diagnosis—A survey. Pattern Recognit. 2018, 83, 134–149. [Google Scholar] [CrossRef]
  61. Levine, A.B.; Schlosser, C.; Grewal, J.; Coope, R.; Jones, S.; Yip, S. Rise of the Machines: Advances in Deep Learning for Cancer Diagnosis. Trends Cancer 2019, 5, 157–169. [Google Scholar] [CrossRef]
  62. Wadhwa, A.; Bhardwaj, A.; Verma, V.S. A review on brain tumor segmentation of MRI images. Magn. Reson. Imaging 2019, 61, 247–259. [Google Scholar] [CrossRef]
  63. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
  64. Rajpurkar, P.; Irvin, J.; Bagul, A.; Ding, D.; Duan, T.; Mehta, H.; Yang, B.; Zhu, K.; Laird, D.; Ball, R.L.; et al. Mura: Large Dataset for Abnormality Detection in Musculoskeletal Radiographs. arXiv 2018, arXiv:1712.06957. [Google Scholar]
  65. Armato, S.G., III; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The Cancer Imaging Archive (TCIA). Available online: https://wiki.cancerimagingarchive.net/display/Public/LIDC-IDRI (accessed on 16 August 2021).
  66. Twinanda, A.; Shehata, S.; Mutter, D.; Marescaux, J.; de Mathelin, M.; Padoy, N. Endonet: A deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 2017, 36, 86–97. [Google Scholar] [CrossRef] [Green Version]
  67. Yu, F.; Croso, G.S.; Kim, T.S.; Song, Z.; Parker, F.; Hager, G.D.; Reiter, A.; Vedula, S.S.; Ali, H.; Sikder, S. Assessment of Automated Identification of Phases in Videos of Cataract Surgery Using Machine Learning and Deep Learning Techniques. JAMA Netw. Open 2019, 2, e191860. [Google Scholar] [CrossRef] [PubMed]
  68. Marban, A.; Srinivasan, V.; Samek, W.; Fernandez, J.; Casals, A. Estimating position amp; velocity in 3d space from monocular video sequences using a deep neural network. Proceedings of IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 1460–1469. [Google Scholar]
  69. Sarikaya, D.; Corso, J.J.; Guru, K.A. Detection and Localization of Robotic Tools in Robot-Assisted Surgery Videos Using Deep Neural Networks for Region Proposal and Detection. IEEE Trans. Med. Imaging 2017, 36, 1542–1549. [Google Scholar] [CrossRef] [PubMed]
  70. Lee, D.; Yu, H.W.; Kwon, H.; Kong, H.; Lee, K.; Kim, H. Evaluation of surgical skills during robotic surgery by deep learning-based multiple surgical instrument tracking in training and actual operations. J. Clin. Med. 2020, 9, 1964. [Google Scholar] [CrossRef] [PubMed]
  71. Garcıa-Peraza-Herrera, L.; Li, W.; Gruijthuijsen, C.; Devreker, A.; Attilakos, G.; Deprest, J.; Vander Poorten, E.; Stoyanov, D.; Vercauteren, T.; Ourselin, S. Real-time segmentation of non-rigid surgical tools based on deep learning and tracking. In Computer-Assisted and Robotic Endoscopy; Peters, T., Yang, G., Navab, N., Mori, K., Luo, X., Reichl, T., McLeod, J., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  72. Shvets, A.; Rakhlin, A.; Kalinin, A.; Iglovikov, V. Automatic instrument segmentation in robot-assisted surgery using deep learning. Proceedings of 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 624–628. [Google Scholar]
  73. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 1–26 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  74. Ward, T.M.; Mascagni, P.; Madani, A.; Padoy, N.; Perretta, S.; Hashimoto, D.A. Surgical data science and artificial intelligence for surgical education. J. Surg. Oncol. 2021, 124, 221–230. [Google Scholar] [CrossRef]
  75. Birkhoff, D.C.; van Dalen, A.S.H.; Schijven, M.P. A Review on the Current Applications of Artificial Intelligence in the Operating Room. Surg. Innov. 2021, 28. [Google Scholar] [CrossRef]
  76. Pangal, D.J.; Kugener, G.; Shahrestani, S.; Attenello, F.; Zada, G.; Donoho, D.A. A Guide to Annotation of Neurosurgical Intraoperative Video for Machine Learning Analysis and Computer Vision. World Neurosurg. 2021, 150, 26–30. [Google Scholar] [CrossRef]
  77. Kennedy-Metz, L.; Mascagni, P.; Torralba, A.; Dias, R.; Perona, P.; Shah, J.; Padoy, N.N.; Zenati, M. Computer vision in the operating room: Opportunities and caveats. IEEE Trans. Med. Robot. Bionics 2020, 3, 2–10. [Google Scholar] [CrossRef] [PubMed]
  78. Hua, J.; Zeng, L.; Li, G.; Ju, Z. Learning for a Robot: Deep Reinforcement Learning, Imitation Learning, Transfer Learning. Sensors 2021, 21, 1278. [Google Scholar] [CrossRef]
  79. Zhang, T.; Mo, H. Reinforcement learning for robot research: A comprehensive review and open issues. Int. J. Adv. Robot. Syst. 2021, 18, 17298814211007305. [Google Scholar] [CrossRef]
  80. Datta, S.; Li, Y.; Ruppert, M.M.; Ren, Y.; Shickel, B.; Ozrazgat-Baslanti, T.; Rashidi, P.; Bihorac, A. Reinforcement learning in surgery. Surgery 2021, 170, 329–332. [Google Scholar] [CrossRef]
  81. Gao, X.; Jin, Y.; Dou, Q.; Heng, P.-A. Automatic Gesture Recognition in Robot-assisted Surgery with Reinforcement Learning and Tree Search. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 31 May–31 August 2020; pp. 8440–8446. [Google Scholar]
  82. Sato, M.; Koga, K.; Fujii, T.; Osuga, Y. Can Reinforcement Learning Be Applied to Surgery. In Artificial Intelligence—Emerging Trends and Applications; IntechOpen: London, UK, 2018. [Google Scholar]
  83. Richter, F.; Member, S.; Orosco, R.K.; Yip, M.C. dVRL: Reinforcement Learning Environments for Surgical Robotics. arXiv 2019, arXiv:1903.02090. [Google Scholar]
  84. Barnoy, Y.; O’Brien, M.; Wang, W.; Hager, G. Robotic Surgery with Lean Reinforcement Learning. arXiv 2021, arXiv:2105.01006. [Google Scholar]
  85. Varier, V.M.; Rajamani, D.K.; Goldfarb, N.; Tavakkolmoghaddam, F.; Munawar, A.; Fischer, G.S. Collaborative Suturing: A Reinforcement Learning Approach to Automate Hand-off Task in Suturing for Surgical Robots. In Proceedings of the 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 1380–1386. [Google Scholar]
  86. Hussein, A.; Gaber, M.M.; Elyan, E. Imitation Learning: A Survey of Learning Methods. ACM Comput. Surv. 2017, 50, 1–35. [Google Scholar] [CrossRef]
  87. Hussein, A.; Elyan, E.; Gaber, M.M.; Jayne, C. Deep imitation learning for 3D navigation tasks. Neural Comput. Appl. 2018, 29, 389–404. [Google Scholar] [CrossRef] [Green Version]
  88. Kim, J.W.; Zhang, P.; Gehlbach, P.; Iordachita, I.; Kobilarov, M. Towards Autonomous Eye Surgery by Combining Deep Imitation Learning with Optimal Control. arXiv 2020, arXiv:2011.07778. [Google Scholar]
  89. Luongo, F.; Hakim, R.; Nguyen, J.H.; Anandkumar, A.; Hung, A.J. Deep learning-based computer vision to recognize and classify suturing gestures in robot-assisted surgery. Surgery 2021, 169, 1240–1244. [Google Scholar] [CrossRef] [PubMed]
  90. Madapana, N.; Low, T.; Voyles, R.M.; Xue, Y.; Wachs, J.; Rahman, M.; Sanchez-Tamayo, N.; Balakuntala, M.V.; Gonzalez, G.; Bindu, J.P.; et al. DESK: A Robotic Activity Dataset for Dexterous Surgical Skills Transfer to Medical Robots. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 6928–6934. [Google Scholar]
  91. Hashimoto, D.A.; Rosman, G.; Witkowski, E.R.; Stafford, C.; Navarette-Welton, A.J.; Rattner, D.W.; Lillemoe, K.D.; Rus, D.L.; Meireles, O.R. Computer Vision Analysis of Intraoperative Video: Automated Recognition of Operative Steps in Laparoscopic Sleeve Gastrectomy. Ann. Surg. 2019, 270, 414–421. [Google Scholar] [CrossRef] [PubMed]
  92. Angelos, P. Complications, Errors, and Surgical Ethics. World J. Surg. 2009, 33, 609–611. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  93. Ross, S.; Robert, M.; Harvey, M.-A.; Farrell, S.; Schulz, J.; Wilkie, D.; Lovatsis, D.; Epp, A.; Easton, B.; McMillan, B.; et al. Ethical issues associated with the introduction of new surgical devices, or just because we can, doesn’t mean we should. J. Obstet. Gynaecol. Can. 2008, 30, 508–513. [Google Scholar] [CrossRef]
  94. Ali, Y.R.H. Ethical impact of the technology on the healthcare system. J. Clin. Investig. Stud. 2020, 3, 1–2. [Google Scholar]
  95. Ross, S.; Weijer, C.; Gafni, A.; Ducey, A.; Thompson, C.; LaFreniere, R. Ethics, economics and the regulation and adoption of new medical devices: Case studies in pelvic floor surgery. BMC Med. Ethic 2010, 11, 14. [Google Scholar] [CrossRef] [Green Version]
  96. Cardenas, D. Surgical ethics: A framework for surgeons, patients, and society. Rev. Colégio Bras. Cirurg. 2020, 47, e20202519. [Google Scholar] [CrossRef]
  97. Bengio, Y.; Lecun, Y.; Hinton, G. Turing Lecture 2018: Deep Learning for AI. Commun. ACM 2021, 64, 58–65. [Google Scholar] [CrossRef]
  98. Wagner, M.; Bihlmaier, A.; Kenngott, H.G.; Mietkowski, P.; Scheikl, P.M.; Bodenstedt, S.; Schiepe-Tiska, A.; Vetter, J.; Nickel, F.; Speidel, S.; et al. A learning robot for cognitive camera control in minimally invasive surgery. Surg. Endosc. 2021, 35, 5365–5374. [Google Scholar] [CrossRef] [PubMed]
  99. Lao, J.; Chen, Y.; Li, Z.-C.; Li, Q.; Zhang, J.; Liu, J.; Zhai, G. A Deep Learning-Based Radiomics Model for Prediction of Survival in Glioblastoma Multiforme. Sci. Rep. 2017, 7, 1–8. [Google Scholar] [CrossRef] [PubMed]
  100. Rodríguez-Ruiz, A.; Lång, K.; Gubern-Merida, A.; Broeders, M.; Gennaro, G.; Clauser, P.; Helbich, T.H.; Chevalier, M.; Tan, T.; Mertelmeier, T.; et al. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison With 101 Radiologists. J. Natl. Cancer Inst. 2019, 111, 916–922. [Google Scholar] [CrossRef] [PubMed]
  101. Degrave, J.; Hermans, M.; Dambre, J.; Wyffels, F. A Differentiable Physics Engine for Deep Learning in Robotics. Front. Neurorobot. 2019, 13, 6. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  102. Bauernschmitt, R.; Schirmbeck, E.U.; Knoll, A.; Mayer, H.; Nagy, I.; Wessel, N.; Wildhirt, S.M.; Lange, R. Towards robotic heart surgery: Introduction of autonomous procedures into an experimental surgical telemanipulator system. Int. J. Med. Robot. Comput. Assist. Surg. 2005, 1, 74–79. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  103. Treat, M.R.; Amory, S.E.; Downey, P.E.; Taliaferro, D.A. Initial clinical experience with a partly autonomous robotic surgical instrument server. Surg. Endosc. 2006, 20, 1310–1314. [Google Scholar] [CrossRef]
  104. Bethea, B.T.; Okamura, A.; Kitagawa, M.; Fitton, T.P.; Cattaneo, S.M.; Gott, V.L.; Baumgartner, W.A.; Yuh, D.D. Application of Haptic Feedback to Robotic Surgery. J. Laparoendosc. Adv. Surg. Tech. 2004, 14, 191–195. [Google Scholar] [CrossRef]
  105. King, C.-H.; Culjat, M.O.; Franco, M.L.; Bisley, J.; Carman, G.P.; Dutson, E.P.; Grundfest, W.S. A Multielement Tactile Feedback System for Robot-Assisted Minimally Invasive Surgery. IEEE Trans. Haptics 2008, 2, 52–56. [Google Scholar] [CrossRef]
  106. Hong, M.B.; Jo, Y.H. Design and Evaluation of 2-DOF Compliant Forceps with Force-Sensing Capability for Minimally Invasive Robot Surgery. IEEE Trans. Robot. 2012, 28, 932–941. [Google Scholar] [CrossRef]
  107. Kim, U.; Lee, D.-H.; Yoon, W.J.; Hannaford, B.; Choi, H.R. Force Sensor Integrated Surgical Forceps for Minimally Invasive Robotic Surgery. IEEE Trans. Robot. 2015, 31, 1214–1224. [Google Scholar] [CrossRef]
  108. Qasaimeh, M.; Sokhanvar, S.; Dargahi, J.; Kahrizi, M. PVDF-Based Microfabricated Tactile Sensor for Minimally Invasive Surgery. J. Microelectromech. Syst. 2008, 18, 195–207. [Google Scholar] [CrossRef]
  109. Zhao, Y.; Guo, S.; Wang, Y.; Cui, J.; Ma, Y.; Zeng, Y.; Liu, X.; Jiang, Y.; Li, Y.; Shi, L.; et al. A CNN-based prototype method of unstructured surgical state perception and navigation for an endovascular surgery robot. Med. Biol. Eng. Comput. 2019, 57, 1875–1887. [Google Scholar] [CrossRef]
  110. Zheng, J.-Q.; Zhou, X.-Y.; Riga, C.; Yang, G.-Z. Towards 3D Path Planning from a Single 2D Fluoroscopic Image for Robot Assisted Fenestrated Endovascular Aortic Repair. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 8747–8753. [Google Scholar]
  111. Li, Z.; Dankelman, J.; De Momi, E. Path planning for endovascular catheterization under curvature constraints via two-phase searching approach. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 619–627. [Google Scholar] [CrossRef]
  112. Mahmoodian, N.; Schaufler, A.; Pashazadeh, A.; Boese, A.; Friebe, M.; Illanes, A. Proximal detection of guide wire perforation using feature extraction from bispectral audio signal analysis combined with machine learning. Comput. Biol. Med. 2019, 107, 10–17. [Google Scholar] [CrossRef] [PubMed]
  113. Chen, C.-H.; Sühn, T.; Illanes, A.; Maldonado, I.; Ahmad, H.; Wex, C.; Croner, R.; Boese, A.; Friebe, M. Proximally placed signal acquisition sensoric for robotic tissue tool interactions. Curr. Dir. Biomed. Eng. 2018, 4, 67–70. [Google Scholar] [CrossRef]
  114. Illanes, A.; Boese, A.; Maldonado, I.; Pashazadeh, A.; Schaufler, A.; Navab, N.; Friebe, M. Novel clinical device tracking and tissue event characterization using proximally placed audio signal acquisition and processing. Sci. Rep. 2018, 8, 1–11. [Google Scholar] [CrossRef]
  115. Chen, C.; Sühn, T.; Kalmar, M.; Maldonado, I.; Wex, C.; Croner, R.; Boese, A.; Friebe, M.; Illanes, A. Texture differentiation using audio signal analysis with robotic interventional instruments. Comput. Biol. Med. 2019, 112, 103370. [Google Scholar] [CrossRef]
  116. Illanes, A.; Schaufler, A.; Sühn, T.; Boese, A.; Croner, R.; Friebe, M. Surgical audio information as base for haptic feedback in robotic-assisted procedures. Curr. Dir. Biomed. Eng. 2020, 6, 1–5. [Google Scholar] [CrossRef]
  117. Schaufler, A.; Illanes, A.; Maldonado, I.; Boese, A.; Croner, R.; Friebe, M. Surgical Audio Guidance: Feasibility Check for Robotic Surgery Procedures. Curr. Dir. Biomed. Eng. 2020, 6, 571–574. [Google Scholar] [CrossRef]
  118. Schaufler, A.; Sühn, T.; Esmaeili, N.; Boese, A.; Wex, C.; Croner, R.; Friebe, M.; Illanes, A. Automatic differentiation between Veress needle events in laparoscopic access using proximally attached audio signal characterization. Curr. Dir. Biomed. Eng. 2019, 5, 369–371. [Google Scholar] [CrossRef]
  119. Sühn, T.; Pandey, A.; Friebe, M.; Illanes, A.; Boese, A.; Lohman, C. Acoustic sensing of tissue-tool interactions—potential applications in arthroscopic surgery. Curr. Dir. Biomed. Eng. 2020, 6, 595–598. [Google Scholar] [CrossRef]
  120. Marcus, H.J.; Payne, C.J.; Hughes-Hallett, A.; Gras, G.; Leibrandt, K.; Nandi, D.; Yang, G.Z. Making the Leap: The Translation of Innovative Surgical Devices from the Laboratory to the Operating Room. Ann. Surg. 2016, 263, 1077–1078. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Artificial intelligence surgery Gartner hype cycle: natural orifice transluminal endoscopic surgery (NOTES), single incision laparoscopic surgery (SILS), machine learning (ML), deep learning (DL), natural language processing, computer vision (CV). Adapted from the Gartner Hype Cycle for Artificial Intelligence, 2019 gartner.com/smarterwithgartner and modifications by Oostehoff et al. Adapted from Ref. [1].
Figure 1. Artificial intelligence surgery Gartner hype cycle: natural orifice transluminal endoscopic surgery (NOTES), single incision laparoscopic surgery (SILS), machine learning (ML), deep learning (DL), natural language processing, computer vision (CV). Adapted from the Gartner Hype Cycle for Artificial Intelligence, 2019 gartner.com/smarterwithgartner and modifications by Oostehoff et al. Adapted from Ref. [1].
Sensors 21 05526 g001
Figure 2. Traditional CV methods (top) vs. deep learning approach (bottom).
Figure 2. Traditional CV methods (top) vs. deep learning approach (bottom).
Sensors 21 05526 g002
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gumbs, A.A.; Frigerio, I.; Spolverato, G.; Croner, R.; Illanes, A.; Chouillard, E.; Elyan, E. Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery? Sensors 2021, 21, 5526. https://doi.org/10.3390/s21165526

AMA Style

Gumbs AA, Frigerio I, Spolverato G, Croner R, Illanes A, Chouillard E, Elyan E. Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery? Sensors. 2021; 21(16):5526. https://doi.org/10.3390/s21165526

Chicago/Turabian Style

Gumbs, Andrew A., Isabella Frigerio, Gaya Spolverato, Roland Croner, Alfredo Illanes, Elie Chouillard, and Eyad Elyan. 2021. "Artificial Intelligence Surgery: How Do We Get to Autonomous Actions in Surgery?" Sensors 21, no. 16: 5526. https://doi.org/10.3390/s21165526

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop