Next Article in Journal
Generalized Spoof Detection and Incremental Algorithm Recognition for Voice Spoofing
Previous Article in Journal
Multi-Objective Collaborative Control Method for Multi-Axle Distributed Vehicle Assisted Driving
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Coordinate-Regression-Based Deep Learning Model for Catheter Detection during Structural Heart Interventions

1
Dalio Institute of Cardiovascular Imaging, NewYork-Presbyterian Hospital and Weill Cornell Medicine, New York, NY 10021, USA
2
Department of Radiology, Weill Cornell Medicine, New York, NY 10021, USA
3
AWS, Amazon, Seattle, WA 98170, USA
4
School of Electrical and Computer Engineering, Cornell University, Ithaca, NY 10021, USA
5
Division of Cardiology, Department of Medicine, Weill Cornell Medicine, New York, NY 10021, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7778; https://doi.org/10.3390/app13137778
Submission received: 9 May 2023 / Revised: 8 June 2023 / Accepted: 19 June 2023 / Published: 30 June 2023

Abstract

Featured Application

Assisting surgeons by automatically detecting the position of the catheter tip based on fluoroscopic images during minimally invasive surgeries.

Abstract

With a growing geriatric population estimated to triple by 2050, minimally invasive procedures that are image-guided are becoming both more popular and necessary for treating a variety of diseases. To lower the learning curve for new procedures, it is necessary to develop better guidance systems and methods to analyze procedure performance. Since fluoroscopy remains the primary mode of visualizations, the ability to perform catheter tracking from fluoroscopic images is an important part of this endeavor. This paper explores the use of deep learning to perform the landmark detection of a catheter from fluoroscopic images in 3D-printed heart models. We show that a two-stage deep-convolutional-neural-network-based model architecture can provide improved performance by initially locating a region of interest before determining the coordinates of the catheter tip within the image. This model has an average error of less than 2% of the image resolution and can be performed within 4 milliseconds, allowing for its potential use for real-time intraprocedural tracking. Coordinate regression models have the advantage of directly outputting values that can be used for quantitative tracking in future applications and are easier to create ground truth values (~50× faster), as compared to semantic segmentation models that require entire masks to be made. Therefore, we believe this work has better long-term potential to be used for a broader class of cardiac devices, catheters, and guidewires.

1. Introduction

With a growing geriatric population estimated to triple by 2050, minimally invasive procedures that are image-guided are becoming both more popular and necessary for treating a variety of diseases [1,2,3]. Currently, over 400 million individuals worldwide are suffering from cardiovascular disease [4], and the number of annual deaths is estimated to be 23.3 million by 2030 [5]. Of these, well over a million patients will undergo a minimally invasive percutaneous procedure to address a structural heart defect, which is defined as an abnormality of the cardiac wall, valve, chamber, or one of the major arteries of the heart [6]. Despite being minimally invasive, these procedures can still result in death or life-threatening complications due to accidental punctures and device embolization, or can have limited efficacy due to the misalignment of an implanted device [7]. Percutaneous and minimally invasive procedures are used for a wide variety of cardiovascular ailments and are typically guided by imaging modalities, such as X-ray fluoroscopy and echocardiography [8], since they provide real-time imaging [9]. However, most organs are transparent to fluoroscopy, so contrast agents, which transiently opacify structures of interest, must be used to visualize the surrounding tissue. Furthermore, fluoroscopy only provides a two-dimensional (2D) projection of the catheter and device, and, therefore, no information on their depth [10]. Echocardiography can directly image anatomic structures and blood flow, so it is often used as a complementary imaging modality. Although this method can provide useful intraoperative 2D and 3D images, this technology requires skilled operators and general anesthesia is a prerequisite for transesophageal echocardiographic (TEE) guidance. The limitations of these techniques increase the complexity of procedures, which often require the interventionalist to determine the position of the catheter/device by analyzing multiple imaging angles and modalities [11,12]. The added coordination of different specialties also increases resource utilization and costs. Preoperative 3D imaging modalities provide detailed anatomic information and are often displayed on separate screens or overlaid on real-time imaging modalities to improve image-guided interventions [9,12,13]. However, this method of fusion imaging obstructs the view of the real-time image during the procedure [12]. Furthermore, all of these images are displayed on 2D screens, which fundamentally mitigate the ability to perceive depth and orientation [14].
Our group [15,16,17], among others [18,19,20,21], have shown the promise of augmented and mixed-reality visualization to address the limitations in depth perception for cardiac interventions by using 3D holographic headsets. In order to leverage mixed-reality technology as a navigation tool for fluoroscopy-guided percutaneous procedures, catheter tracking is a cornerstone of the image guidance system. In our earlier publications, we demonstrated a mixed-reality-based training simulator for structural heart disease interventions using a 3D-printed heart phantom [15]. In this work, the catheter was coupled, at three distinct locations along its distal end, with three electromagnetic (EM) sensors, each allowing the real-time simultaneous tracking of three spatial positions (X, Y, and Z) and three orientation angles (azimuth, elevation, and roll). Although utilizing EM sensors is advantageous for portability and to minimize radiation from fluoroscopy, affordable systems (<USD 10 k) have a low accuracy (up to ~5 mm) and require manual integration of sensors into a catheter, and thus are not a general solution to address the many types of cardiac interventional devices available on the market.
To address these limitations with EM sensor tracking, we previously presented a novel deep-learning-driven method for tracking a catheter in a 3D-printed heart phantom from biplane fluoroscopic images that were acquired during a mock procedure [17]. In this work, we trained a U-Net [22] model on the 3D-printed heart phantom to segment a radiopaque marker at the tip of the catheter. A postprocessing step was used to analytically calculate the Z-coordinate by leveraging the two simultaneous views of the catheter. Although this was an accurate method to perform the tip detection of a catheter, it is not scalable to many types of catheters and/or devices, since semantic segmentation models require the time-consuming manual annotation of ground truth images. Annotating masks (i.e., selecting all pixels in the catheter tip) takes ~50 times longer than defining the landmark as a single point on the image. We, therefore, explored the use of coordinate regression models that directly output the X- and Y-coordinate of the catheter tip, while utilizing the same analytical Z-coordinate calculation. This method has the benefits of simpler ground truth annotations and mitigating the need for the postprocessing of the output mask.
Deep convolutional neural networks (CNNs) have been applied to a multitude of computer vision problems. In medical imaging, CNNs have been effective in many tasks, such as classification [23], localization [24], tracking [25], and image segmentation [22,26]. Ronneberger et al. [22] proposed the U-Net [22], which replaced the pooling operators in fully convolutional networks (FCNs) [27] with up-sampling operators, allowing the input image resolution to be retained in the output. The performance of the U-Net [22] for segmenting medical images, notably with even small training datasets, demonstrated the potential of such an encoder–decoder architecture. The U-Net [22] was later applied to other medical settings, including the Xenopus kidney [28], MRI volume segmentation of the prostate [29], retinal vessels, liver and tumors in CT scans, ischemic stroke lesion, and intervertebral disc and pancreas [30,31,32,33,34,35,36,37,38,39,40,41].
Landmark localization plays a vital role in medical image analysis. Several deep learning methods have been proposed for landmark localization that employ regression or classification [42,43,44,45,46,47,48]. In regression-based localization, a hard threshold is utilized to detect the presence of a landmark in image slices, patches, or voxels. This threshold converts the model’s coordinate prediction (pixel-wise or mm-based) to the assigned annotated labels. Therefore, these methods usually rely on the careful consideration of a final threshold value, which may be data- or task-specific. Zheng et al. [49] localized a landmark by classifying image voxels with multilayer perceptrons, while Xu et al. [50] localized landmarks based on their relative position (up, down, left, or right) to the landmark of interest. In another work, Yang et al. [51] predicted the location of a landmark based on intersecting the classification outputs from all axial, coronal, and sagittal image slices. Another existing approach for coordinate regression is to generate a heatmap output from a CNN and apply a loss directly to the heatmap rather than the numerical coordinates [52]. Even though this approach offers good spatial generalization, it has the drawback that gradient flow begins at the heatmap rather than at the numerical coordinates. This creates a disconnect between the heatmap loss optimization and the true goal of reducing the coordinate error distances [52]. Another coordinate regression approach is to utilize fully connected (FC) layers that produce numerical coordinates [53]. Most notably, FC layers allow end-to-end backpropagation from the predicted numerical coordinates to the input image. However, the fully connected layer weights are highly dependent on the spatial distribution of the inputs during training. To address this issue, our group used another approach that leveraged a segmentation network followed by a postprocessing step [17]. Despite providing promising results, the time-consuming process of making masks for the ground truth labels prevented the long-term scalability of this work for other devices and expanding the dataset to allow for greater generalizability (multiple scanners, background images from patients, etc.). Thus, we propose pursuing a direct identification of the catheter tip without any masking annotation/analysis.
There has also been some work involving coordinate regression in medical imaging in recent years, including localizing in areas related to neurology [54] and arthrology [55] sciences. Dünnwald et al. conducted segmentation and localization on the locus coeruleus (LC), a small nucleus in the brain stem. They proposed CoRe-Unet, a type of 3D U-Net to predict coordinates of a voxel in the input volume. The network relies on correlating prominent structures to detect the actual position of the LC through a localization network trained to detect the center-of-mass (COM) coordinates [56]. Li et al. [55] also focused on landmark detection in the temporomandibular joints (TJ) through a two-stage end-to-end localization network based on an attention-guided mechanism. Their method includes global and local stages, based on learning both local features around landmarks and estimating landmark coordinates. The network consists of a differentiable spatial to numerical transform (DSNT) layer attached to a 3D U-Net, enabling the conversion of heatmaps to coordinate detection [55]. Despite some achievements in the mentioned line of work, models based on heatmap analysis have two main drawbacks: they are computationally expensive and are sensitive to outliers, whilst coordinate regression provides a faster estimation of the landmark [57].
There has been work involving guidewire tracking, usually used in combination with catheter tracking in related studies. Researchers indicate that many frameworks used for catheter tracking can also be applied to guidewire tracking [58]. Traditional methods in endovascular and neurosurgery fields have guided researchers to use either intensity- and learning-based models or the segmentation and movement of the guidewire [58,59,60,61,62]. However, with emergent segmentation and deep learning techniques, new approaches have taken a turn in this direction. These techniques range from supervised to unsupervised methods. Supervised methods can consist of instance segmentation, region of interest [63,64,65,66], two-stage region of interest, and target segmentation [67], whilst unsupervised methods have used optical flow and a U-Net trained in a Siamese network [68]. Although all these methods are relevant to the field, they include semantic segmentation, which falls out of the scope of this paper.
In another work, researchers utilized a CNN to explore the possibility to detect motion between two fluoroscopic frames in catheterization procedures [69]. They were able to compare their CNN-based catheter tip detection with normalized cross correlation (CC) and found a mean absolute error (MAE) of 8.7 ± 2.5 pixels or 3.0 ± 0.9 mm between methods, with the CNN outperforming CC. However, the researchers state that the correlation between the predictions and tracking results is not obvious. In another study, automated catheter localization for ultrasound-guided high-dose-rate prostate brachytherapy was pursued. They used a U-NET model to localize implanted catheters on transverse images on 3500 manually localized implanted catheters. They reported 80% reconstruction accuracy within 2 mm along 90% of the catheter length; however, they also mentioned that the catheter tip was often not detected and required extrapolation [70].
To overcome the issues in the previously pursued works, this paper presents the implementation of a landmark localization method using coordinate regression for catheter detection from fluoroscopic images acquired during a mock procedure in a 3D-printed heart phantom.
We can summarize the motivation of the paper as follows:
  • Tracking the tip of a catheter for future use in a mixed-reality navigation system.
  • Addressing the limited accuracy, low availability, and high cost of EM sensor tracking systems.
  • Proposing a catheter tip coordinate regression detection methodology leveraging deep convolutional neural networks to reduce the time-consuming task of generating ground truth masks.

2. Materials and Methods

2.1. Dataset

We collected a fluoroscopic image dataset for our custom-made, 3D-printed heart model during a mock procedure in the catheterization lab at NewYork-Presbyterian Hospital. The custom model contained a heart that was transparent under X-ray and a metal spray-painted spine that was visible under X-ray, which is the same as in our previous publication [15]. In addition, some metal spheres were present as fiducial markers, but those were not used in this study. It should be noted that the 3D-printed heart model was within an acrylic box that produced additional artifacts seen in the image, but these artifacts will not be present in clinical images. Our dataset contained a total of 3408 JPEG images at 512 × 512 spatial resolution, where the catheter was moved along the entire range of the image. In all images, the entire catheter tip was visible. The dataset included 62 paired images which were taken from different views at the same time. These 124 images were used to test the model in 3D space. Image contrast, clarity, and brightness varied across sets taken on different days.
To prove that landmark detection can be achieved, we initially focused solely on the radiopaque marker band of the catheter tip. The marker band formed different shapes depending on the angle of the image, and, therefore, can be seen as a rectangle with varying width, a thin-walled circle, or an oval. The catheter can also take on different shapes in the image as it is curved, rotated, and translated by the user (Figure 1).
The dataset (containing 3408 images) was randomly divided into three sets: training (2182 images), validation (546 images), and test (680 images). In addition, we had an additional set, the “fixed-test” set, which was based on removing any incorrectly predicted regions from the first stage of the 2-stage network architecture (to determine the best base accuracy of the second stage of the network). For deep learning model evaluation, 20% of the samples were set aside at the beginning as the test set. The training and validation sets consisted of the remaining 80% of the samples, which were trained and evaluated based on a 5-fold cross-validation technique. Therefore, in each iteration, 80% of the training and validation samples were dedicated to training the model and 20% of them were set as the validation set. The best parameters for the training process were chosen based on best practices and the assessment of the validation outcomes. The training parameters included the number of layers, neurons in each layer, and epochs, and the learning rate. The best values for these parameters were achieved by varying them and checking for the best accuracy in the outcome. The detailed architecture is discussed in Section B.

2.2. Architectures

The CNN-based models developed for this problem are intended to locate the center of the marker band in less time than the frame rate of the acquisition system (generally 15 frames per second). Since it is understood that landmark detection will be more accurate on detailed images, we implemented two deep networks working in series. The first network, called the region selection network, predicts a subregion of the image that contains the marker band; a second network, called the localizer network, finds the exact position of the landmark (outputting its X- and Y-coordinates). Although the two networks were used sequentially during inference, training was performed separately. This way, the image was used in more detail without using the larger field of view of the entire image. The details of these networks are explained in the following two sections.

2.2.1. Region Selection Network

The region selection network seeks to detect the region of the image that contains the marker band of the catheter. To select the target region, the images are first divided into n columns and n rows. Since the model is supervised, it needs the ground truth positions of the targets in the new splitting system. Therefore, the target positions are converted to the region number in which it resides. Thus, the model output is an n × n vector with a softmax activation, with each element representing the probability of the target point being in that region.
The network consists of a VGG-16 [71] backbone followed by three convolution layers and three fully connected layers. The first convolution layer consists of a 3 × 3 kernel with 1024 channels, followed by a rectified linear unit (ReLU) activation function. The second layer includes a 3 × 3 kernel, 512 channels, and a ReLU activation function. The final layer consists of a 1 × 1 kernel, 9 channels, and a sigmoid activation function (Figure 2). Afterward, three fully connected (FC) layers are implemented. These layers have 128, 16, and n × n (representing the total region number) neurons. The FC layers have a dropout probability of 0.20 and 0.05 as shown in Figure 2.
The input images were originally 512 × 512 pixels. However, as the input to a VGG network [71] is an image with dimensions 224 × 224 × 3, we initially resized our images to be in accordance with the VGG dimensions. For transfer learning purposes, the weights of all layers were frozen with weights pretrained on ImageNet [72], whilst the remaining layers of the network were fine-tuned on our dataset. Training was performed with 80 epochs per 5-fold split (where the highest accuracy was achieved), with the longest training occurring with 400 epochs. The Adam optimizer was used with a 0.00001 constant learning rate, and we used a cross-entropy loss function.

2.2.2. Localizer Network

The localizer network finds the best point in the image for the center of the marker band in the selected region. In this network (Figure 3), the input is a region of the original image that was selected in the first network, albeit it is the ground truth region in the training procedure. These subimages are loaded with a higher resolution to match the needed size of 224 × 224 pixels. The network’s output is the X- and Y-coordinates of this subimage (i.e., between 0 and 223). Since the landmark location is in the local coordinate system of the subimage, it is then converted to the global coordinates of the original image. The architecture of this network is similar to the previous network apart from the last layer, which now has two FC layers with 9 and 2 neurons, respectively. Here, training is performed with 100 epochs per 5-fold split (with the lowest radial distance error as cost function), with the longest training occurring with 500 epochs. The network’s output provides a prediction based on the last two neurons, which represent the X- and Y-coordinates of the landmark in the local coordinate system of the subimage.

2.3. Dual Network Inference

During inference, the two networks work in series (Figure 4), where the first network receives the images sequentially with a size of 224 × 244 pixels and designates the region in which the landmark has the highest probability of existing. It should be noted that if this first network predicts incorrectly, the second network will not be able to select the correct catheter tip. The selected region’s resolution is updated to match the VGG-16 network architecture and entered as the 224 × 224 × 3 input to the network. This provides a connection between the localizer network and the predicted landmark’s coordinates of the region. A transformation occurs to convert the network output to the local landmark coordinates (marked as the center of the marker and indicated with a red arrow) and, consequently, the global coordinates of the original image.
Both networks’ designs were implemented in Google Colab utilizing the Python programming language. Google hosted Colab for the artificial intelligent applications with many inbuilt libraries and free GPU and TPU accelerators. Concerning hardware acceleration, we ran the model on Colab with a GPU which proved to have the highest efficiency. Colab also offers TPU computing; however, the TPU takes up a considerably more training time compared to the GPU mode typically in small batch sizes.

3. Results and Discussion

The overall contribution of this work is showing how a two-stage architecture can improve the accuracy of landmark detection for a coordinate regression network. Figure 5 shows the distribution of errors that occur for all input images (indicated by the number of samples in the vertical axis) for when only a single region (i.e., the entire image) is fed into the second stage versus nine regions. As can be seen, the results for nine regions had a distribution that was shifted towards the left, indicating a higher accuracy as compared to the distribution of one region, which was further shown by their average accuracy of 1.75 pixels and 7.36 pixels, respectively. Therefore, utilizing the two-stage architecture is necessary to optimize the accuracy of this landmark detection.
To further optimize the performance of this two-stage architecture, we varied the size of the square region array from n = 3 (9 regions) to n = 10 (100 regions). Figure 6 shows the accuracy of the region selection network for their training, validation, and test sets, for each of the characterized region sizes. Training and validation sets are the last sets provided by the 5-fold validation. As can be seen, with an increase in the number of regions, the accuracy of detection for the first stage decreased. The incorrect predictions were usually occurring when the marker band was located at the border of the region, such that it was overlapping with another region (Figure 7). In this example of the figure, the model was predicting from 25 possible regions, and predicted correctly in the figure on the left (Figure 7A) but incorrectly in the figure on the right (Figure 7B). To state a simple example, in the case of 100 regions, the model misclassified 11 (out of 680) images. In brief, almost all the predicted regions touched the catheter tip, even if they did not point to the correct region.
In the localizer network, the expectation is to predict the exact location of the center of the marker band within the target region. Figure 8 presents the outcomes obtained from the network. This chart indicates the averages of errors between the ground truth coordinate and the predicted landmark coordinate by the model. This error is reported separately for the training, validation, test, and fixed-test results. The “fixed-test” results remove all the incorrectly selected regions to demonstrate the highest accuracy the second model can achieve. The error is reported based on the 512 × 512 pixel size versions of the original images. The test dataset contained the average of the results of both the correctly and incorrectly selected regions.
Overall, the data convey two overall trends: (i) the larger the region number, the less accurately the region selection performs, and (ii) the larger the region number, the more accurately the localizer network performs. Given these two trends, it is seen that either n = 5 or n = 7 provides the optimal results, with n = 7 providing the highest accuracy for the fixed-test set. These conclusions are further supported by the plots in Figure A1, which show the distribution of error for all points; it can be seen that the majority of images were accurately predicted, with a minimal number of large outliers. As we plotted these errors with their ground truth coordinates of the image (Figure A2), we did not see any obvious trends due to spatial positioning within the image.
The inference time to predict landmark position in this model (both networks consecutively) is 4 milliseconds on average. The frame rate of the acquisition system is typically 15 frames per second (67 ms). Therefore, the inference time is significantly less than the system’s acquisition rate and this method can be used for real-time catheter tracking. Compared to other well-known networks, the results of average accuracy, average training time, and average inference time of some models are shown in Table 1. Our method, which is the two-stage and VGG-based model, outperforms all the mentioned models.

4. Conclusions

We have shown that landmark detection using coordinate regression deep learning models can be used to perform catheter localizing on fluoroscopy images. Our results suggest that these models can provide ~1% positional accuracy at speeds much faster (>10×) than commercial acquisition systems used in cardiac interventions, and have the fastest training and inference times compared to other models. Furthermore, our method requires less time in preparing ground truth training datasets as compared to semantic segmentation methods (as typically used for U-Net). Although the results are promising, there are several limitations that need future improvement before these models can be used for clinical applications: (i) Nonclinical images: These models were trained on images acquired from 3D-printed models and, thus, do not have clinically relevant backgrounds; thus, clinical use should only be performed after the model has been optimized and validated on clinical images. (ii) Limited degrees of freedom: Currently, the model only predicts a single point on the catheter, giving a single 3D coordinate but no orientation; future models should predict two or more points to allow for the catheter’s 3D orientation to be calculated. (iii) Generalizability: This model was trained on a set of images acquired from a single type of catheter, imaging system, and phantom model; future work needs to train the model on an expanded set of data. (iv) Accelerating and parallel processing must be considered in future works.
Despite the limitations listed above, this work provides the primary advantage of developing deep learning models that can be optimized without the tedious task of developing image masks, as typically needed by U-Net models [22]. Furthermore, these models can be used to perform catheter tracking for training purposes for custom 3D-printed heart models [15] or on commercial training systems (Biomodex Inc., Paris, France).

Author Contributions

M.A.: Designed and implemented the DL model, programming code; A.K.: wrote the paper, statistical analysis; S.H.: programming code, designed the DL model; M.T. and A.C.: developed the control U-Net model results; Y.X. and H.R.: designed and implemented the DL model; S.C.W. and M.R.: acquired the ground truth images; T.M., B.L., A.W. and M.S.: statistical analysis, manuscript revision; B.M.: project overview, wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets and models used and/or analyzed during the current study are available on Github from the following link: https://github.com/mosadeghlabwcm/2023-Publication_Coordinate-Regression-Tracking, accessed on 16 February 2023.

Acknowledgments

We thank the Dalio Institute of Cardiovascular Imaging for their continued support.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Histogram of error for each image in the test set for all evaluated region numbers.
Figure A1. Histogram of error for each image in the test set for all evaluated region numbers.
Applsci 13 07778 g0a1
Figure A2. Magnitude of errors and the predicted location of the marker band for all region numbers.
Figure A2. Magnitude of errors and the predicted location of the marker band for all region numbers.
Applsci 13 07778 g0a2

References

  1. Schmitto, J.D.; Mokashi, S.A.; Cohn, L.H. Minimally-invasive valve surgery. J. Am. Coll. Cardiol. 2010, 56, 455–462. [Google Scholar] [CrossRef]
  2. Endo, Y.; Nakamura, Y.; Kuroda, M.; Ito, Y.; Hori, T. The Utility of a 3D Endoscope and Robot-Assisted System for MIDCAB. Ann. Thorac. Cardiovasc. Surg. 2019, 25, 200–204. [Google Scholar] [CrossRef]
  3. Dieberg, G.; Smart, N.A.; King, N. Minimally invasive cardiac surgery: A systematic review and meta-analysis. Int. J. Cardiol. 2016, 223, 554–560. [Google Scholar] [CrossRef]
  4. Jiang, Z.; Qu, H.; Zhang, Y.; Zhang, F.; Xiao, W.; Shi, D.; Gao, Z.; Chen, K. Efficacy and Safety of Xinyue Capsule for Coronary Artery Disease after Percutaneous Coronary Intervention: A Systematic Review and Meta-Analysis of Randomized Clinical Trials. Evid. Based. Complement. Altern. Med. 2021, 2021, 6695868. [Google Scholar] [CrossRef]
  5. Bansilal, S.; Castellano, J.M.; Fuster, V. Global burden of CVD: Focus on secondary prevention of cardiovascular disease. Int. J. Cardiol. 2015, 201 (Suppl. S1), S1–S7. [Google Scholar] [CrossRef]
  6. Little, S.H. Structural Heart Interventions. Methodist Debakey Cardiovasc. J. 2017, 13, 96–97. [Google Scholar] [CrossRef]
  7. Wasmer, K.; Zellerhoff, S.; Kobe, J.; Monnig, G.; Pott, C.; Dechering, D.G.; Lange, P.S.; Frommeyer, G.; Eckardt, L. Incidence and management of inadvertent puncture and sheath placement in the aorta during attempted transseptal puncture. Europace 2017, 19, 447–457. [Google Scholar] [CrossRef] [PubMed]
  8. Faletra, F.F.; Pedrazzini, G.; Pasotti, E.; Moccetti, T. Side-by-side comparison of fluoroscopy, 2D and 3D TEE during percutaneous edge-to-edge mitral valve repair. JACC Cardiovasc. Imaging 2012, 5, 656–661. [Google Scholar] [PubMed]
  9. Arujuna, A.V.; Housden, R.J.; Ma, Y.; Rajani, R.; Gao, G.; Nijhof, N.; Cathier, P.; Bullens, R.; Gijsbers, G.; Parish, V. Novel system for real-time integration of 3-D echocardiography and fluoroscopy for image-guided cardiac interventions: Preclinical validation and clinical feasibility evaluation. IEEE J. Transl. Eng. Health Med. 2014, 2, 1–10. [Google Scholar]
  10. Sra, J.; Krum, D.; Choudhuri, I.; Belanger, B.; Palma, M.; Brodnick, D.; Rowe, D.B. Identifying the third dimension in 2D fluoroscopy to create 3D cardiac maps. JCI Insight 2016, 1, e90453. [Google Scholar] [CrossRef] [PubMed]
  11. Celi, S.; Martini, N.; Emilio Pastormerlo, L.; Positano, V.; Berti, S. Multimodality imaging for interventional cardiology. Curr. Pharm. Design 2017, 23, 3285–3300. [Google Scholar]
  12. Biaggi, P.; Fernandez-Golfín, C.; Hahn, R.; Corti, R. Hybrid imaging during transcatheter structural heart interventions. Curr. Cardiovasc. Imaging Rep. 2015, 8, 1–14. [Google Scholar]
  13. Falk, V.; Mourgues, F.; Adhami, L.; Jacobs, S.; Thiele, H.; Nitzsche, S.; Mohr, F.W.; Coste-Manière, È. Cardio navigation: Planning, simulation, and augmented reality in robotic assisted endoscopic bypass grafting. Ann. Thorac. Surg. 2005, 79, 2040–2047. [Google Scholar] [PubMed]
  14. Muraru, D.; Badano, L.P. Physical and technical aspects and overview of 3D-echocardiography. In Manual of 3D Echocardiography; Springer: Berlin/Heidelberg, Germany, 2017; pp. 1–44. [Google Scholar]
  15. Jang, S.-J.; Torabinia, M.; Dhrif, H.; Caprio, A.; Liu, J.; Wong, S.C.; Mosadegh, B. Development of a Hybrid Training Simulator for Structural Heart Disease Interventions. Adv. Intell. Syst. 2020, 2, 2000109. [Google Scholar] [CrossRef]
  16. Liu, J.; Al’Aref, S.J.; Singh, G.; Caprio, A.; Moghadam, A.A.A.; Jang, S.-J.; Wong, S.C.; Min, J.K.; Dunham, S.; Mosadegh, B. An augmented reality system for image guidance of transcatheter procedures for structural heart disease. PLoS ONE 2019, 14, e0219174. [Google Scholar] [CrossRef]
  17. Torabinia, M.; Caprio, A.; Jang, S.-J.; Ma, T.; Tran, H.; Mekki, L.; Chen, I.; Sabuncu, M.; Wong, S.C.; Mosadegh, B. Deep learning-driven catheter tracking from bi-plane X-ray fluoroscopy of 3D printed heart phantoms. Mini-Invasive Surg. 2021, 5, 32. [Google Scholar] [CrossRef]
  18. Southworth, M.K.; Silva, J.R.; Silva, J.N.A. Use of extended realities in cardiology. Trends Cardiovas. Med. 2020, 30, 143–148. [Google Scholar] [CrossRef]
  19. Jung, C.; Wolff, G.; Wernly, B.; Bruno, R.R.; Franz, M.; Schulze, P.C.; Silva, J.N.A.; Silva, J.R.; Bhatt, D.L.; Kelm, M. Virtual and Augmented Reality in Cardiovascular Care: State-of-the-Art and Future Perspectives. JACC Cardiovasc. Imaging 2022, 15, 519–532. [Google Scholar] [CrossRef]
  20. Kasprzak, J.D.; Pawlowski, J.; Peruga, J.Z.; Kaminski, J.; Lipiec, P. First-in-man experience with real-time holographic mixed reality display of three-dimensional echocardiography during structural intervention: Balloon mitral commissurotomy. Eur. Heart J. 2020, 41, 801. [Google Scholar] [CrossRef]
  21. Arjomandi Rad, A.; Vardanyan, R.; Thavarajasingam, S.G.; Zubarevich, A.; Van den Eynde, J.; Sa, M.; Zhigalov, K.; Sardiari Nia, P.; Ruhparwar, A.; Weymann, A. Extended, virtual and augmented reality in thoracic surgery: A systematic review. Interact. Cardiovasc. Thorac. Surg. 2022, 34, 201–211. [Google Scholar] [CrossRef] [PubMed]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015. [Google Scholar]
  23. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. (Eds.) Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014. [Google Scholar]
  24. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. arXiv 2013, arXiv:13126229. [Google Scholar]
  25. Chandan, G.; Jain, A.; Jain, H. Real time object detection and tracking using Deep Learning and OpenCV. In Proceedings of the 2018 International Conference on Inventive Research in Computing Applications (ICIRCA), Coimbatore, India, 11–12 July 2018. [Google Scholar]
  26. Rajchl, M.; Lee, M.C.; Oktay, O.; Kamnitsas, K.; Passerat-Palmbach, J.; Bai, W.; Damodaram, M.; Rutherford, M.A.; Hajnal, J.V.; Kainz, B. Deepcut: Object segmentation from bounding box annotations using convolutional neural networks. IEEE Trans. Med. Imaging 2016, 36, 674–683. [Google Scholar] [CrossRef] [PubMed]
  27. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  28. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  29. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016. [Google Scholar]
  30. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.-W.; Heng, P.-A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef]
  31. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  32. Zhang, J.; Jin, Y.; Xu, J.; Xu, X.; Zhang, Y. Mdu-net: Multi-scale densely connected u-net for biomedical image segmentation. arXiv 2018, arXiv:181200352. [Google Scholar] [CrossRef]
  33. Jin, Q.; Meng, Z.; Pham, T.D.; Chen, Q.; Wei, L.; Su, R. DUNet: A deformable network for retinal vessel segmentation. Knowl.-Based Syst. 2019, 178, 149–162. [Google Scholar]
  34. Jin, Q.; Meng, Z.; Sun, C.; Cui, H.; Su, R. RA-UNet: A hybrid deep attention-aware network to extract liver and tumor in CT scans. Front. Bioeng. Biotechnol. 2020, 8, 1471. [Google Scholar] [CrossRef]
  35. Dolz, J.; Ben Ayed, I.; Desrosiers, C. Dense multi-path U-Net for ischemic stroke lesion segmentation in multiple image modalities. In International MICCAI Brainlesion Workshop; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  36. Guo, J.; Deng, J.; Xue, N.; Zafeiriou, S. Stacked dense u-nets with dual transformers for robust face alignment. arXiv 2018, arXiv:181201936. [Google Scholar]
  37. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S. nnu-net: Self-adapting framework for u-net-based medical image segmentation. arXiv 2018, arXiv:180910486. [Google Scholar]
  38. Clèrigues, A.; Valverde, S.; Bernal, J.; Freixenet, J.; Oliver, A.; Lladó, X. Acute and sub-acute stroke lesion segmentation from multimodal MRI. Comput. Meth. Prog. Biol. 2020, 194, 105521. [Google Scholar] [CrossRef]
  39. Dolz, J.; Desrosiers, C.; Ben Ayed, I. IVD-Net: Intervertebral disc localization and segmentation in MRI with a multi-modal UNet. In International Workshop and Challenge on Computational Methods and Clinical Applications for Spine Imaging; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  40. Zhuang, J. LadderNet: Multi-path networks based on U-Net for medical image segmentation. arXiv 2018, arXiv:181007810. [Google Scholar]
  41. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:180403999. [Google Scholar]
  42. Urschler, M.; Ebner, T.; Štern, D. Integrating geometric configuration and appearance information into a unified framework for anatomical landmark localization. Med. Image Anal. 2018, 43, 23–36. [Google Scholar] [CrossRef] [PubMed]
  43. Xue, H.; Artico, J.; Fontana, M.; Moon, J.C.; Davies, R.H.; Kellman, P. Landmark Detection in Cardiac MRI by Using a Convolutional Neural Network. Radiol. Artif. Intell. 2021, 3, e200197. [Google Scholar] [CrossRef] [PubMed]
  44. Dabbah, M.A.; Murphy, S.; Pello, H.; Courbon, R.; Beveridge, E.; Wiseman, S.; Wyeth, D.; Poole, I. Detection and location of 127 anatomical landmarks in diverse CT datasets. Med. Imaging Image Process. 2014, 9034, 284–294. [Google Scholar] [CrossRef]
  45. Ibragimov, B.; Likar, B.; Pernus, F.; Vrtovec, T. Computerized Cephalometry by Game Theory with Shape-and Appearance-Based Landmark Refinement. In Proceedings of the International Symposium on Biomedical imaging (ISBI), Prague, Czech Republic, 13–16 April 2016. [Google Scholar]
  46. Zheng, Y.; John, M.; Liao, R.; Nottling, A.; Boese, J.; Kempfert, J.; Walther, T.; Brockmann, G.; Comaniciu, D. Automatic aorta segmentation and valve landmark detection in C-arm CT for transcatheter aortic valve implantation. IEEE Trans. Med. Imaging 2012, 31, 2307–2321. [Google Scholar] [CrossRef]
  47. Oktay, O.; Bai, W.; Guerrero, R.; Rajchl, M.; de Marvao, A.; O’Regan, D.P.; Cook, S.A.; Heinrich, M.P.; Glocker, B.; Rueckert, D. Stratified Decision Forests for Accurate Anatomical Landmark Localization in Cardiac Images. IEEE Trans. Med. Imaging 2017, 36, 332–342. [Google Scholar] [CrossRef]
  48. Rohr, K. Landmark-Based Image Analysis using Geometric and Intensity Models; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  49. Zheng, Y.F.; Liu, D.; Georgescu, B.; Nguyen, H.; Comaniciu, D. 3D Deep Learning for Efficient and Robust Landmark Detection in Volumetric Data. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Volume 9349, pp. 565–572. [Google Scholar] [CrossRef]
  50. Xu, Z.; Huang, Q.; Park, J.; Chen, M.; Xu, D.; Yang, D.; Liu, D.; Zhou, S.K. Supervised Action Classifier: Approaching Landmark Detection as Image Partitioning. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017. [Google Scholar]
  51. Yang, D.; Zhang, S.T.; Yan, Z.N.; Tan, C.W.; Li, K.; Metaxas, D. Automated Anatomical Landmark Detection on Distal Femur Surface Using Convolutional Neural Network. In Proceedings of the 2015 IEEE 12th International Symposium on Biomedical Imaging (Isbi), Brooklyn, NY, USA, 16–19 April 2015; pp. 17–21. [Google Scholar]
  52. Nibali, A.; He, Z.; Morgan, S.; Prendergast, L. Numerical coordinate regression with convolutional neural networks. arXiv 2018, arXiv:180107372. [Google Scholar]
  53. Toshev, A.; Szegedy, C. DeepPose: Human Pose Estimation via Deep Neural Networks. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  54. Chu, X.; Yang, W.; Ouyang, W.L.; Ma, C.; Yuille, A.L.; Wang, X.G. Multi-Context Attention for Human Pose Estimation. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (Cvpr 2017), Honolulu, HI, USA, 21–26 July 2017; pp. 5669–5678. [Google Scholar] [CrossRef]
  55. Li, J.; Wang, Y.; Mao, J.; Li, G.; Ma, R. (Eds.) End-to-End Coordinate Regression Model with Attention-Guided Mechanism for Landmark Localization in 3D Medical Images. International Workshop on Machine Learning in Medical Imaging; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  56. Dünnwald, M.; Betts, M.J.; Düzel, E.; Oeltze-Jafra, S. Localization of the Locus Coeruleus in MRI via Coordinate Regression. In Bildverarbeitung für die Medizin 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 10–15. [Google Scholar]
  57. Jin, H.; Liao, S.; Shao, L. Pixel-in-pixel net: Towards efficient facial landmark detection in the wild. Int. J. Comput. Vis. 2021, 129, 3174–3194. [Google Scholar] [CrossRef]
  58. Ramadani, A.; Bui, M.; Wendler, T.; Schunkert, H.; Ewert, P.; Navab, N. A survey of catheter tracking concepts and methodologies. Med. Image Anal. 2022, 82, 102584. [Google Scholar] [CrossRef]
  59. Lessard, S.; Lau, C.; Chav, R.; Soulez, G.; Roy, D.; de Guise, J.A. Guidewire tracking during endovascular neurosurgery. Med. Eng. Phys. 2010, 32, 813–821. [Google Scholar] [CrossRef]
  60. Vandini, A.; Glocker, B.; Hamady, M.; Yang, G.-Z. Robust guidewire tracking under large deformations combining segment-like features (SEGlets). Med. Image Anal. 2017, 38, 150–164. [Google Scholar] [CrossRef] [PubMed]
  61. Wang, P.; Chen, T.; Zhu, Y.; Zhang, W.; Zhou, S.K.; Comaniciu, D. Robust guidewire tracking in fluoroscopy. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  62. Zweng, M.; Fallavollita, P.; Demirci, S.; Kowarschik, M.; Navab, N.; Mateus, D. Automatic guide-wire detection for neurointerventions using low-rank sparse matrix decomposition and denoising. In Proceedings of the Workshop on Augmented Environments for Computer-Assisted Interventions, Munich, Germany, 6 October 2015. [Google Scholar]
  63. Ambrosini, P.; Ruijters, D.; Niessen, W.J.; Moelker, A.; Walsum, T.v. Fully automatic and real-time catheter segmentation in X-ray fluoroscopy. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017. [Google Scholar]
  64. Nguyen, A.; Kundrat, D.; Dagnino, G.; Chi, W.; Abdelaziz, M.E.; Guo, Y.; Ma, Y.; Kwok, T.M.; Riga, C.; Yang, G.-Z. End-to-end real-time catheter segmentation with optical flow-guided warping during endovascular intervention. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 1 June 2020. [Google Scholar]
  65. Subramanian, V.; Wang, H.; Wu, J.T.; Wong, K.C.; Sharma, A.; Syeda-Mahmood, T. Automated detection and type classification of central venous catheters in chest X-rays. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  66. Zhou, Y.-J.; Xie, X.-L.; Zhou, X.-H.; Liu, S.-Q.; Bian, G.-B.; Hou, Z.-G. A real-time multifunctional framework for guidewire morphological and positional analysis in interventional X-ray fluoroscopy. IEEE Trans. Cognit. Dev. Syst. 2020, 13, 657–667. [Google Scholar] [CrossRef]
  67. Li, R.-Q.; Bian, G.; Zhou, X.; Xie, X.; Ni, Z.; Hou, Z. A two-stage framework for real-time guidewire endpoint localization. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019. [Google Scholar]
  68. Vlontzos, A.; Mikolajczyk, K. Deep segmentation and registration in X-ray angiography video. arXiv 2018, arXiv:180506406. [Google Scholar]
  69. Vernikouskaya, I.; Bertsche, D.; Rottbauer, W.; Rasche, V. Deep learning-based framework for motion-compensated image fusion in catheterization procedures. Comput. Med. Imaging Graph. 2022, 98, 102069. [Google Scholar] [CrossRef]
  70. Liu, D.; Tupor, S.; Singh, J.; Chernoff, T.; Leong, N.; Sadikov, E.; Amjad, A.; Zilles, S. The challenges facing deep learning–based catheter localization for ultrasound guided high-dose-rate prostate brachytherapy. Med. Phys. 2022, 49, 2442–2451. [Google Scholar] [CrossRef] [PubMed]
  71. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  72. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition (CVPR2009), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  73. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:170404861. [Google Scholar]
  74. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  75. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
Figure 1. Variability of catheter images. (A) Different positions and orientations of the catheter. (B) Different orientations of the marker band. Catheter outer diameter is 4.6 mm for scale.
Figure 1. Variability of catheter images. (A) Different positions and orientations of the catheter. (B) Different orientations of the marker band. Catheter outer diameter is 4.6 mm for scale.
Applsci 13 07778 g001
Figure 2. Schematic visualization of image segmentation in region selection network.
Figure 2. Schematic visualization of image segmentation in region selection network.
Applsci 13 07778 g002
Figure 3. Schematic visualization of image segmentation in localizer network.
Figure 3. Schematic visualization of image segmentation in localizer network.
Applsci 13 07778 g003
Figure 4. Schematic visualization of dual network. Red box shows the region of interest and red arrow points the marker.
Figure 4. Schematic visualization of dual network. Red box shows the region of interest and red arrow points the marker.
Applsci 13 07778 g004
Figure 5. Distribution of errors in landmark detection for 1 region and 9 regions.
Figure 5. Distribution of errors in landmark detection for 1 region and 9 regions.
Applsci 13 07778 g005
Figure 6. Graph of accuracy of region selection for various numbers of regions.
Figure 6. Graph of accuracy of region selection for various numbers of regions.
Applsci 13 07778 g006
Figure 7. Example of incorrect (A) and correct (B) selection of region containing target.
Figure 7. Example of incorrect (A) and correct (B) selection of region containing target.
Applsci 13 07778 g007
Figure 8. Average percentage error for different regions.
Figure 8. Average percentage error for different regions.
Applsci 13 07778 g008
Table 1. Comparison of proposed method, in terms of accuracy, training time, and inference time, with other methods.
Table 1. Comparison of proposed method, in terms of accuracy, training time, and inference time, with other methods.
ModelRegion NumberMean ErrorStandard Deviation of ErrorTraining TimeInference TimeOutput Type
Pixels from 512 × 512 ImageAverage for Each Image
Mobile Net [73]5 × 519.299.773.8 s3.6 msLandmark
ResNet [74]5 × 53.356.234.9 s4.4 msLandmark
Dense Net [75]5 × 53.867.2921.5 s11.2 msLandmark
U-Net [17]11.006.130.2 s10 msMask
VGG [71]17.364.902.1 s1.7 msLandmark
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aghasizade, M.; Kiyoumarsioskouei, A.; Hashemi, S.; Torabinia, M.; Caprio, A.; Rashid, M.; Xiang, Y.; Rangwala, H.; Ma, T.; Lee, B.; et al. A Coordinate-Regression-Based Deep Learning Model for Catheter Detection during Structural Heart Interventions. Appl. Sci. 2023, 13, 7778. https://doi.org/10.3390/app13137778

AMA Style

Aghasizade M, Kiyoumarsioskouei A, Hashemi S, Torabinia M, Caprio A, Rashid M, Xiang Y, Rangwala H, Ma T, Lee B, et al. A Coordinate-Regression-Based Deep Learning Model for Catheter Detection during Structural Heart Interventions. Applied Sciences. 2023; 13(13):7778. https://doi.org/10.3390/app13137778

Chicago/Turabian Style

Aghasizade, Mahdie, Amir Kiyoumarsioskouei, Sara Hashemi, Matin Torabinia, Alexandre Caprio, Muaz Rashid, Yi Xiang, Huzefa Rangwala, Tianyu Ma, Benjamin Lee, and et al. 2023. "A Coordinate-Regression-Based Deep Learning Model for Catheter Detection during Structural Heart Interventions" Applied Sciences 13, no. 13: 7778. https://doi.org/10.3390/app13137778

APA Style

Aghasizade, M., Kiyoumarsioskouei, A., Hashemi, S., Torabinia, M., Caprio, A., Rashid, M., Xiang, Y., Rangwala, H., Ma, T., Lee, B., Wang, A., Sabuncu, M., Wong, S. C., & Mosadegh, B. (2023). A Coordinate-Regression-Based Deep Learning Model for Catheter Detection during Structural Heart Interventions. Applied Sciences, 13(13), 7778. https://doi.org/10.3390/app13137778

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop