Next Article in Journal
ACL Reconstruction in Skeletally Immature Athletes: Current Concepts
Previous Article in Journal
Exploring Perceptions and Experiences of Patients Undergoing Transcranial Magnetic Stimulation (TMS) for Depression and Adjustment Disorder in Romanian Private Practices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

The Artificial Intelligence-Enhanced Echocardiographic Detection of Congenital Heart Defects in the Fetus: A Mini-Review

by
Khadiza Tun Suha
1,
Hugh Lubenow
1,
Stefania Soria-Zurita
2,
Marcus Haw
2,
Joseph Vettukattil
1,2,* and
Jingfeng Jiang
1,*
1
Biomedical Engineering Department, Michigan Technological University, Houghton, MI 49931, USA
2
Betz Congenital Heart Center, Helen DeVos Children’s Hospital, Grand Rapids, MI 49503, USA
*
Authors to whom correspondence should be addressed.
Medicina 2025, 61(4), 561; https://doi.org/10.3390/medicina61040561
Submission received: 12 January 2025 / Revised: 7 March 2025 / Accepted: 18 March 2025 / Published: 21 March 2025
(This article belongs to the Section Cardiology)

Abstract

Artificial intelligence (AI) is rapidly gaining attention in radiology and cardiology for accurately diagnosing structural heart disease. In this review paper, we first outline the technical background of AI and echocardiography and then present an array of clinical applications, including image quality control, cardiac function measurements, defect detection, and classifications. Collectively, we answer how integrating AI technologies and echocardiography can help improve the detection of congenital heart defects. Particularly, the superior sensitivity of AI-based congenital heart defect (CHD) detection in the fetus (>90%) allows it to be potentially translated into the clinical workflow as an effective screening tool in an obstetric setting. However, the current AI technologies still have many limitations, and more technological developments are required to enable these AI technologies to reach their full potential. Also, integrating diagnostic AI technologies into the clinical workflow should resolve ethical concerns. Otherwise, deploying diagnostic AI may not address low-resource populations’ healthcare access disadvantages. Instead, it will further exacerbate the access disparities. We envision that, through the combination of tele-echocardiography and AI, low-resource medical facilities may gain access to the effective detection of CHD at the prenatal stage.

1. Introduction

According to data from the Centers for Disease Control, congenital heart defects (CHDs), or diseases arising from abnormal heart formation, afflict approximately 1% of newborns annually in the United States [1,2]. CHD can involve a problem with the structure or function of the heart, altering normal blood flow. Some CHDs, such as single ventricle defects, represent critical heart conditions. Therefore, a swift prenatal diagnosis and timely interventions can reduce associated morbidity and mortality.
Echocardiography is a non-invasive diagnostic ultrasound imaging technique widely used to screen for structural and functional anomalies. Obstetric fetal CHD screening commonly includes a four-chamber (4ch) view and an outflow track evaluation. In contrast, a fetal echocardiogram is a dedicated and focused diagnostic ultrasound of the fetal heart performed and analyzed by subspecialized pediatric congenital cardiologists and sonographers. According to a meta-analysis by Zhang et al. [3], prenatal echocardiography for CHD demonstrates only moderate sensitivity (68.5%). However, the sensitivity for detecting CHD within low-risk pregnant women (i.e., in a typically obstetric setting) is considerably lower (36.1% [3]), suggesting that a significant portion of CHDs may be missed during the regular obstetric ultrasound [4,5,6].
Approximately one-sixth to one-fourth of these unidentified infants have a critical CHD, which is defined as a severe and life-threatening heart condition requiring an intervention in the first month of life [7]. Unfortunately, 25% of infants with an unidentified CHD are missed in postnatal screening and are discharged from the hospital [7,8], resulting in many infant deaths in the United States [2]. Missed critical CHDs are considered a gap in the current workflow of the diagnosis of CHD.
In the era of artificial intelligence (AI), AI-assisted technologies, including advanced algorithms, analytics, and imaging technologies, can potentially impact clinical practices in radiology. At the same time, these technologies hold promises for advancing fetal and neonatal care, particularly for closing the “diagnostic gap” mentioned above.
This article reviews recent developments and elevated capabilities in the use of AI with echocardiography for the prompt and precise prenatal detection of CHDs. The rest of this paper is structured as follows: Section 2 provides technical background on AI applications in echocardiography, followed by a discussion in Section 3 on how AI technologies are utilized to enhance the echocardiography-based diagnosis of CHDs. Section 4 presents a critical analysis of existing studies, followed by a discussion of ethical concerns in Section 5. Section 6 provides the current limitations of AI for detecting CHDs before the closing remarks in Section 7.

2. Technical Background

2.1. Techniques of AI

AI is a broad term covering machine-based algorithms and models designed to mimic human logic and intelligence. Machine Learning (ML) is a subset of AI [9]. In the last few decades, ML has increasingly been used in biomedicine to detect pathological conditions and assist in making diagnoses [10,11,12]. In this context, ML algorithms teach a computer to process echocardiographic data and extract useful information or features (also known as pattern recognition [13]), just like a medical doctor does, so the computer can predict the presence of a CHD, given the echocardiographic data.
ML can be divided into subfields that include (1) supervised learning, (2) unsupervised learning, (3) semi-supervised learning, and (4) reinforcement learning, as shown in Figure 1.
In supervised learning, a computer is trained to learn features or information associated with labeled data, i.e., echocardiograms with or without CHD diagnosed by medical doctors. Combining these learned features allows ML algorithms to make predictions or decisions. Logistic regression, random forests, gradient boosting, and support vector machines are commonly used supervised learning models. After an ML model has been trained, the model’s accuracy is tested with an unseen dataset [14,15]. Supervised learning inherently relies on feature extracts, so the training outcomes may be sensitive to input data.
The second ML method, unsupervised learning, allows a computer algorithm to discover correlations or relationships among unlabeled data [15]. Unsupervised ML algorithms can be applied to clustering, finding similarity, and dimensionality reduction. Because the true answers from data are unknown, quantitative performance assessments of these unsupervised ML methods are not often sought. However, such assessments become essential in biomedical research when labeled data becomes infeasible and when the assessments would offer insight into biomedical data. Investigating the plausibility and salience of identified correlations or relationships is still useful for knowledge discovery.
Semi-supervised learning falls between supervised and unsupervised learning. Semi-supervised learning uses labeled and unlabeled data to train ML models; it is often used when there is not enough labeled data, largely due to the costs associated with generating them. Consequently, semi-supervised learning can use a small amount of labeled data to guide model training, while a large amount of unlabeled data can refine the model.
The last subfield in ML is reinforcement learning (RL), often used in robotics and control systems. In this subcategory, the algorithm interacts with the environment, receives penalties for specific actions, and learns through trial and error. A non-technical description of RL can be found in Appendix A.
Deep learning (DL), a popular ML subfield, is often tied to neural networks mimicking how the human brain works. In DL, the term “deep” indicates the network’s depth or the number of layers of these “artificial” neurons. “Artificial” neurons can be activated, and connections among these neurons are associated with weights. DL models need to be trained with a lot of data, akin to teaching a child to identify an apple by showing them many pictures of apples. The training process allows a DL network to achieve a pre-determined goal. Because the training must be cast in a mathematical optimization framework, the pre-determined goal must be written out as a function known as a loss function.
During training, the DL model makes predictions based on the input data, and we measure its performance using the loss function. The lower the loss, the better our model performs. This process, also known as “backpropagation”, helps minimize the loss function by adjusting the DL model’s internal parameters (i.e., weights and neuron activation).
Loosely speaking, a trained DL model acts like a carefully designed filtration system; different neuron layers in the system represent varying features extracted from the input data. For example, suppose we input an echocardiogram into a DL model. In that case, some early layers might recognize simple features, such as the gross shape (e.g., heart chambers, valves, etc.), and later layers might learn how to combine separate features to resemble the object of interest, i.e., the heart and a diagnosis associated with the heart. DL models are widely used for medical applications, such as medical image processing, speech recognition, and biomedical signal processing.
It is worth noting that, as shown in Figure 1, DL is a methodology parallel to the four categories mentioned above. In other words, DL models can be trained under supervised, unsupervised, or semi-supervised frameworks. The training process iteratively adjusts weights associated with neurons. More recently, the DL methodology has been expanded into reinforced learning as well, as shown in Figure 1.
Notably, convolutional neural networks (CNNs) are a class of core DL methods for handling imaging data. A non-technical explanation is provided in Appendix A. CNNs have numerous applications in medical imaging, particularly in image classification, object detection, and image segmentation. Specifically, variants of CNNs have been used to calculate biometric features, segment lesions, detect anatomical structures, and control image quality in conjunction with fetal echocardiography, as reviewed below.
To this end, in AI-augmented echocardiography, one common approach is first to process ultrasound images or videos. Then, these preprocessed ultrasound data are trained and validated to build deployable AI models that can be translated into the clinical setting. The choice of the actual ML algorithms depends on the specific clinical application and clinical data, as reviewed below.

2.2. Echocardiogram Background

Echocardiography employs ultrasound technology to image cardiac structure and function, and it is an essential tool in diagnosing, assessing, and managing heart conditions. This technique utilizes soundwaves ranging from a 2 to 12 MHz frequency by an ultrasound transducer (also known as a probe), with higher frequencies enhancing the image resolution at the expense of decreasing the image depth and balancing clarity with anatomical coverage. During echocardiography, the ultrasound transducer transmits acoustic waves. The acoustic waves travel through the biological tissues and often reflect and scatter due to differences in acoustic impedance. The returned echoes (through reflections and scattering) can be received by the same ultrasound transducer and displayed as ultrasound brightness images to reveal tissue boundaries and internal compositions. These rendered brightness images of the heart are B-mode echocardiograms, enabling the real-time internal visualization of heart conditions and dynamics. Fetal echocardiograms are specialized applications of echocardiography meant to assess fetal heart structure and function during pregnancy (Figure 2).
These scans can be performed as early as 18 weeks of gestation, once heart structures have developed, thus allowing the early detection of CHD. If an issue is identified, ultrasound scans after 24 weeks are useful for initial treatment planning. During pregnancy, various imaging techniques can be used depending on the gestational age, including abdominal and endovaginal approaches, which involve differing the transducer placement on the patient. Sonographers view structures, such as the walls, septum, chambers, and valves, utilizing multiple views, such as the four-chamber view, abdominal situs view, left ventricular outflow tract view, right ventricular outflow tract view, three-vessel view, and trachea view, for comprehensive evaluation.
The sonographer manually operates frame selection, which is critical for cardiac image analysis. Different phases of the cardiac cycle are independently evaluated and selected to detect the end diastole frame, which denotes the largest cardiac cavity area, and the end-systolic frame, which is used to assess changes in cardiac dimensions and volume to evaluate heart function. The frame sets are used as reference points to determine cardiac biometrics. Each chamber and the wall separating them is individually assessed to detect any significant dropouts denoting septal defects. The chamber and valve movements are observed throughout the cardiac cycle to detect any size variations, conduction abnormalities, myocardium abnormalities, ventricular dysfunction, or abnormal position or the formation of the great vessels and valves.

3. Artificial Intelligence (AI) Applications

In recent years, numerous studies have been conducted on the effective use of AI methods in fetal echocardiography to diagnose congenital heart disease. We used relevant terms to search through WEB of Science, Scopus, and PUBMED using the following criteria: (1) publication data range from 2020 to the present (i.e., last five years); (2) the selected topic is related to applying AI to echocardiographic data for detecting CHD at the prenatal stage; and (3) the publication language is English. The details of key phrases used in the literature search can be found in Appendix B below.
The initial search was performed by two individuals (KTS and HL). They obtained 57 and 35 papers, respectively. Then, their lists merged, and they reviewed abstracts to determine their fitness for the abovementioned topic. After the abstract review, the number of papers was reduced to 27. Finally, three individuals (KTS, HL, and JJ) reviewed the twenty-seven papers to finalize their fitness to our topic, resulting in the twenty papers in Table 1 below.
Table 1 briefly outlines these twenty papers. Our review below focuses on the effectiveness and relevance of AI technologies in detecting CHDs.

3.1. Image Quality Control and View Selection

Image quality is crucial in detecting anatomic variations and diagnosing abnormalities in fetal echocardiographic images. Echocardiography quality is influenced by involuntary fetal movements, unusual heart size, ultrasonic speckle noise and artifacts, and the knowledge level of the fetal echocardiographer or primary physician [38]. An image quality control framework is essential for obtaining the best results in the fetal echocardiographic detection of CHDs. This research has motivated several recent studies.
Dong et al. developed a framework composed of three CNN-based networks (B-CNN [basic CNN], D-CNN [deeper CNN], and ARVB Net [aggregated residual visual block network]) for quality control in the four-chamber detection of the fetal heart. Their experiment on fetal echo datasets shows the framework’s effectiveness, especially ARVBNet, for image quality control. The algorithm provides a quantitative assessment and real-time image optimization information for novices. Although a three-step framework showed superior accuracy, an error in the first rough classification can affect the overall performance. The authors suggest that a multi-task learning model studied in future research may solve this problem to an extent [23].
Other researchers also focused on developing the quality of the four-chamber view. Notably, Qiao and colleagues creatively used generative adversarial network models to improve the quality of four-chamber view echocardiograms [16,24]. In the first study, Qiao et al. created a Pseudo-Siamese Feature Fusion Generative Adversarial Network (PSFFGAN) to synthesize four-chamber views with artifacts, ultrasonic characteristics, and speckle noises. Specifically, the PSFFGAN uses a Triplet Generative Adversarial Loss Function (TGALF), which optimizes the PSFFGAN to completely extract the cardiac anatomical structure information and obtain high-quality fetal four-chamber views using four-chamber sketch images [16]. In their second study, Qiao et al. created an AI model to collect diversified and high-quality fetal four-chamber views. The Enhanced-Style Consistent Modulation (ESCM) module learns two sets of modulation parameters for sketch and semantic contours. They also added a Generative Adversarial Network that performs image synthesis and quality evaluation at different resolutions. Their proposed sketch mass loss function effectively suppresses the loss values in the training process of the four-chamber detection [24].
Another application of AI in fetal echocardiography quality control is enhancing echocardiograms by increasing the brightness and contrast. Sutarno et al. built a CNN-based Low Light Image Enhancement (LLIE) architecture to enhance the light intensity of echocardiographic images. They also added a CNN-based classifier named “FetalNet” to create a stacked architecture that could improve the success rate of fetal heart defect prediction. The model (post-enhancement) could increase the success rate by approximately 25% and produce 100% sensitivity using testing data [19].
Selecting high-quality views from clinically acquired 2D echocardiograms is critically important to diagnose CHD accurately. The work by Li et al. [21] is promising. They proposed a DL neural network model that can simultaneously perform multiview classification and image quality assessment. Their deep learning neural network model consists of a backbone network, a neck network, a view classification branch, and a quality assessment branch. The backbone network is for feature extraction, while the neck network fuses features. The view classification branch can classify echocardiograms into seven different views: axial four-chamber (A4C), parasternal view of the pulmonary artery (PSPA), parasternal long axis (PLAX), parasternal short axis at the mitral valve level (PSAX-MV), parasternal short axis at the papillary muscle level (PSAX-PM), parasternal short axis at the apical level (PSAX-AP), and other views. Using the 170,311 echocardiograms acquired, their testing results showed an overall classification accuracy of 97.8%.
Wu et al. also applied DL technologies to recognize anatomical structures for view selections [34]. Specifically, they attempted to recognize five different views: apical four-chamber view, three-vessel view, three-vessel trachea view, right ventricular outflow tract view, and left ventricular outflow tract view. Overall, their method achieved remarkable performance, i.e., high sensitivity (75–100%), high specificity (>93%), and high positive predictive value (90.0–99.5%).

3.2. Detection and Classification of Fetal Heart Defects

The AI detection of fetal heart defects from echocardiograms is a vital step in classifying features for the potential diagnosis of CHD and other abnormalities. To accomplish this, several methods have been developed to obtain the possible views and features of fetal echocardiography with their inherent benefits and drawbacks. To build a DL model capable of image analysis, it needs to be trained using a controlled dataset of varying types. This would set the benchmark for the DL to complete the task when correctly asked. These datasets must be sufficiently large to configure the DL properly; however, there is a decrease in benefits for each addition. Several types of DL algorithms attempt to identify fetal hearts differently and are, therefore, trained uniquely.
Truong et al. [25] propose using ML with a random forest algorithm framework that employs tenfold cross-validation to train models for assessing the presence of CHDs. This method demonstrates that ML using RF algorithms can improve sensitivity with a value of 0.85, a specificity of 0.88, positive prediction rates of 0.55, and a negative prediction rate of 0.97 to detect CHD (B). Wong et al. [26] utilized a version of V-Net deep learning for multiview and multiclass image segmentation. Through the careful selection of datasets based on images following verified landmark conditions, they showed how V-Net alone can accurately segment fetal echocardiograms in multiple views and abnormal conditions. Their training of the modified V-Net used 598 images with 14 anatomical labels, excluding some abnormalities that did not contain the structure necessary for the label, between the gestational ages of 18 + 0 and 24 + 6 weeks. These techniques have varying benefits, especially with image analysis and potential differences in the resource draw and needs. Various detection methods can be more or less suited for the classification of different fetal heart features from different views. The classification of different features, such as septum, heart walls, ventricles, and vessels, from different views is vital for diagnosing CHD. They provide the information necessary to derive a diagnosis of CHD or other potential abnormalities [28]. Wong et al. used a method that searches for 14 anatomical landmarks in a combined view model of 4CHV and 3VTV. They noted that some segmentations are more difficult to segment out, such as the trachea, interatrial septum, and mitral valve, with Dice coefficients lower than 70%. They successfully segmented anatomical landmarks within a 1% Dice coefficient of both separate models’ available landmarks, concluding that using multiple models for multiple views is unnecessary.

3.3. Measurements of Cardiac Function and Parameters

Defining the heart’s condition through the measurement and assessment of fetal biometry allows for reliable diagnostic procedures to occur. The reliable positive and negative prediction of CHD and other abnormalities in the fetal heart enables crucial early intervention and treatment.
Several different measurements are used to quantify the possibility of CHD or other abnormalities. Doppler traces can show the flow from the aortic arch and arterial duct, across the oval foramen and in the pulmonary veins, inflow through the atrioventricular valves, and outflow through the arterial valves. The verification and measurement of cardiac structures and volumes are crucial for diagnosis. The identification of anatomical landmarks and the quantization of their size are achieved through methods such as cropping and segmentation [28]. L. Yan et al. used a two-stage method to mark and segment three vessel samples through an object identification tool called Yolov5, which localizes the Region of Interest (ROI) within the original full-sized echocardiographic images. Subsequently, a modified Deeplabv3 equipped with their novel AMFF (Attentional Multi-scale Feature Fusion) module is applied in the second stage to segment the three vessels within the cropped ROI images. Furthermore, measurements can be drawn from time intervals in echocardiography and are used in several measurements to observe periods of activity within the fetal heart, as many actions within the fetal heart are periodic.
The segmentation of a view identifies anatomical landmarks, from which measurements can be extrapolated. As shown in Figure 3, the segmentation of the heart from an echocardiogram can be achieved by either a DL segmentation algorithm (e.g., ZoomNet [39]) or a DL object detection algorithm (e.g., YOLO [40]).
Measuring the parameters and functions identified and quantified in the AI detection of fetal biometry is integral to diagnosing positive and negative predictions of CHD. Many of these measurements are made to encompass the possibility of abnormalities present or absent in a cardiac system. Without a comprehensive assessment through various parameters, CHD detection is less reliable.

3.4. Diagnosis of Specific Abnormalities

The analysis of a specific abnormality in the fetal cardiac structure is significant in determining congenital heart disease. AI can assist physicians in diagnosing malformations in the fetal heart. Ventricular septal defect (VSD) is a deformity in which there is a deficiency in the wall separating the right and left ventricles. Small- to medium-sized VSDs are missed even by expert fetal cardiologists. Atrioventricular septal defects (AVSDs) are more serious defects that, if left undetected, can lead to severe consequences after birth. Several studies have focused on applying AI models in fetal echocardiographic images to detect these defects. Veronese et al. analyzed the performance of Fetal Intelligent Navigation Echocardiography (FINE) used for the prenatal identification of AVSDs. They used spatiotemporal image correlation (STIC) volume datasets from four-chamber views. This navigation system successfully determined the nine standard fetal views after it was applied to the input datasets. The accuracy was 100%, as it was able to identify abnormal fetal hearts with atrioventricular septal defects and a common atrioventricular valve [36].
Tetralogy of Fallot (TOF) is also a severe condition that can lead to infant death resulting from pulmonary artery obstruction, right ventricle thickening, septal anomalies with an overriding aorta, and many other anomalies. In severe cases, TOF has higher mortality rates than VSD. Yu et al. differentiated TOF and VSD cases using four convolutional neural network models, VGG19, ResNet50, NTS-Net, and the weakly supervised data augmentation network (WSDAN), on fetal echocardiography images. WSDAN, with an AUC of 0.873, showed the best performance among all the models in classifying the images [18].
Another kind of CHD found in infants is total anomalous pulmonary venous connection (TAPVC). Wang et al. aimed to predict TAPVC using a parameter called the PLAS ratio from fetal echocardiography. They analyzed if the calculation of the PLAS (the ratio of the epicardium-descending aortic distance to the center of the heart-descending aortic distance) using CNN would be a useful feature to detect TAPVC. The dataset included echocardiographic images of 287 fetuses with normal hearts and 32 fetuses with isolated TAPVC (25.6 ± 2.7 weeks). The computerized calculation used five DL segmentation models, including DeepLabv3+, FastFCN, PSPNet, and DenseASPP. The results showed that the PLAS ratio was higher in the TAPVC cases than in the controlled group in both manual and DL processes [17].
One of the most critical CHDs is hypoplastic left heart syndrome (HLHS). Day et al. trained CNN to automatically detect fetal HLHS using a specific single-center fetal database. This retrospective fetal database included 161 fetal patients and provided single-expert-identified ‘ideal’ four-chamber (4ch) view frames to identify and extract up to 10 frames chronologically before and after the ‘ideal’ 4ch frame. A total of 10,248 frames were obtained and analyzed (3960 with HLHS). Their results suggest that while specific AI models are not ready or good enough to operate independently, they will augment human performance [31].
In another study by Li et al. [33], AI detected VSDs very accurately, i.e., mAP@0.5 (a variant definition of accuracy) reaching 0.926. Interestingly, under AI guidance, the VSD detection accuracy improved by 6.7% and 2.8% among junior- and intermediate-level doctors, respectively.

3.5. Identification of Normal Heart from CHD

Accurately distinguishing normal fetal hearts from complex congenital CHDs is an essential and widespread need. Poor sensitivity in this task can reduce treatment options, negatively impact outcomes, and lead to unnecessary tests and referrals. Several studies have created AI models that differentiate between normal and abnormal fetal hearts. Arnaout et al. [20] showed that trained ensemble learning models performed well in detecting CHD. They used CNN on multimodal images to segment different views of the fetal heart and then distinguish between normal hearts and complex CHD cases. The model achieved an AUC of 0.99, 95% sensitivity, and 100% negative predictive value, working on 4108 fetal test cases in classifying normal and abnormal hearts.
Using a variant CNN model, the You Only Look Once (YOLO) model [40] shows a noteworthy performance in identifying CHD. Yang at el. utilized the YOLOv5 model to distinguish normal and abnormal fetal heart ultrasound images accurately. The dataset included normal fetal cases and VSD test cases where YOLOv5 showed 92.79% accuracy in identifying abnormal fetal hearts [32]. In addition to a high accuracy and prediction rate, less complexity in the computational process is also a desirable factor in working with large datasets. Magesh et al. proposed a new model integrating deep Reg net with CNN to extract features from ultrasound images and classify them into normal and CHD with less computational complexity. The ultrasound images from 363 pregnant women were also preprocessed using a SCRAB filter to eliminate the noise artifacts. The model could classify the images into two groups with 98.4% accuracy in identifying the CHD cases. It showed better performance compared with other DL networks used in fetal heart CHD classification [35].
While DL models show a higher accuracy in classifying normal and abnormal hearts from echocardiographic images, considering high-volume clinical indicators can also be important for validating the results. Qu et al. performed a data analysis using ML on pregnant women who had fetuses with CHDs. An explainable boosting machine (EBM) was used for prediction, and around 1127 predictors were included, with the top predictors selected according to contributions and performances. The results showed 0.65 accuracy, 0.74 sensitivity, and 0.65 specificity, where maternal UA, glucose, and coagulation were the most consistent predictors [37]. In their study, echocardiographic data read by two experts were used as the ground truth for diagnosing CHDs.
Fetal arrhythmia may cause fetal death, abortion, or hydrops. Although electrocardiogram (ECG) is the leading modality for assessing heart rhythm it is technically challenging in fetuses. An echocardiogram can detect arrhythmias using M-mode across the atria and ventricle to detect an AV block or atrial arrhythmias. Pulse Wave Doppler (PWD), as developed by Yang et al., uses an intelligent quantification system (HR-IQS). Specifically, they used a DL-based method to detect PWD signals automatically. The arrivals of E, A, and V waves are detected, and fetal cardiac time intervals (CTI) are computed. The resultant HR-IQS system was validated using multicenter PWD data, specifically from 6498 PWD spectrums over the junction between the left ventricular inflow and outflow tracts from 2630 fetuses. They reported a “significant positive correlation (p < 0.001) and moderate-to-excellent consistency (p < 0.001) between the manual and HR-IQS automated measurements of CTIs” [22]. The preliminary results indicated that the trained HR-IQS system has the potential to detect fetal arrhythmia [22].

4. Critical Analysis

A detailed analysis of the eighteen (18) AI models reviewed above was conducted, and their characteristics are summarized in Table 2 below. Two papers in Table 1 are not included, as justified below. In [36], AI was applied to medical testing data, and echocardiographic data were used to verify the presence of CHD. Ref. [37] was a case study and only included four cases. Thus, neither study fits the scope of this critical analysis, in which echocardiographic data are used as inputs for AI.

4.1. Categories of AI Models

According to the American Society of Echocardiography (ASE), AI products can be divided into (1) Assisted, (2) Augmentative, and (3) Autonomous models [41]. In the first category, AI products will produce only clinically relevant data. For instance, an AI product measuring the deformation of the heart is considered an Assisted AI product. An Augmentative AI product analyzes raw echo data and produces clinically relevant data. Thus, if the heart deformation is converted to a summary metric reflecting myocardial infarction, the AI model will be elevated to an Augmentative model. Models in the first and second categories require physicians’ involvement in decision-making and reporting. In contrast, autonomous models will make clinically relevant decisions using an AI model without the involvement of a medical doctor.
Interestingly, there are no Augmentative AI models. Two AI models are synthetic echocardiogram generators that can be used for training. The rest of the papers are split between Assisted (7/14) and Autonomous (9/14) AI models.

4.2. Selection of Training Data

As shown in Table 2, only four studies (4/18) used more than 10,000 images for training. Among them, three studies included data from more than one ultrasound vendor. Nearly half of these studies (7/18) used data from a single center with small data sizes (<1000 echocardiograms). Using a large amount of data acquired across imaging vendors and multiple medical centers will enhance the validity of these ML algorithms.
Only a small fraction of these studies (approximately 10%; 2/18) provided sufficient information regarding data acquisition and collection methods. Evaluating ML models in CHD detection without knowing detailed information about the data hampers our efforts to truly validate these ML methods and translate them into the clinical workflow.
Unlike other ML-based medical research (e.g., segmentation of brain cancer), none of these studies used a public database. Although important, the establishment of public echocardiography databases for fetal applications is still in its infancy. Public databases with comprehensive data are crucial for further developing AI algorithms to detect CHD and understand the potential of developed ML models.

4.3. ML Models and Completeness of Their Descriptions

As shown in Table 1, all studies except one used DL-based methods. In particular, due to the tremendous achievements of CNN-based models in Radiology, most studies adopted one or more variants of CNN models. One of the significant advantages of CNN models is that features extracted from echocardiograms are automated. Thus, these models can be integrated into the clinical workflow with full automation. In contrast, the study by Truong et al. [25] adopted Random Forest, a non-DL-based method, to classify fetuses into normal and abnormal groups. Twenty (20) manually crafted features (e.g., cardiothoracic ratio, maximum Doppler Velocity at aortic artery, etc.) are fed into the Random Forest model. These manually crafted features prevent the proposed ML model from being fully automated unless fully automated methods are developed to create these features.
After reviewing the eighteen papers listed in Table 2, none of these studies performed a formal computational complexity analysis. We found that ML models can be divided into three groups based on the completeness of the description of the ML model. Source codes are provided by three (3 out of 18) papers, and readers can potentially verify their neural network configurations. In contrast, in four (4 out of 18) papers, three studies used neural network models, and their configurations were not discussed. One paper used a Random Forest classifier. Only two of the remaining eleven papers in the last group provided sufficient details to reproduce their neural network models. Although most of the other nine papers used published models, the information provided by the authors may not allow others to grasp their ML models fully. For instance, the classical ResNet model [42] has many variants. Simply stating that the ResNet model was used does not provide sufficient information for further assessments.

4.4. Generalization and Computational Requirements of Reviewed ML Models

Generalization, defined as AI’s ability to apply knowledge to new data [43], has been a key challenge in many diagnostic AI applications. Although no formal studies of AI generalization for detecting CHDs have been identified, most studies (14/18) only used data from one center, and only one-third of the studies used data from more than one (echocardiography) vendor. Also, two-thirds of the studies did not describe their data acquisition protocols. Alarmingly, one study [35] included little information about the data procurement process. Limitations in data collection pose difficulties in generalizing these AI models. Furthermore, as shown in Table 2, eight studies (out of 18) used a small amount of data (<1000 images). Thus, these models may be too closely tuned for the training data, i.e., overfitting, making the generalization of these models challenging. Consequently, we submit that more work is needed to demonstrate that the results of these studies (Table 2) can be generalized.
To improve the generalization, AI algorithms should be validated using diverse sources of data. Since collecting data and subsequently making accurate annotations requires considerable resources, data augmentation has been attempted. Sfakianakis et al. [43] used DL-generated synthetic ultrasound data to improve training for image segmentation. In contrast, Chen et al. [44] adopted more traditional image data augmentation methods, such as rotation, to improve the training of DL-based object detection. The GAN-based echocardiogram synthesis methods by others [16,24] can also be used for data augmentation. In the literature, advanced modeling methods [45,46] have been used to generate synthetic data for ultrasound elastography applications [47]. Given appropriate developments, these advanced modeling methods can also potentially train DL-based CHD detection using echocardiographic data.
Although data augmentation may be appropriate for certain applications, such as image segmentation and object detection (e.g., certain views of echocardiograms), the generation of pathologies requires more consensus. That is why federated learning [48] has been explored. Federated learning (also known as collaborative learning) is a specific setting where multiple institutions collaboratively train an AI model without requiring physical data exchanges and centralized data storage. Without the legal paperwork or procedures required to exchange data, the collaborative training of novel AI methods can be accelerated. Goto et al. [49] adopted federated learning to access data from Massachusetts General Hospital, the University of California San Francisco, and Keio University Hospital. Their results showed that access to data from two additional institutions improved the AI’s performance. Consequently, we think the federated learning approach is a practical approach by which computational scientists can access diverse data needed to generalize their AI models.
As we found in Section 4.4, many models did not provide sufficient information to assess their computational complexity formally. However, two-dimensional echocardiograms were mostly used as inputs for these AI models in all applications. Thus, these studies did not require high-end Graphic Processing Units (GPUs) (Table 2). Fifteen studies used non-professional-grade GPUs (i.e., gaming GPUs: GTX 1080, GTX 2080, RTX 3050, and RTX 3090; Nivida Inc., Santa Clara, CA, USA). Two studies did not disclose the GPU equipment. The remaining two studies used the V100 GPU (Nivida Inc., CA, USA), which only has an onboard GPU-accessible 16 GB memory. In 16 studies where the GPU details are provided, the GPU memory varies from 11 GB to 48 GB. This finding suggests that most AI models can be implemented into clinical workstations.

4.5. AI Performance Versus Clinician’s Performance in Clinical Workflow

In the eighteen studies shown in Table 1 and Table 2, no direct comparisons to clinical studies can be found for nine studies. The rest of the nine studies can be divided into three groups: (1) heart structure and vessel delineation [26,28], (2) anatomical measurement [17], and (3) CHD differentiation [20,25,31,32,33,35].
The first two studies [26,28] investigated the application of DL methods to segment either heart chambers or vessels connecting to the heart. Their Dice scores ranged from 0.74 to 0.89. The Dice score measures the (area) overlap rate between the AI-segmented structures/vessels and the ground truth (i.e., clinical experts delineated structure/vessel). A Dice score of one indicates a perfect overlap (i.e., identical), while a Dice score of zero denotes no overlap. The Dice scores ranging from 0.74 to 0.89 suggest that DL methods obtain highly consistent results compared to those obtained by clinical experts. Compared to the anatomical delineation using other modalities (i.e., often above 0.9), such as vessel segmentation using X-ray digital subtraction angiography [50], CT angiography [51], and intravascular ultrasound [52], and brain cancer delineation using MRI data [53], the Dice scores reported in [26,28] were slightly lower.
In one study by Wang et al. [17], in the second group, the post-left atrium space (PLAS) ratios detected by the AI method were comparable to those obtained by human experts (accuracy: 89% vs. 88% and area under the curve [AUC]: 0.88 vs. 0.87).
In the last group, six studies [20,25,31,32,33,35] investigated the detection of CHDs. Two of them were designed to detect specific CHDs. In the study by Day et al. [31] the sensitivity and specificity of the AI-based detection of HLHS were comparable to those reported in the clinical literature. For example, in the UK’s NHS fetal anomaly screening program, the sensitivity for detecting HLHS is around 92.7% [54], while the AI algorithm described by Day et al. has a sensitivity of 94.3%. Thus, the AI algorithm slightly outperforms the clinicians regarding the sensitivity for screening for HLHS. In another study by Li et al. [33], the detection of VSDs was investigated; since its performance metrics differed from the standard metrics in clinical studies, direct comparisons are difficult. However, its performance was excellent (e.g., a variant of the accuracy measure is close to 0.92).
The four studies [20,25,32,35] in the last group were designed to differentiate abnormal fetal hearts from normal ones. All these four studies performed well. Three studies [20,25,35] reported excellent AUC values (i.e., >0.94) and overall accuracy (i.e., >0.88), as shown in Table 3 below. In the other study by Yang et al. [32], the AUC was not reported and the total accuracy was 80%. The AUC values of these three studies are on par with the overall summary AUC value in a meta-analysis of echocardiography-based CHD detection studies [3]. Notably, Arnaout et al. [20] reported an AUC of 0.99, 95% sensitivity, 96% specificity (95% CI, 95–97%), and 100% negative predictive value in distinguishing normal from abnormal hearts.
The specificity values of these three studies were lower than the average specificity of the meta-analysis (93% [average of three studies] vs. 99.8%), as shown in Table 3. However, the sensitivity of the four AI studies was considerably higher (90.5% [average of four studies] vs. 68.5%). It is generally recognized that echocardiography’s sensitivity in screening CHDs varied significantly from prior studies (e.g., from 92% [55] to 16% [56], on average 68.5%). A lower sensitivity (on average 36.1%) typically comes from low-risk pregnant women (i.e., a typically obstetric setting), as the meta-analysis results suggested ([3]). Consequently, an AI-based detection method with a reasonable specificity value (e.g., 90%) can be developed into an effective screening tool in the obstetric setting.

4.6. One Case Study [57]

We found one study in which AI’s performance was compared with multiple human readers in a point-of-care setting for a low-risk population. Between 2015 and 2016, 108 fetuses (42 normal and 66 abnormal) from the northwestern region of the Netherlands were studied. CHDs were confirmed at birth. To mimic an initial screening of CHDs at the prenatal stage, stored images were shown to three fetal cardiac experts as live imaging in a point-of-care setting. The CHD status was hidden from all three fetal cardiac experts. It is also important to note that a DL model previously trained by Arnaout et al. [20] was used in this study. “all CHD status” was hidden from the DL model as well”.
During the initial clinical assessment by OB doctors, 31 abnormal cases (out of 66) were missed, translating to approximately a 50% miss rate. In this low-risk screen setting, a sensitivity of 50% is consistent with the results shown in Table 3. In contrast, the DL model’s sensitivity and specificity were 91% and 78%, respectively. Three blinded human experts achieved a sensitivity and specificity of 55 ± 10% (mean ± SD; range, 47–67%) and 71 ± 13% (range, 57–83%), respectively. They concluded that the previously trained DL model had higher sensitivity than the initial clinical assessment in detecting CHDs in the cohort of 108 fetuses in a low-risk population. Also, the DL model outperformed three fetal cardiac experts, particularly in sensitivity.
Moreover, these abnormal cases included 19 lesions the DL model had not encountered during its training. The encouraging outcome is that the DL model correctly identified 16 of them (out of 19), suggesting the training is sufficient. Recall that the DL model by Arnaout et al. [20] is one of the best DL models to detect CHDs (see Table 3).

4.7. Scientist–Physician Partnership

For all studies listed in Table 2, based on the authors’ affiliations, we found that 17 out of 18 publications resulted from collaborations between computational scientists/engineers and clinicians. This trend reflects the fact that applying AI models to CHD detection is multidisciplinary in nature; thus, collaborations between computational scientists and clinicians are needed to advance this field and translate AI technologies into the clinical workflow.

5. Ethical Concerns of Artificial Intelligence

AI has great potential to improve the detection of fetal CHDs; however, it also raises many ethical dilemmas. The first question that can be raised is the “black box” phenomenon. ML techniques, including DL in medicine, are heavily criticized for their “black box” nature since the way in which many ML models arrive at decisions cannot be explained, and the decision process is not transparent [58,59]. Transparent ML models ensure safety, fairness, and accountability, making them more attractive for medical doctors to use with confidence. This issue can only be solved using more interpretable models and transparency in the validation process. However, regarding detecting CHDs, the “black box” nature of ML models may not be as significant if we use an ML model as a screening tool. We can adjust the ML model to reach a high sensitivity (e.g., >99%) if the false positive rate is not too high (e.g., 30%).
Another ethical concern is that biases exist among ML models. As explained before, supervised ML models learn from input data. Thus, one primary source of biases stems from the biases presented in their training data, which may be related to race, ethnicity, genetics, region, or gender [60]. Moreover, ML models that use echocardiographic data may also need to sample training data across a spectrum of ultrasound vendors, as each vendor often uses proprietary techniques to process echocardiograms before they are displayed on its scanners. Furthermore, data from low-resource medical facilities are rarely used in all the studies evaluated; this is another form of data selection bias. To address the bias problem, we should consider a standard protocol for training data selection [60]. Creating open-source benchmark training data in which different CHDs from various patient populations are appropriately sampled may allow us to mitigate this bias problem. However, eliminating the sources of biases is challenging.
In the era of ML, deploying or training ML models relies on a huge amount of patient data. Consequently, ensuring patient privacy and confidentiality is a primary ethical concern. For instance, CHD detection may be enhanced if electrical health records (EHRs) can be added to the ML model. Then, should patients’ consent be required before using EHR data for building an ML model? The Health Insurance Portability and Accountability Act (HIPAA) requires such consent in the United States. However, such a requirement may not be universal. Patients should be allowed to opt out in countries where such consent is not required.
As shown in Table 2, many studies aimed to develop Autonomous AI models. These models may fundamentally change the patient–clinician relationship. Relying on Autonomous AI can misguide physicians, increasing the risk of causing harm and exposing them to potential legal action [61].
Additionally, there is the issue of fair accessibility [62]. For example, like many other modern technologies, the deployment of diagnostic AI typically starts with areas with sufficient resources. Thus, access to advanced diagnostic AI may further widen the healthcare disparities.

6. Discussion

6.1. General Challenges

Reviewing the available literature, we found that AI algorithms may aid in detecting and classifying CHDs. Although applying ML and DL algorithms in fetal echocardiography has numerous benefits, researchers have encountered several challenges.
First, as shown in Table 2, most AI models are not fully tested. Most studies were conducted with limited data (e.g., a single center and a small number of echocardiograms). The demographic information of these data is not clear. Moreover, biases in the training process cannot be evaluated without knowing the data selection. For instance, Taksøe-Vester et al. highlight the inevitable bias associated with supervised learning models, as these models are trained on datasets manually segmented by humans [63]. Thus, the superior accuracy reported by some of these studies may not be generalized. The development of such AI models will benefit from creating public databases with comprehensive patient information, allowing all AI models to be tested transparently.
Second, we observed that all the studies shown in Table 2 were outcomes of multidisciplinary collaborations. These collaborations should be applauded. Medical domain knowledge will assist computational scientists in developing better AI models. Also, the collective knowledge from physicians may help avoid the biases of the AI model and facilitate appropriate clinical decisions [58]. Along with the research on AI models used in fetal echocardiography, a framework for rigorous clinical assessment is also required [64]. Recall that the current testing metrics, such as sensitivity, accuracy, etc., are mainly designed for generic ML studies. Thus, metrics more suited for CHD detection can be further investigated.
Third, the adoption of AI-based CHD detection is still low. One limiting factor is the opaqueness of many AI models. The interpretation of AI algorithms is complex and operator-dependent, as “black box” models are created using raw data directly [14,15]. To overcome the “black box” issue and increase clinicians’ acceptance, steps should be taken to build clinically interpretable systems. For example, Moorfield’s Eye Hospital has developed a different approach. In its system, a segmentation network generates an intermediate tissue representation, allowing clinicians to visualize and quantify key areas of retinal pathology [65], increasing physicians’ confidence.
Our current study is a mini-review. Although we extensively searched the relevant literature, our selection could be biased, and our observations and opinions are bound by the papers selected.

6.2. Clinical Translation/Implementation

According to what was reported by Cardiovascular Business on 10 January 2025 [66], the US Food and Drug Agency (FDA) has cleared 161 AI products in cardiology up to the date of their publication. Our search found that one product, namely FETOLY-HEART, is relevant. FETOLY-HEART is an AI-assisted product based on ASE categorization [41] and can automatically detect certain heart views. Since this AI tool only received its clearance in August 2024, its clinical outcomes have not been reported in the literature.
Since most AI products for detecting CHDs have minimal or moderate risks, they are likely Class II medical devices and will only require FDA 510K clearance for regulatory approval. In other words, clinical trials may not be required to allow AI products for CHD detection to enter the market. That explains why very few clinical trials are known. Through a search of clinicaltrials.gov, we only found one clinical trial register, and the results have not been updated. Thus, we expect that more AI products of this kind will be cleared for clinical use in the near future.
Given the minimal or moderate risks involving in using AI products for detecting CHDs, their integration into the clinical workflow mainly resides in logistical issues, such as the integration into existing software platforms, e.g., picture archiving and communication system (PACS). Blezek et al. discussed various scenarios for deploying diagnostic AI products into the clinical workflow [67]. One scenario suggested by them may offer a plausible pathway for deploying an AI product aiming at CHD detection as summarized below.
If a cardiologist wishes to view the AI results, the cardiologist can launch an AI product by clicking a (software) button integrated into the PACS system. Then, the AI product opens and can automatically select one or multiple views to make a classification. Finally, the AI will show its results with some explanations. Cardiologists can see these AI results under a rapid review mode. The rapid review mode will include all selected images and the AI’s explanations during the decision-making process. Finally, the cardiologist can decide to accept or reject the AI’s results. Explained AI has been attempted by others [37,68]; these results are encouraging. Thus, its integration is possible and will enhance the usability of AI-based software for CHD detection.

7. Future Outlook and Closing Remarks

We consider AI-based CHD detection a leap forward for reducing diagnostic errors and improving clinical outcomes, particularly regarding sensitivity (see Section 4.5). If the work reviewed above is implemented into the clinical workflow, many studies will often require physicians to conduct many tasks manually (e.g., entering ultrasound measurements and selecting ultrasound images). Further developments can be focused on automation to streamline the workflow, including view selection, image enhancement, and quality assessment. Based on the studies reviewed, AI technologies for view selection, image enhancement, and quality assessment are available; thus, integrating these sub-components can offer a fully automated system for detecting CHDs in real time.
In addition to being used as a screening tool, another practical initial approach may be to set up AI platforms to identify specific CHDs that needs interventions soon after birth in a tertiary care setting (single ventricle, critical valvar stenosis, etc.). This would significantly enhance perinatal care and survival with CHDs.
Moreover, the availability of automated AI platforms for detecting CHDs could significantly benefit low-resource medical facilities, such as in rural areas and developing countries. As discussed above, although deploying diagnostic AI in low-resource areas, like ones for CHD detection, should be given priority, such deployment could be limited for financial reasons. Hence, the integration of AI into the traditional framework of tele-echocardiography [69] might bridge the gap for low-resource areas. In tele-echocardiography, an ultrasound technologist or radiologist first obtains cardiovascular ultrasound images from a patient. Then, these images are transmitted to a remote tertiary care setting for diagnosis, review, and image interpretation by a cardiologist. In this sense, an AI system can be deployed in a tertiary care setting and as a remote “cardiologist”, enabling rapid and accurate screening or decision-making. Indeed, the telemedicine [70] that has benefited low-resource areas may be further enhanced by the rapid development of diagnostic AI.
We conclude that AI-based CHD detection is feasible with the existing technologies. However, limitations restrict AI models’ potential for clinical applications and possibly dampen their clinical effectiveness. If these limitations are acknowledged and the process is approached cautiously, applying AI algorithms to detect and classify CHDs will be integral to future pediatric cardiac care.
Although the AI studies reviewed above have achieved encouraging results, there is still room for improvement. The improvement could be achieved if physiological data from the fetus and other medical test results were added to the AI-based predictive modeling. Recall that in [37], several medical testing results (e.g., maternal serum uric acid, glucose, etc.) were shown to be independent predictors of CHD. However, this is a single study, and others should verify their results. Also, the complex roles and associations of these biomarkers regarding CHD [71] should be further investigated. Furthermore, wearable biosensors can continuously collect physiological data about a fetus in the womb. In a recent publication, the American Heart Association’s Scientific Advisory Committee advocated using wearable technology to detect early signs of CHD [72]. Although wearable devices are routinely used to monitor maternal health, they have not been systematically used for detecting CHDs at the prenatal stage [73].

Author Contributions

Conceptualization, J.J. and J.V.; investigation, K.T.S., H.L., S.S.-Z., J.J. and J.V.; resources, all authors; data curation, all authors; writing—original draft preparation, K.T.S., H.L., S.S.-Z., J.V. and J.J.; writing—review and editing, all authors; visualization, K.T.S., H.L. and J.J.; supervision, M.H., J.V. and J.J.; project administration, M.H., J.V. and J.J.; funding acquisition, M.H., J.V. and J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research is partially funded through a research collaboration agreement with Corewell Health (formerly Spectrum Health Foundation).

Acknowledgments

We acknowledge discussions with colleagues at the Betz Congenital Health Center at Helen DeVos Children’s Hospital (Grand Rapids, Michigan, USA). Teng Yu from Michigan Technological University contributed to the preparation of Figure 3.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AUCArea of the curve
CHDCongenital Heart Defects
CNNConvolutional Neural Network
DLDeep Learning
GANGenerative Adversarial Network
GPUGraphic Processing Unit
HLHSHypoplastic left Heart Syndrome
HIPAAHealth Insurance Portability and Accountability Act
ML Machine Learning
NNNeural Network
PACSPicture archiving and communication system
PWDPulse Wave Doppler
RFRandom Forest
VSDVentricular Septal Defect
YOLOYou Only Look Once

Appendix A. Non-Technical Explanation of ML Models

Appendix A.1. Convolution Neural Network (CNN)

A CNN model is a multiple-layered computer algorithm seeking an image’s basic “elements” (also called features). Like the human visual cortex, a CNN model has a hierarchical structure, with simple features computed in early layers and more complex features found in deeper layers. Loosely speaking, a CNN model decomposes an image using basic “elements”, such as various edges, corners, shapes, etc., to capture its contents.

Appendix A.2. Generative Adversarial Network (GAN)

In the context of echocardiogram synthesis, a GAN model trains two competing neural networks to generate visually “realistic” echocardiograms. Realistic means synthesized echocardiograms are similar to “real” echocardiograms used for training. A GAN mode has two competing/adversarial networks: one generates new echocardiograms, and the other detects these synthesized echograms. Such competition allows the first network to improve its ability to generate synthesized echocardiograms that can be undetectable by the second network.

Appendix A.3. Reinforcement Learning (RL)

RL is one of the major ML methods in AI. As its name suggests, the central scheme of RL is to reward agents for their “good” behavior or actions, thereby reinforcing these good behaviors or actions. Collectively, a computer can learn how to navigate through a complex problem.

Appendix B. Key Phrases Used in the Literature Search

Keywords used in the literature search are shown in Table A1 below.
Table A1. Two lists of key phrases used in the literature search.
Table A1. Two lists of key phrases used in the literature search.
Person 1Personal 2
Deep learning in fetal echocardiography
Key anatomical structure detection in fetal
echocardiography
Standard plane recognition in fetal heart
Fetal heart standard views
Deep learning in fetal heart echocardiography
Fetal echocardiography using artificial intelligence
Fetal echocardiography
Artificial intelligence
Congenital heart disease
Deep learning
Machine learning
Structural heart disease detection
Congenital heart disease
detection
Heart Disease

References

  1. CDC. Congential Heart Defects Data and Statistics. Available online: https://www.cdc.gov/heart-defects/data/index.html#:~:text=Babies%20born%20with%20heart%20defects&text=Heart%20defects%20affect%20nearly%201,year%20in%20the%20United%20States.&text=The%20prevalence%20of%20some%20heart,is%20a%20ventricular%20septal%20defect (accessed on 21 February 2025).
  2. Plana, M.N.; Zamora, J.; Suresh, G.; Fernandez-Pineda, L.; Thangaratinam, S.; Ewer, A.K. Pulse oximetry screening for critical congenital heart defects. Cochrane Database Syst. Rev. 2018. [Google Scholar] [CrossRef] [PubMed]
  3. Zhang, Y.-F.; Zeng, X.-L.; Zhao, E.-F.; Lu, H.-W. Diagnostic Value of Fetal Echocardiography for Congenital Heart Disease: A Systematic Review and Meta-Analysis. Medicine 2015, 94, e1759. [Google Scholar] [CrossRef] [PubMed]
  4. Sklansky, M.; DeVore, G.R. Fetal Cardiac Screening. J. Ultrasound Med. 2016, 35, 679–681. [Google Scholar] [CrossRef] [PubMed]
  5. Sun, H.Y.; Proudfoot, J.A.; McCandless, R.T. Prenatal detection of critical cardiac outflow tract anomalies remains suboptimal despite revised obstetrical imaging guidelines. Congenit. Heart Dis. 2018, 13, 748–756. [Google Scholar] [CrossRef]
  6. van Nisselrooij, A.E.L.; Teunissen, A.K.K.; Clur, S.A.; Rozendaal, L.; Pajkrt, E.; Linskens, I.H.; Rammeloo, L.; van Lith, J.M.M.; Blom, N.A.; Haak, M.C. Why are congenital heart defects being missed? Ultrasound Obstet. Gynecol. 2020, 55, 747–757. [Google Scholar] [CrossRef]
  7. Wren, C.; Reinhardt, Z.; Khawaja, K. Twenty-year trends in diagnosis of life-threatening neonatal cardiovascular malformations. Arch. Dis. Child.-Fetal Neonatal Ed. 2008, 93, F33–F35. [Google Scholar] [CrossRef]
  8. Riede, F.T.; Wörner, C.; Dähnert, I.; Möckel, A.; Kostelka, M.; Schneider, P. Effectiveness of neonatal pulse oximetry screening for detection of critical congenital heart disease in daily clinical routine—Results from a prospective multicenter study. Eur. J. Pediatr. 2010, 169, 975–981. [Google Scholar] [CrossRef]
  9. Davis, A.; Billick, K.; Horton, K.; Jankowski, M.; Knoll, P.; Marshall, J.E.; Paloma, A.; Palma, R.; Adams, D.B. Artificial Intelligence and Echocardiography: A Primer for Cardiac Sonographers. J. Am. Soc. Echocardiogr. 2020, 33, 1061–1066. [Google Scholar] [CrossRef]
  10. Goecks, J.; Jalili, V.; Heiser, L.M.; Gray, J.W. How Machine Learning Will Transform Biomedicine. Cell 2020, 181, 92–101. [Google Scholar] [CrossRef]
  11. Abuelezz, I.; Hassan, A.; Jaber, B.A.; Sharique, M.; Abd-Alrazaq, A.; Househ, M.; Alam, T.; Shah, Z. Contribution of Artificial Intelligence in Pregnancy: A Scoping Review. Stud. Health Technol. Inf. 2022, 289, 333–336. [Google Scholar] [CrossRef]
  12. Islam, M.N.; Mustafina, S.N.; Mahmud, T.; Khan, N.I. Machine learning to predict pregnancy outcomes: A systematic review, synthesizing framework and future research agenda. BMC Pregnancy Childbirth 2022, 22, 348. [Google Scholar] [CrossRef]
  13. Mennickent, D.; Rodríguez, A.; Opazo, M.C.; Riedel, C.A.; Castro, E.; Eriz-Salinas, A.; Appel-Rubio, J.; Aguayo, C.; Damiano, A.E.; Guzmán-Gutiérrez, E.; et al. Machine learning applied in maternal and fetal health: A narrative review focused on pregnancy diseases and complications. Front. Endocrinol. 2023, 14, 1130139. [Google Scholar] [CrossRef]
  14. Alsharqi, M.; Woodward, W.J.; Mumith, J.A.; Markham, D.C.; Upton, R.; Leeson, P. Artificial intelligence and echocardiography. Echo Res. Pract. 2018, 5, R115–R125. [Google Scholar] [CrossRef]
  15. Barry, T.; Farina, J.M.; Chao, C.-J.; Ayoub, C.; Jeong, J.; Patel, B.N.; Banerjee, I.; Arsanjani, R. The Role of Artificial Intelligence in Echocardiography. J. Imaging 2023, 9, 50. [Google Scholar] [CrossRef] [PubMed]
  16. Qiao, S.; Pan, S.; Luo, G.; Pang, S.; Chen, T.; Singh, A.K.; Lv, Z. A Pseudo-Siamese Feature Fusion Generative Adversarial Network for Synthesizing High-Quality Fetal Four-Chamber Views. IEEE J. Biomed. Health Inform. 2023, 27, 1193–1204. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, X.; Yang, T.-Y.; Zhang, Y.-Y.; Liu, X.-W.; Zhang, Y.; Sun, L.; Gu, X.-Y.; Chen, Z.; Guo, Y.; Xue, C.; et al. Diagnosis of fetal total anomalous pulmonary venous connection based on the post-left atrium space ratio using artificial intelligence. Prenat. Diagn. 2022, 42, 1323–1331. [Google Scholar] [CrossRef]
  18. Yu, X.; Ma, L.; Wang, H.; Zhang, Y.; Du, H.; Xu, K.; Wang, L. Deep learning-based differentiation of ventricular septal defect from tetralogy of Fallot in fetal echocardiography images. Technol. Health Care 2024, 32, 457–464. [Google Scholar] [CrossRef]
  19. Sutarno, S.; Nurmaini, S.; Partan, R.U.; Sapitri, A.I.; Tutuko, B.; Naufal Rachmatullah, M.; Darmawahyuni, A.; Firdaus, F.; Bernolian, N.; Sulistiyo, D. FetalNet: Low-light fetal echocardiography enhancement and dense convolutional network classifier for improving heart defect prediction. Inform. Med. Unlocked 2022, 35, 101136. [Google Scholar] [CrossRef]
  20. Arnaout, R.; Curran, L.; Zhao, Y.; Levine, J.C.; Chinn, E.; Moon-Grady, A.J. An ensemble of neural networks provides expert-level prenatal detection of complex congenital heart disease. Nat. Med. 2021, 27, 882–891. [Google Scholar] [CrossRef]
  21. Li, X.; Zhang, H.; Yue, J.; Yin, L.; Li, W.; Ding, G.; Peng, B.; Xie, S. A multi-task deep learning approach for real-time view classification and quality assessment of echocardiographic images. Sci. Rep. 2024, 14, 20484. [Google Scholar] [CrossRef]
  22. Yang, X.; Huang, X.; Wei, C.; Yu, J.; Yu, X.; Dong, C.; Chen, J.; Chen, R.; Wu, X.; Yu, Z.; et al. An intelligent quantification system for fetal heart rhythm assessment: A multicenter prospective study. Heart Rhythm. 2024, 21, 600–609. [Google Scholar] [CrossRef] [PubMed]
  23. Dong, J.; Liu, S.; Liao, Y.; Wen, H.; Lei, B.; Li, S.; Wang, T. A Generic Quality Control Framework for Fetal Ultrasound Cardiac Four-Chamber Planes. IEEE J. Biomed. Health Inform. 2020, 24, 931–942. [Google Scholar] [CrossRef]
  24. Qiao, S.; Pang, S.; Luo, G.; Xie, P.; Yin, W.; Pan, S.; Lyu, Z. A progressive growing generative adversarial network composed of enhanced style-consistent modulation for fetal ultrasound four-chamber view editing synthesis. Eng. Appl. Artif. Intell. 2024, 133, 108438. [Google Scholar] [CrossRef]
  25. Truong, V.T.; Nguyen, B.P.; Nguyen-Vo, T.-H.; Mazur, W.; Chung, E.S.; Palmer, C.; Tretter, J.T.; Alsaied, T.; Pham, V.T.; Do, H.Q.; et al. Application of machine learning in screening for congenital heart diseases using fetal echocardiography. Int. J. Cardiovasc. Imaging 2022, 38, 1007–1015. [Google Scholar] [CrossRef] [PubMed]
  26. Wong, K.C.; Sinkovskaya, E.; Abuhamad, A.; Syeda-Mahmood, T. Multiview and Multiclass Image Segmentation Using Deep Learning in Fetal Echocardiography; SPIE: Bellingham, WA, USA, 2021; Volume 11597. [Google Scholar]
  27. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  28. Yan, L.; Ling, S.; Mao, R.; Xi, H.; Wang, F. A deep learning framework for identifying and segmenting three vessels in fetal heart ultrasound images. BioMedical Eng. OnLine 2024, 23, 39. [Google Scholar] [CrossRef]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W., Frangi, A., Eds.; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  30. Dozen, A.; Komatsu, M.; Sakai, A.; Komatsu, R.; Shozu, K.; Machino, H.; Yasutomi, S.; Arakaki, T.; Asada, K.; Kaneko, S.; et al. Image Segmentation of the Ventricular Septum in Fetal Cardiac Ultrasound Videos Based on Deep Learning Using Time-Series Information. Biomolecules 2020, 10, 1526. [Google Scholar] [CrossRef]
  31. Day, T.G.; Budd, S.; Tan, J.; Matthew, J.; Skelton, E.; Jowett, V.; Lloyd, D.; Gomez, A.; Hajnal, J.V.; Razavi, R.; et al. Prenatal diagnosis of hypoplastic left heart syndrome on ultrasound using artificial intelligence: How does performance compare to a current screening programme? Prenat. Diagn. 2024, 44, 717–724. [Google Scholar] [CrossRef]
  32. Yang, Y.; Wu, B.; Wu, H.; Xu, W.; Lyu, G.; Liu, P.; He, S. Classification of normal and abnormal fetal heart ultrasound images and identification of ventricular septal defects based on deep learning. J. Perinat. Med. 2023, 51, 1052–1058. [Google Scholar] [CrossRef]
  33. Li, F.; Li, P.; Liu, Z.; Liu, S.; Zeng, P.; Song, H.; Liu, P.; Lyu, G. Application of artificial intelligence in VSD prenatal diagnosis from fetal heart ultrasound images. BMC Pregnancy Childbirth 2024, 24, 758. [Google Scholar] [CrossRef]
  34. Wu, H.; Wu, B.; Lai, F.; Liu, P.; Lyu, G.; He, S.; Dai, J. Application of Artificial Intelligence in Anatomical Structure Recognition of Standard Section of Fetal Heart. Comput. Math. Methods Med. 2023, 2023, 5650378. [Google Scholar] [CrossRef]
  35. Magesh, S.; RajaKumar, P.S. Fetal Heart Disease Detection Via Deep Reg Network Based on Ultrasound Images. J. Appl. Eng. Technol. Sci. 2023, 5, 439–450. [Google Scholar] [CrossRef]
  36. Veronese, P.; Guariento, A.; Cattapan, C.; Fedrigo, M.; Gervasi, M.T.; Angelini, A.; Riva, A.; Vida, V. Prenatal Diagnosis and Fetopsy Validation of Complete Atrioventricular Septal Defects Using the Fetal Intelligent Navigation Echocardiography Method. Diagnostics 2023, 13, 456. [Google Scholar] [CrossRef] [PubMed]
  37. Qu, Y.; Deng, X.; Lin, S.; Han, F.; Chang, H.H.; Ou, Y.; Nie, Z.; Mai, J.; Wang, X.; Gao, X.; et al. Using Innovative Machine Learning Methods to Screen and Identify Predictors of Congenital Heart Diseases. Front. Cardiovasc. Med. 2022, 8, 797002. [Google Scholar] [CrossRef] [PubMed]
  38. Ma, M.; Sun, L.-H.; Chen, R.; Zhu, J.; Zhao, B. Artificial intelligence in fetal echocardiography: Recent advances and future prospects. IJC Heart Vasc. 2024, 53, 101380. [Google Scholar] [CrossRef]
  39. Pang, Y.; Zhao, X.; Xiang, T.-Z.; Zhang, L.; Lu, H. Zoom In and Out: A Mixed-scale Triplet Network for Camouflaged Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 2150–2160. [Google Scholar]
  40. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  41. ASE. Policy Statement on AI. Available online: https://www.asecho.org/ase-policy-statements/ (accessed on 20 February 2025).
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Sfakianakis, C.; Simantiris, G.; Tziritas, G. GUDU: Geometrically-constrained Ultrasound Data augmentation in U-Net for echocardiography semantic segmentation. Biomed. Signal Process. Control. 2023, 82, 104557. [Google Scholar] [CrossRef]
  44. Chen, S.-H.; Weng, K.-P.; Hsieh, K.-S.; Chen, Y.-H.; Shih, J.-H.; Li, W.-R.; Zhang, R.-Y.; Chen, Y.-C.; Tsai, W.-R.; Kao, T.-Y. Optimizing Object Detection Algorithms for Congenital Heart Diseases in Echocardiography: Exploring Bounding Box Sizes and Data Augmentation Techniques. Rev. Cardiovasc. Med. 2024, 25, 335. [Google Scholar] [CrossRef]
  45. Wang, Y.; Helminen, E.; Jiang, J. Building a virtual simulation platform for quasistatic breast ultrasound elastography using open source software: A preliminary investigation. Med. Phys. 2015, 42, 5453–5466. [Google Scholar] [CrossRef]
  46. Wang, Y.; Peng, B.; Jiang, J. Building an open-source simulation platform of acoustic radiation force-based breast elastography. Phys. Med. Biol. 2017, 62, 1949. [Google Scholar] [CrossRef]
  47. Peng, B.; Xian, Y.; Zhang, Q.; Jiang, J. Neural-network-based Motion Tracking for Breast Ultrasound Strain Elastography: An Initial Assessment of Performance and Feasibility. Ultrason. Imaging 2020, 42, 74–91. [Google Scholar] [CrossRef]
  48. Yurdem, B.; Kuzlu, M.; Gullu, M.K.; Catak, F.O.; Tabassum, M. Federated learning: Overview, strategies, applications, tools and future directions. Heliyon 2024, 10, e38137. [Google Scholar] [CrossRef]
  49. Goto, S.; Solanki, D.; John, J.E.; Yagi, R.; Homilius, M.; Ichihara, G.; Katsumata, Y.; Gaggin, H.K.; Itabashi, Y.; MacRae, C.A.; et al. Multinational Federated Learning Approach to Train ECG and Echocardiogram Models for Hypertrophic Cardiomyopathy Detection. Circulation 2022, 146, 755–769. [Google Scholar] [CrossRef]
  50. Mu, N.; Lyu, Z.; Rezaeitaleshmahalleh, M.; Tang, J.; Jiang, J. An attention residual u-net with differential preprocessing and geometric postprocessing: Learning how to segment vasculature including intracranial aneurysms. Med. Image Anal. 2023, 84, 102697. [Google Scholar] [CrossRef] [PubMed]
  51. Mu, N.; Lyu, Z.; Rezaeitaleshmahalleh, M.; Bonifas, C.; Gosnell, J.; Haw, M.; Vettukattil, J.; Jiang, J. S-Net: A multiple cross aggregation convolutional architecture for automatic segmentation of small/thin structures for cardiovascular applications. Front. Physiol. 2023, 14, 1209659. [Google Scholar] [CrossRef]
  52. Zhu, F.; Gao, Z.; Zhao, C.; Zhu, H.; Nan, J.; Tian, Y.; Dong, Y.; Jiang, J.; Feng, X.; Dai, N.; et al. A Deep Learning-based Method to Extract Lumen and Media-Adventitia in Intravascular Ultrasound Images. Ultrason. Imaging 2022, 44, 191–203. [Google Scholar] [CrossRef] [PubMed]
  53. Cui, S.; Mao, L.; Jiang, J.; Liu, C.; Xiong, S. Automatic Semantic Segmentation of Brain Gliomas from MRI Images Using a Deep Cascaded Neural Network. J. Healthc. Eng. 2018, 2018, 4940593. [Google Scholar] [CrossRef] [PubMed]
  54. Aldridge, N.; Pandya, P.; Rankin, J.; Miller, N.; Broughan, J.; Permalloo, N.; McHugh, A.; Stevens, S. Detection rates of a national fetal anomaly screening programme: A national cohort study. BJOG Int. J. Obstet. Gynaecol. 2023, 130, 51–58. [Google Scholar] [CrossRef]
  55. Copel, J.A.; Pilu, G.; Green, J.; Hobbins, J.C.; Kleinman, C.S. Fetal echocardiographic screening for congenital heart disease: The importance of the four-chamber view. Am. J. Obstet. Gynecol. 1987, 157, 648–655. [Google Scholar] [CrossRef]
  56. Buskens, E.; Grobbee, D.E.; Frohn-Mulder, I.M.E.; Stewart, P.A.; Juttmann, R.E.; Wladimiroff, J.W.; Hess, J. Efficacy of Routine Fetal Ultrasound Screening for Congenital Heart Disease in Normal Pregnancy. Circulation 1996, 94, 67–72. [Google Scholar] [CrossRef]
  57. Athalye, C.; van Nisselrooij, A.; Rizvi, S.; Haak, M.C.; Moon-Grady, A.J.; Arnaout, R. Deep-learning model for prenatal congenital heart disease screening generalizes to community setting and outperforms clinical detection. Ultrasound Obstet. Gynecol. 2024, 63, 44–52. [Google Scholar] [CrossRef]
  58. Liyanage, H.; Liaw, S.-T.; Jonnagaddala, J.; Schreiber, R.; Kuziemsky, C.; Terry, A.L.; de Lusignan, S. Artificial Intelligence in Primary Health Care: Perceptions, Issues, and Challenges. Prim. Health Care Inform. Work. Group Contrib. Yearb. Med. Inform. 2019, 28, 041–046. [Google Scholar] [CrossRef]
  59. Sullivan, H.R.; Schweikart, S.J. Are Current Tort Liability Doctrines Adequate for Addressing Injury Caused by AI? AMA J. Ethics 2019, 21, E160–E166. [Google Scholar] [CrossRef]
  60. Abdullah, Y.I.; Schuman, J.S.; Shabsigh, R.; Caplan, A.; Al-Aswad, L.A. Ethics of Artificial Intelligence in Medicine and Ophthalmology. Asia Pac. J. Ophthalmol. 2021, 10, 289–298. [Google Scholar] [CrossRef] [PubMed]
  61. Luxton, D. Should Watson be consulted for a second opinion? AMA J. Ethics 2019, 21, 131–137. [Google Scholar] [CrossRef]
  62. Nordling, L. A fairer way forward for AI in health care. Nature 2019, 573, S103. [Google Scholar] [CrossRef] [PubMed]
  63. Taksoee-Vester, C.A.; Mikolaj, K.; Bashir, Z.; Christensen, A.N.; Petersen, O.B.; Sundberg, K.; Feragen, A.; Svendsen, M.B.S.; Nielsen, M.; Tolsgaard, M.G. AI supported fetal echocardiography with quality assessment. Sci. Rep. 2024, 14, 5809. [Google Scholar] [CrossRef]
  64. Day, T.G.; Kainz, B.; Hajnal, J.; Razavi, R.; Simpson, J.M. Artificial intelligence, fetal echocardiography, and congenital heart disease. Prenat. Diagn. 2021, 41, 733–742. [Google Scholar] [CrossRef]
  65. Ting, D.S.W.; Pasquale, L.R.; Peng, L.; Campbell, J.P.; Lee, A.Y.; Raman, R.; Tan, G.S.W.; Schmetterer, L.; Keane, P.A.; Wong, T.Y. Artificial intelligence and deep learning in ophthalmology. Br. J. Ophthalmol. 2019, 103, 167–175. [Google Scholar] [CrossRef]
  66. Cardiovascular Business. FDA Has Now Cleared More than 1,000 AI Models, Including Many in Cardiology. Available online: https://cardiovascularbusiness.com/topics/artificial-intelligence/fda-has-cleared-more-1000-ai-algorithms-many-cardiology (accessed on 6 March 2025).
  67. Blezek, D.J.; Olson-Williams, L.; Missert, A.; Korfiatis, P. AI Integration in the Clinical Workflow. J. Digit. Imaging 2021, 34, 1435–1446. [Google Scholar] [CrossRef]
  68. Mu, N.; Rezaeitaleshmahalleh, M.; Lyu, Z.; Wang, M.; Tang, J.; Strother, C.M.; Gemmete, J.J.; Pandey, A.S.; Jiang, J. Can we explain machine learning-based prediction for rupture status assessments of intracranial aneurysms? Biomed. Phys. Eng. Express 2023, 9, 037001. [Google Scholar] [CrossRef]
  69. Satou, G.M.; Rheuban, K.; Alverson, D.; Lewin, M.; Mahnke, C.; Marcin, J.; Martin, G.R.; Mazur, L.S.; Sahn, D.J.; Shah, S.; et al. Telemedicine in Pediatric Cardiology: A Scientific Statement From the American Heart Association. Circulation 2017, 135, e648–e678. [Google Scholar] [CrossRef]
  70. Stagg, A.; Giglia, T.M.; Gardner, M.M.; Offit, B.F.; Fuller, K.M.; Natarajan, S.S.; Hehir, D.A.; Szwast, A.L.; Rome, J.J.; Ravishankar, C.; et al. Initial Experience with Telemedicine for Interstage Monitoring in Infants with Palliated Congenital Heart Disease. Pediatr. Cardiol. 2023, 44, 196–203. [Google Scholar] [CrossRef] [PubMed]
  71. Rajpal, S.; Alshawabkeh, L.; Opotowsky, A.R. Current Role of Blood and Urine Biomarkers in the Clinical Care of Adults with Congenital Heart Disease. Curr. Cardiol. Rep. 2017, 19, 50. [Google Scholar] [CrossRef] [PubMed]
  72. Tandon, A.; Avari Silva, J.N.; Bhatt, A.B.; Drummond, C.K.; Hill, A.C.; Paluch, A.E.; Waits, S.; Zablah, J.E.; Harris, K.C.; Council on Lifelong Congenital Heart Disease and Heart Health in the Young (Young Hearts); et al. Advancing Wearable Biosensors for Congenital Heart Disease: Patient and Clinician Perspectives: A Science Advisory From the American Heart Association. Circulation 2024, 149, e1134–e1142. [Google Scholar] [CrossRef]
  73. Alim, A.; Imtiaz, M.H. Wearable Sensors for the Monitoring of Maternal Health—A Systematic Review. Sensors 2023, 23, 2411. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A schematic diagram showing the categorization of AI methods. Some AI models have been adopted and refined for their use in the detection of CHDs in conjunction with fetal echocardiography.
Figure 1. A schematic diagram showing the categorization of AI methods. Some AI models have been adopted and refined for their use in the detection of CHDs in conjunction with fetal echocardiography.
Medicina 61 00561 g001
Figure 2. (a) An ultrasound image of a fetal heart in the four-chamber view and (b) an ultrasound image of a fetal heart in the aortic arch view. Both ultrasound images are overlaid with Color Doppler.
Figure 2. (a) An ultrasound image of a fetal heart in the four-chamber view and (b) an ultrasound image of a fetal heart in the aortic arch view. Both ultrasound images are overlaid with Color Doppler.
Medicina 61 00561 g002
Figure 3. (a) A frame of a four-chamber view, (b) a segmented fetal heart by ZoomNet, and (c) a detected fetal heart using YOLO (version 7).
Figure 3. (a) A frame of a four-chamber view, (b) a segmented fetal heart by ZoomNet, and (c) a detected fetal heart using YOLO (version 7).
Medicina 61 00561 g003
Table 1. A concise summary of all selected studies reviewed in Section 3. GAN is known as a generative adversarial network. Non-technical explanations of GAN and CNN can be found in Appendix A.
Table 1. A concise summary of all selected studies reviewed in Section 3. GAN is known as a generative adversarial network. Non-technical explanations of GAN and CNN can be found in Appendix A.
TitlesInput DataTypes of AI AlgorithmsApplicationClinical Relevance
A Pseudo-Siamese Feature Fusion Generative Adversarial Network for Synthesizing High-Quality Fetal Four-Chamber Views [16]Four-chamber view ultrasound imagesGAN: Pseudo-Siamese Feature Fusion Generative Adversarial Network (PSFFGAN)Synthesize echocardiogramsEnrich training database for designing better diagnostic CHD AI tools
Diagnosis of fetal total anomalous pulmonary venous connection based on the post-left atrium space ratio using artificial intelligence [17]Four-chamber view time-resolved ultrasound videosCNN: DeepLabv3+, FastFCN, PSPNet, and DenseASPP Segmentation, biometric measurement, and diagnosis of specific abnormalityDiagnosis of fetal total anomalous pulmonary venous connection
Deep learning-based differentiation of ventricular septal defect from tetralogy of Fallot in fetal echocardiography images [18]ultrasound imagesCNN: VGG19, ResNet50, NTS-Net, and WSDANDiagnosis of abnormalityDifferentiation of ventricular septal defect from tetralogy of Fallot
FetalNet: Low-light fetal echocardiography enhancement and dense convolutional network classifier for improving heart defect prediction [19]Four-chamber view ultrasound imagesCNN-based LLIE modelImage enhancement A DL-based image processing to improve AI-based CHD detection
An ensemble of neural networks provides expert-level prenatal detection of complex congenital heart disease [20]Ultrasound images with five standard viewsCNN: Ensemble NetworkDetection of cardiac views and classification of normal and abnormal heartsDifferentiate complex CHD from normal hearts
A multi-task deep learning approach for real-time view classification and quality assessment of echocardiographic images [21]Ultrasound images with multiple viewsA multi-task CNN model with four components: a backbone network, a neck network, a view classification branch, and a quality assessment branch Selection of high-quality viewsImprove automation for view selection in the clinical workflow
An intelligent quantification system for fetal heart rhythm assessment: A multicenter prospective study [22]Ultrasound Pulse Wave DopplerR-CNNCalculate fetal cardiac time intervalsAssess fetal rhythm and function
A Generic Quality Control Framework for Fetal Ultrasound Cardiac Four-Chamber Planes [23]Ultrasound images with multiple viewsVarying CNN modelsQuality assessment of ultrasound images Improve automation for image selection in the clinical workflow
A progressive growing generative adversarial network composed of enhanced style-consistent modulation for fetal ultrasound four-chamber view editing synthesis [24]Four-chamber view ultrasound imagesAn enhanced Generative Adversarial Network (GAN)Synthesize echocardiogramsEnrich training database for designing better diagnostic CHD AI tools
Application of Machine Learning in screening for congenital heart diseases using fetal echocardiography [25]Manually measured parametersRandom Forest Detecting CHDScreen
for CHDs
Multiview and multiclass image segmentation using deep learning in fetal echocardiography [26]Ultrasound images with multiple viewsV-Net [27]Heart structure segmentationImprove the automation for detecting CHDs
A deep learning framework for identifying and segmenting three vessels in fetal heart ultrasound images [28]Ultrasound images with three-vessel viewVarying U-Net [29] modelsSegmenting heart structure Improve the automation for detecting CHDs
Image segmentation of the ventricular septum in fetal cardiac ultrasound videos based on deep learning using time-series information [30]Four-chamber view ultrasound imagesCNN: Cropping–Segmentation–Calibration (CSC)Detection of ventricular septumFoundation for detecting VSD
Prenatal diagnosis of hypoplastic left heart syndrome on ultrasound using artificial intelligence: How does performance compare to a current screening programme [31]Four-chamber view ultrasound imagesResNet Detecting hypoplastic left heart syndrome Detect hypoplastic left heart syndrome
Classification of normal and abnormal fetal heart ultrasound images and identification of ventricular septal defects based on deep learning [32]Ultrasound images with five standard viewsYOLO (version 5) and other CNN models Classification of normal and abnormal fetal heartsScreen fetal CHDs
Application of artificial intelligence in VSD prenatal diagnosis from fetal heart ultrasound images [33]Ultrasound images with the four-chamber view and left ventricular outflow viewResNet-18, DenseNet, and MobileNet, Detect VSDAuto-detection of VSD
Application of Artificial Intelligence in Anatomical Structure Recognition of Standard Section of Fetal Heart [34]5 views of ultrasound imagesU-Y-Net derived from YOLO (version 5)View recognition of standard ultrasound viewsFoundation for detecting CHD
Fetal Heart Disease Detection Via Deep Reg Network Based on Ultrasound Images [35]Not specified AlexNet, ResNet-50
VGG-16, DenseNet, MobileNet, and RegNet
Characterize normal vs. abnormal fetal heartsScreen fetal CHDs
Prenatal Diagnosis and Fetopsy Validation of Complete Atrioventricular Septal Defects Using the Fetal Intelligent Navigation Echocardiography Method [36]Four-chamber view ultrasound imagesSpatial image correlation Ultrasound View Navigation Improve visualization of ventricular walls
Using Innovative Machine Learning Methods to Screen and Identify Predictors of Congenital Heart Diseases [37]Self-reported questionnaires and routine clinical laboratory test resultsExplainable Boosting MachineIdentify predictors of CHDsImprove the detection of fetal CHDs
Table 2. A summary of the characteristics used in eighteen (18) papers. ASE stands for the American Society of Echocardiography. N/P and N/A denote not provided and not applicable, respectively.
Table 2. A summary of the characteristics used in eighteen (18) papers. ASE stands for the American Society of Echocardiography. N/P and N/A denote not provided and not applicable, respectively.
Paper CenterVendorData
Acquisition
Protocol
Training
Data Size
ASE Category [41]
Qiao et al. [16]1N/PN/P~1000N/A
Wang et al. [17]11Minimal~300Assisted
Yu et al. [18]1N/PN/P~200Autonomous
Sutarno et al. [19]11N/P~500Assisted
Arnaout et al. [20]24Yes~100,000Autonomous
Li, et al. [21]14N/P~100,000Autonomous
Yang et al. [22] 1142Yes~10,000Autonomous
Dong et al. [23] 1N/PN/P~7000Assisted
Qiao et al. [24] 1N/PN/P~600N/A
Truong et al. [25] 11N/P~4000Autonomous
Wong et al. [26]N/PN/PN/P~300Assisted
Yan et al. [28]11Minimal~500Assisted
Dozen et al. [30]11Minimal~600Assisted
Day et al. [31]11N/P~10,000Autonomous
Yang et al. [32]13Minimal~1800Autonomous
Li, et al. [33]15N/P~1500Autonomous
Wu, et al. [34]15N/P~3400Assisted
Magesh, et al. [35]N/PN/PN/P~400Autonomous
1 Doppler spectra were used to detect irregular heart rhythm.
Table 3. A summary of the performance of four AI studies, which differentiate abnormal fetal hearts from normal fetal hearts. AUC, PPV, and NPV stand for the area under the curve, positive predictive value, and negative predictive value, respectively. N/P denotes not provided.
Table 3. A summary of the performance of four AI studies, which differentiate abnormal fetal hearts from normal fetal hearts. AUC, PPV, and NPV stand for the area under the curve, positive predictive value, and negative predictive value, respectively. N/P denotes not provided.
PaperAccuracyAUCSensitivitySpecificityPPVNPV
Arnaout et al. [20]~100%0.9995%96%20%100%
Truong et al. [25]88%0.9485%88%55%97%
Yang et al. [32]80%N/P90%N/P90%N/P
Magesh, et al. [35]97%0.9792%95%N/PN/P
Meta-analysis of clinical studies [3]N/P0.9968.5%99.8%N/PN/P
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Suha, K.T.; Lubenow, H.; Soria-Zurita, S.; Haw, M.; Vettukattil, J.; Jiang, J. The Artificial Intelligence-Enhanced Echocardiographic Detection of Congenital Heart Defects in the Fetus: A Mini-Review. Medicina 2025, 61, 561. https://doi.org/10.3390/medicina61040561

AMA Style

Suha KT, Lubenow H, Soria-Zurita S, Haw M, Vettukattil J, Jiang J. The Artificial Intelligence-Enhanced Echocardiographic Detection of Congenital Heart Defects in the Fetus: A Mini-Review. Medicina. 2025; 61(4):561. https://doi.org/10.3390/medicina61040561

Chicago/Turabian Style

Suha, Khadiza Tun, Hugh Lubenow, Stefania Soria-Zurita, Marcus Haw, Joseph Vettukattil, and Jingfeng Jiang. 2025. "The Artificial Intelligence-Enhanced Echocardiographic Detection of Congenital Heart Defects in the Fetus: A Mini-Review" Medicina 61, no. 4: 561. https://doi.org/10.3390/medicina61040561

APA Style

Suha, K. T., Lubenow, H., Soria-Zurita, S., Haw, M., Vettukattil, J., & Jiang, J. (2025). The Artificial Intelligence-Enhanced Echocardiographic Detection of Congenital Heart Defects in the Fetus: A Mini-Review. Medicina, 61(4), 561. https://doi.org/10.3390/medicina61040561

Article Metrics

Back to TopTop