Next Article in Journal
From Pollutant Removal to Renewable Energy: MoS2-Enhanced P25-Graphene Photocatalysts for Malathion Degradation and H2 Evolution
Previous Article in Journal
Rutting Behavior of Dual-Layer Asphalt Pavements Subjected to Variable Temperature
Previous Article in Special Issue
Fabrication of ZrO2 Armor Ceramics by 3D Printing Accompanied with Microwave Sintering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Cycle Fatigue Life Prediction of Additive Manufacturing Inconel 718 Alloy via Machine Learning

1
School of Aeronautics and Astronautics, Tianjin Sino-German University of Applied Sciences, Tianjin 300350, China
2
Wheel Rail Center, Tianjin Research Institute for Advanced Equipment, Tsinghua University, Tianjin 300300, China
3
School of Materials Science and Engineering, Tianjin University, Tianjin 300350, China
*
Author to whom correspondence should be addressed.
Materials 2025, 18(11), 2604; https://doi.org/10.3390/ma18112604
Submission received: 16 April 2025 / Revised: 14 May 2025 / Accepted: 22 May 2025 / Published: 3 June 2025
(This article belongs to the Special Issue High Temperature-Resistant Ceramics and Composites)

Abstract

:
This study established a machine learning framework to enhance the accuracy of very-high-cycle fatigue (VHCF) life prediction in selective laser melted Inconel 718 alloy by systematically comparing the use of generative adversarial networks (GANs) and variational auto-encoders (VAEs) for data augmentation. We quantified the influence of critical defect parameters (dimensions and stress amplitudes) extracted from fracture analyses on fatigue life and compared the performance of GANs versus VAEs in generating synthetic training data for three regression models (ANN, Random Forest, and SVR). The experimental fatigue data were augmented using both generative models, followed by hyperparameter optimization and rigorous validation against independent test sets. The results demonstrated that the GAN-generated data significantly improved the prediction metrics, with GAN-enhanced models achieving superior R2 scores (0.91–0.97 vs. 0.86 ± 0.87) and lower MAEs (1.13–1.62% vs. 2.00–2.64%) compared to the VAE-based approaches. This work not only establishes GANs as a breakthrough tool for AM fatigue prediction but also provides a transferable methodology for data-driven modeling of defect-dominated failure mechanisms in advanced materials.

1. Introduction

Additive manufacturing (AM) has revolutionized component fabrication through computer-aided design and layer-wise material deposition, offering advantages such as design flexibility, cost efficiency, and rapid prototyping for aerospace, automotive, and biomedical applications [1,2,3,4]. Among AM-processed materials, Inconel 718 alloys are widely used in critical components like turbine blades and nuclear structures due to their high-temperature resistance. However, fatigue failure caused by internal defects (e.g., porosity, unmelted particles, and lack-of-fusion voids) remains a major challenge, often leading to catastrophic economic losses [5,6,7]. These defects act as stress concentrators, significantly reducing the very-high-cycle fatigue (VHCF) life of AM components [6].
To address this issue, two dominant approaches have emerged: physics-based modeling and data-driven machine learning (ML). Classical physics models, including Mayer’s model Equation (1) [8,9], the Z-parameter model Equation (3) [10], and Wang’s model Equation (2) [11], correlate fatigue life with defect size and material hardness by treating inclusions as initial cracks. While these models provide mechanistic insights, their accuracy relies heavily on empirical parameter calibration and simplified assumptions about defect morphology [12]. Recent hybrid approaches, such as Jia-Le Fan et al.’s physics-informed neural network (PINN) [13], integrate physical equations (e.g., Z-parameter) into ML frameworks to augment limited experimental data. Nevertheless, such methods remain constrained by the inherent limitations of the underlying physical models.
[ σ ( a r e a ) 1 6 ] α N f = C   ( M a y e r s   m o d e l )
N f = 9 G K t h 2 2 E ( Δ σ Δ σ D R ) 2 a 0 + a 0 ( 1 n / 2 ) C Δ σ n β 1 n π n / 2 ( n / 2 1 )   ( W a n g s   m o d e l )
[ Y σ ( a r e a ) 1 6 D β ] α N f = C   ( Z   p a r a m e t e r   m o d e l )
In parallel, purely data-driven ML techniques—such as Random Forest (RF), XGBoost, and support vector regression (SVR)—have demonstrated promising results in fatigue life prediction by directly mapping defect characteristics (e.g., size, location, and morphology) to fatigue properties [14,15]. For instance, Y.W. Luo et al. [15] quantified the influence of pore size, density, and proximity to surfaces on the VHCF life of SLM Inconel 718 using experimental datasets, highlighting the critical role of defect statistics. However, ML models require large, high-quality datasets for robust training, which are often unavailable due to the time-intensive nature of fatigue testing.
The methods mentioned above, no matter it is the prediction of fatigue life based on physical models [8,9,10,11,12] or the training of machine learning models after augmenting data through physical models [13], both methods encounter difficulties in building fatigue datasets and are constrained by the parameters of the physical model. Therefore, to explore the relationship between fatigue life and other parameters not encompassed by the existing physical model, it is necessary to re-optimize the established physical model. However, this is a complex task despite the existence of various machine learning methods designed to study complex fatigue behavior involving multiple parameters [13,14,15,16].
The model training for machine learning is still limited to the construction of fatigue dataset. Deep Generative Models (DGMs) are a class of machine learning models [16,17,18] that aim to learn the probability distribution of data, thereby generating new samples that resemble the training data. DGMs have widely application in fields such as image synthesis [19], natural language processing [20], and the lifetime estimation of mechanical parts [21]. This paper introduces DGMs to tackle the challenge of insufficient data and to enhance the accuracy of machine learning models [22]. Specifically, the study employed generative adversarial networks (GANs) and variational auto-encoders (VAEs) for the purpose of data augmentation. GANs comprise a generator and a discriminator; the former produces realistic samples while the latter distinguishes between real and generated samples [23]. GAN variants include CTGAN [24], CopulaGAN [25], TGAN [26], and MedGAN [27,28]. Conversely, VAEs integrate the probability distribution into the potential space during training, producing samples that are more interpretable [29]. In this work, DGMs are introduced to establish a fatigue life model for machine learning based on an existing smaller dataset.

2. Materials and Methods

2.1. Original Dataset Construction

The material under investigation in this study is additively manufactured Inconel 718 that was produced via Selective Laser Melting (SLM). As a widely used nickel-based superalloy, Inconel 718 exhibits exceptional high-temperature performance and has been extensively applied in mission-critical aerospace components, power generation systems, and automotive turbochargers [30,31,32]. The fabrication process employed an EOS M290 SLM (Krailling, Germany) system equipped with a 400 W fiber laser, utilizing high-purity argon as the shielding gas to prevent oxidation. The feedstock material consisted of gas-atomized Inconel 718 powder with a spherical morphology and controlled particle size distribution (15–45 μm). Prior to manufacturing, the substrate underwent sequential surface preparation including abrasive grinding, ultrasonic cleaning in ethanol, and vacuum drying to ensure interfacial integrity.
Cylindrical specimens (10 mm diameter × 80 mm length) were fabricated using a bidirectional scanning strategy with 90° interlayer rotation to minimize anisotropic effects. The optimized SLM parameters were as follows: laser power = 260 W, scan speed = 1000 mm/s, hatch spacing = 0.11 mm, and layer thickness = 40 μm. The post-processing heat treatment followed a three-stage protocol: (1) solution treatment at 950–980 °C for 1 h with air cooling, (2) aging at 720 ± 5 °C for 8 h, and (3) furnace cooling at 50 °C/h to 620 ± 5 °C, followed by 8 h isothermal holding and final air cooling [33]. The chemical composition (Table 1) and microstructural characteristics (Figure 1).
Very-high-cycle fatigue (VHCF) testing was conducted under fully reversed loading (R = −1) using ultrasonic fatigue instrumentation, with compressed air cooling employed to maintain the specimen temperature during high-frequency cycling. Fractographic analysis via scanning electron microscopy (SEM) was used to quantify critical defect characteristics, including (1) critical inclusion size (√area), (2) subsurface defect depth, and (3) crack initiation zone ratio. The experimental dataset (Table 2) and corresponding fracture surface morphologies (Figure 1) were extracted from reference [34] with validation through comparative analysis.
The very-high-cycle fatigue life experimental data and scanning electron microscope (SEM) plots for SLM were obtained from the literature [34] as shown in Table 2. Fatigue fracture analysis results of specimens subjected to ultrasonic fatigue experiments at a stress ratio of −1 were obtained including stress amplitude (σa), fatigue life (Nf), defect diameter d, defect size (square root of inclusions area √area), and inclusion distance (the shortest distance from the center of the inclusions to the surface of the specimen. dinc) and are shown in Figure 1. This newly introduced parameter S i (ratio of crack source extension area to specimen area) and other parameters are shown in Table 2.

2.2. Generative Adversarial Network (GAN) Data Expansion

As one of the most representative generative models within Deep Generative Models (DGMs), this study utilized the TensorFlow Keras Sequential API to define the neural network architectures of both the generator and the discriminator. The discriminator accepts data as input and outputs the probability of its authenticity. The GAN model is constructed by integrating the aforementioned functions of the generator and discriminator. It is compiled utilizing the binary cross-entropy loss function and the Adam optimizer. Training of the GAN involves loop iterations, where initially, the discriminator is trained on real data, followed by generated data, to effectively distinguish between the two. The generator, accepting Gaussian-distributed noise Z, aims to produce increasingly realistic data to deceive the discriminator. Upon completion of training, the generator is employed to generate. The objective function for the newly developed generative adversarial network (GAN) targeting synthetic data generation is articulated in Equation (4).
m i n G m a x D V ( D , G ) = E X ~ P d a t a ( x ) [ l o g D ( x ) ] + E Z ~ P z ( z ) [ l o g ( 1 D ( G ( z ) ) ) ]
where G is the parameter of the generator, whose objective is to generate samples that look similar to the real data. D is the parameter of the discriminator. The goal of the discriminator is to give a high probability for the real data and a low probability for the generated data to discriminate whether the samples generated by the generator are real data or not. E X ~ P d a t a ( x ) [ l o g D ( x ) ] this term represents the samples x sampled from the real data distribution p_data(x), and then calculates the logarithm of the discriminator’s predicted probability for these real samples. The goal of the generator is to maximize the discriminator’s predicted probability for the generated samples as being real.
E Z ~ P z ( z ) [ l o g ( 1 D ( G ( z ) ) ) ]
This term represents the input distribution of the generator. p_z (z) represents the noise sampled in the generator, which is then fed into the generator to generate a sample, and then the generated sample is fed into the discriminator. Finally, the logarithm of the discriminator’s predicted probability of these generated samples is calculated. The goal of the generator is to maximize this term so that the generated samples are harder for the discriminator to recognize as generated rather than real. The structure of the GAN is shown in Figure 2.

2.3. Variational Auto-Encoder (VAE) Data Expansion

This study utilized TensorFlow’s Keras Sequential API to define the encoder and decoder architectures and compile the model using the Adam optimizer. The encoder processes input data through a series of dense layers to derive the mean and log variance, subsequently sampling the latent space based on these parameters. Subsequently, the decoder receives samples from the latent space to generate the output. A variational auto-encoder (VAE) model connects the encoder and decoder via a fully connected layer. Sampling of the latent space occurs through a dedicated sampling layer to facilitate new data generation. The sampling mechanism utilizes the mean and log variance to calculate the random bias within the Gaussian distribution. The VAE’s objective function comprises two components, the first being the reconstruction loss, with Mean Squared Error (MSE) serving as the reconstruction loss, as delineated in Equation (6). The second component is the KL (Kullback–Leibler) divergence, which measures the discrepancy between the distribution in the latent space and the standard normal distribution, thereby guiding the learned distribution to approximate the standard normal distribution. This objective function aims to minimize the cumulative sum of these two components. The overall objective function, formulated as the Evidence Lower Bound (ELBO), is defined in Equation (7). The aim is to optimize the ELBO for improved data representation and to ensure meaningful latent variable distributions via KL divergence constraints.
M S E = 1 N i = 1 N ( x i x ^ i ) 2
E L B O = 1 N i = 1 N ( x i x ^ i ) 2 + 1 2 i = 1 d ( 1 + l o g ( σ i 2 ) μ i 2 σ i 2 ) )
where 1 N i = 1 N ( x i x ^ i ) 2 is the reconstruction error calculated from the mean square error, N is the sample size, x i is the original data, and x ^ i is the ta generated by the decoder, 1 2 i = 1 d ( 1 + l o g ( σ i 2 ) μ i 2 σ i 2 ) ) is the KL scatter, d is the dimension of latent variables, and μ i and σ i are the mean and variance of the latent variables, respectively. The VAE structure is shown in Figure 3.

3. Development of Machine Learning Models

3.1. Data Usability

Utilizing the aforementioned methods, the GAN and VAE were expanded to generate 150 datasets each, as illustrated in Figure 4 and Figure 5. Prior to training the ML model, it is crucial to assess the consistency and reliability of the extended data. To validate the usability of the generated data and gain a deeper understanding of how the characteristics of the defect distribution in the specimen affect fatigue life, the Pearson correlation coefficient (PCC) was employed to quantify the relationship between individual defect characteristics and fatigue life.
P = ( X i X ¯ ) ( Y i Y ¯ ) ( X i X ¯ ) 2 ( Y i Y ¯ ) 2
where Xi and Yi are the values of two characteristic variables, and X ¯ and Y ¯ are the means of two characteristic variables.
The original and extended datasets were imported to calculate the Pearson correlation coefficients, and PCC heat maps between the fatigue life and the relevant parameters for each dataset were obtained, as shown in Figure 6. The similarity between the extended datasets in Figure 6b,c and the original dataset in Figure 6a demonstrates the reliability of the data generated by GAN and VAE. It is reasonable to build the expanded dataset using these two generative models.

3.2. Data Preprocessing

With the aforementioned methods, 150 sets of datasets were obtained using the GAN and VAE, each comprising five variables: σa/Mpa, dinc/μm, and √area/μm, where √area is derived from the generated d. Additionally, the ratio of the extended area to the specimen’s area Si and Nf/cycle have different dimensions and units, affecting the uniformity and potentially causing escalated training duration and convergence challenges. Consequently, preprocessing of the dataset prior to its utilization in the ML model for training is imperative. This process aims to mitigate dimensional complexity and redundancy within the data. The normalization in Equation (9) was employed for this purpose. In addition, the logarithm of fatigue life was utilized for both model training and testing, as depicted in Equation (9).
X n o r m = X X m i n X m a x X m i n
where Xnorm, X, Xmin, and Xmax represent the normalized, original, minimum, and maximum values of the variable X, respectively.
N l o g = l o g N f
where Nlog and Nf denote the logarithmic and original values of the fatigue life, respectively.
It is imperative to ensure that the established machine learning models possess both reliability and generalization capabilities. The extended dataset was divided in a 7:3 ratio between the training and validation sets, with the original data serving as the test set and was not included in the training process. The training set is utilized for the iterative training of the model, aiming to develop a fatigue life prediction model, while the validation set facilitates hyperparameter optimization to select the optimal hyperparameter. Furthermore, to mitigate the impact of data randomness on the prediction outcomes, 5-fold cross-validation is implemented on the training data. Finally, the original data excluded from model training is fed into the trained model for prediction, enabling the evaluation of the model’s prediction performance.

3.3. Machine Learning Models

Following data expansion and preprocessing, the training set is utilized for ML model training. Subsequently, upon hyperparameter optimization, the optimal ML models are obtained, and the R2 value of the validation set within the training set is computed to assess the fit of the ML models. This is followed by fatigue life prediction using the optimal ML models, with R2 and MAE serving as metrics to evaluate the model’s generalization ability and accuracy. The process is illustrated in Figure 7.

3.3.1. Support Vector Regression Model

Following data expansion and preprocessing, the data are input into the Support Vector Regression (SVR) model for training, where the objective is to identify an optimal hyperplane that minimizes the model’s prediction error on the training data while tolerating a certain level of error [34]. The training is an optimization problem, wherein the penalty parameter (C), the kernel function parameters (γ), and the tolerance (ε) are fine-tuned using methods like cross-validation. This phase aims to evaluate the model’s performance on validation data and select the optimal hyperparameter configuration to prevent overfitting or underfitting. The available kernel functions include Linear Kernel (linear), Radial Basis Function (RBF), Polynomial Kernel (poly), and Sigmoid Kernel (sigmoid). The regression function is presented in Equation (11) [35].
f ( x ) = i = 1 n ( α i α i * ) · K ( x i , x j ) + b , α i , α i * 0 , C
where α i and α i * are the Lagrange multipliers, b is the bias, and C is the penalty parameter. K ( x i , x j ) denotes the kernel function distribution [36] with radial kernel (rbf), linear kernel, and polynomial kernel [36].
In the SVR structure, σa (MPa), dinc (μm), √area (μm), and Si are used as the x input parameters, while life is employed as the f(x) output dependent variable.

3.3.2. Random Forest Model

The training phase of the RF regression model constructs multiple uncorrelated decision trees. In this study, σa/Mpa, dinc/μm, √area/μm, and Si were used as input parameters, and the lifetime was used as the y R F output dependent variable. The training dataset is divided into multiple sub-datasets. Different sub-datasets are used to train different decision trees, denoted as DTk, where k is a natural number. For each decision tree, the predicted output vector is calculated. Finally, the result is obtained by averaging the predicted output vectors of all decision trees, as shown in Equation (12) [37].
y R F = 1 k p = 1 k y p p r e
There are four important parameters in the RF model, including nestimators, min_samples_split, min_samples_leaf, and max_depth. Grid hyperparameter optimization is used for parameter optimization of the RF parameters.

3.3.3. Artificial Neural Network

Artificial neural networks (ANNs) are composed of numerous neurons, which include an input layer for predicting outcomes. The input layer receives the input data, with each node representing an input feature. The hidden layers facilitate the learning of complex data relationships. There may be multiple hidden layers, each composed of numerous nodes. The output layer, where each node corresponds to an output category or value, produces the predicted results [38]. The neuron is the fundamental unit of an ANN. The input data passes through the input layer and propagates through the hidden layers in sequence, eventually producing results in the output layer. Each node in the hidden layers receives input values from the nodes of the previous layer. These input values are then processed by weights, biases, and activation functions to generate an output, which is finally transmitted to the next node. The regression function yi of the i-th neuron is expressed by the following Equation (13) [39,40,41,42].
y i = f i ω i , j x j t i
where ω i ,j is the weight of the neural network, xj denotes the input of the current neuron node, fj is the activation function, and ti is the threshold parameter.
The general idea is to use a training dataset and continuously adjust the weights through backpropagation until the model is able to accurately predict the input data. Commonly used optimization algorithms include gradient descent, stochastic gradient descent, etc. The weights are updated at each iteration to minimize the error function. The error function is given in Equation (14) [39,40,41,42].
E = 1 2 i = 1 n y i y i t r u e
where E is the error function, n is the number of samples in the dataset, y i is the actual output of the regression function, and y i t r u e is the true value of the fatigue life of the experimental fatigue test. In this study, a multilayer perceptron (MLP) artificial neural network regressor was used.

3.3.4. Model Evaluation Criteria

To evaluate the accuracy of the ML model in predicting fatigue life, the coefficient of determination R2 and the mean absolute error (MAE) were used. R2 is an accurate representation of the degree of fit of the regression model, with values close to 1 reflecting a high accuracy of the model and good fit. The MAE calculates the mean absolute error of the fatigue predicted life and the fatigue experimental life, with values close to 0 indicating high model accuracy. The formulas of R 2 and MAE are described in Equations (15) and (16), respectively.
R 2 ( y , y p r e ) = 1 i = 1 n ( y i t r u e y i p r e ) 2 i = 1 n ( y i t r u e y m e a n ) 2
M A E ( y , y p r e ) = 1 n i = 1 n y i t r u e y i p r e
where y i t r u e denotes the i-th original data, y i p r e denotes the i-th predicted data, and y m a e is the average of the original data.

3.4. Hyperparameter Optimization

The machine learning model training process adjusts the hyperparameters of the model to improve performance and generalization. Hyperparameters are parameters that are set before the model is trained and their values cannot be learned through the training process and need to be adjusted manually. The choice of hyperparameters is crucial for the performance and generalization ability of the model. In order to accurately assess model performance with different hyperparameter settings, cross-validation is often used. The training data are divided into training and validation sets. The training set is used to iteratively train the model, while the validation set assesses the quality of the model’s training. Prior to the optimization process, it is necessary to set the range of values for the hyperparameters and parameters, as well as to select the method for optimizing the hyperparameters, e.g., lattice search [14], stochastic search, and Bayesian optimization [24]. The optimization set for each ML model is described in Table 3.
For the support vector regression model, three hyperparameters C, ε, γ and kernel function selection were optimized. These hyperparameters are crucial for the accuracy and generalization ability of the model. In this study, the penalty parameter C takes 15 values, the ε and the coefficients γ of the kernel function take 5 values in the range of C ϵ [10, 20, ……140, 150], ε ϵ [0.001, 0.01, 0.1, 1, 10], and γ ϵ [0.001, 0.01, 0.1, 1, 10].
Four parameters were optimized for the RF model, namely the total number of trees (n_estimators ϵ [50, 100, 150, 200, 500]), minimum number of samples for node splits (min_samples_split ϵ [2, 4, 6, 8, 10, 16]), minimum number of samples for leaf nodes (min_samples_leaf ϵ [1, 2, 4]), and the maximum depth of the tree (max_depth ϵ [None, 10, 20, 30, 40, 50]).
For the ANN model, the following four parameters were identified as key hyperparameters of the neural network model. Hidden_layer_sizes are defined as a list of tuples, where each tuple specifies the number and size of the hidden layers. For instance, (20) signifies a single hidden layer comprising 20 neurons. Utilizing list comprehension, configurations such as [(i, j, k)] indicate attempts at varying hidden layer sizes with respective neuron counts. The ranges for these configurations are as follows: i ϵ [20, 30...70, 80] for layer sizes, and j, k ϵ [0, 10, 20, 40] and [0, 10, 15, 20] for neuron counts and subsequent layers, respectively. The activation parameter specifies the activation function, which introduces non-linearity into each hidden layer, with the options ‘identity’ (17), ‘RELU’ (18), ‘sigmoid’ (19), and ‘tanh’ (20). The solver parameter was the chosen optimizer for the model, with options such as ADAM, LBFGS, and SGD. Lastly, max_iter defines the maximum number of iterations or training rounds for the model, after which, training will cease regardless of convergence status.
Identity :   ( f ( x ) = x )
RELU :   ( f ( x ) = m a x ( 0 , x ) )
Sigmoid :   ( f ( x ) = 1 / ( 1 + e x p ( x ) ) )
Tan h :   ( f ( x ) = e x e x e x + e x )

4. Results and Discussion

Three ML models, SVR, RF, and ANN, constructed based on deep generative models were developed to predict the fatigue life of Inconel 718 using the two extended datasets mentioned in Section 2.2 and Section 2.3. The initial prediction accuracies of the ML models on the training set were compared. Subsequently, the prediction results of the different models were juxtaposed with the original data to evaluate their performances. These models are denoted by a combination of the deep learning model and machine learning model sizes used for data augmentation in Table 4 and Table 5; for instance, GAN-SVR denotes the SVR model trained on the GAN extended dataset.

4.1. Parameter Configuration and Training Set Performance of the ML Model

In this study, the hyperparameters of the GAN-SVR model are optimally configured for the kernel function using the radial basis function (RBF) (21):
K ( x i , x j ) = e x p ( γ x i x j 2 )
where γ is the nuclear parameter in the optimal configuration of hyperparameters for GAN-SVR.
The validation set performance, with each calculated R2 value, is listed in Table 6. Based on the provided graph illustrating the fatigue life validation set prediction results of the ML model and R2, the optimized model can be obtained. The GAN-ANN model (R2 = 0.965) exhibited superior performance compared to the GAN-SVR model (R2 = 0.965) and GAN-RF model (R2 = 0.965). Meanwhile, the VAE-RF model (R2 = 0.983) exhibited superior performance compared to the VAE-ANN model (R2 = 0.969) and VAE-SVR model (R2 = 0.979).
A comparison of the results of the fatigue life calculated by the optimized ML model (with tuned hyperparameters and parameters) and the experimental results of the validation set is shown in Figure 8.

4.2. Fatigue Life Prediction

The validation set model and the hyperparameter-optimized ML model were compared and finally the optimized model was used to predict the fatigue life using the original dataset to evaluate the generalization ability of the model. The prediction accuracies of the GAN-ML models and VAE-ML models are shown in Figure 9 and Figure 10. The R2 and MAE are presented in Table 7. Based on the values of R2, the GAN-ANN model after GAN data enhancement performed worse (R2 = 0.919, MAE = 1.62%) than the GAN-SVR model (R2 = 0.929, MAE = 1.46%) and GAN-RF model (R2 = 0.975, MAE = 1.13%). The VAE-ANN model based on enhanced VAE data performed worse (R2 = 0.856, MAE = 2.06%) than the VAE-SVR model (R2 = 0.917, MAE = 1.61%) and VAE-RF model (R2 = 0.891, MAE = 1.57%). It can be observed from Figure 10 that the data predicted by the VAE-ML model fell outside the error range in the higher cycle ranges, clearly indicating that the model performed poorly when predicting data in these higher cycle ranges.
The analysis indicated that, in the validation set, the GAN-based machine learning models (GAN-ML) demonstrated lower performance compared to the VAE-based machine learning models (VAE-ML). However, in the test set, the GAN-ML models significantly outperformed the VAE-ML models. The accuracy of the GAN-ML models in the test set was not only higher than that of the VAE-ML models but was also comparable to their performance in the validation set. This consistent performance indicates the superior generalization capabilities of the GAN-ML models. In contrast, although the VAE-ML models showed higher accuracy in the validation set, they were unable to maintain this advantage in the test set. These data suggest that GAN-ML models are better at handling unseen data, reflecting their enhanced generalization capabilities compared to VAE-ML models. While the GAN-RF models exhibited significant underfitting and deliver inferior performance in the validation set, they nevertheless attained a higher R2 score with fewer training sets, which, in turn, makes GAN-ANN and GAN-SVR models appear more reliable by comparison.

5. Conclusions

This study proposes an extensible fatigue life prediction framework for high-temperature nickel-based alloys (demonstrated on Inconel 718) by systematically integrating defect characterization and generative modeling. Fatigue data were derived from SEM-based fracture analysis, defect size/location measurements, and a novel parameter—the fatigue crack ductile zone-to-specimen area ratio—to quantify crack growth behavior. To address data limitations, deep generative models (DGMs), including GANs and VAEs, were employed for data augmentation, followed by Pearson’s correlation analysis to validate parameter–life relationships and ensure the physical consistency of the synthesized datasets. The framework combines advanced experimental characterization, probabilistic data generation, and statistical validation to enhance predictive accuracy in fatigue life assessment.
The volume and quality of the initial dataset critically influence the performance of GAN and VAE models in fatigue life prediction, as observed in this study. A limited data volume restricts the generative models’ ability to capture the full statistical distribution of defect characteristics and loading conditions, leading to suboptimal synthetic data diversity. For instance, the GAN-RF model exhibited underfitting, likely due to insufficient coverage of defect distance interactions in the original dataset. Data quality issues, such as imbalanced defect size ranges or measurement uncertainties in SEM-derived parameters propagate into the generated data, amplifying biases in downstream ML predictions, as evidenced by the VAE-ANN’s overfitting (training MAE = 1.62% vs. test MAE = 2.46%). Future work should prioritize expanding the experimental dataset with controlled defect geometries and inclusion distributions while integrating uncertainty quantification to decouple measurement noise from true physical trends.
The comparative performance analysis of machine learning (ML) models integrated with generative adversarial network (GAN) and variational autoencoder (VAE) frameworks revealed distinct trends: the GAN-based models consistently outperformed their VAE counterparts, with GAN-RF achieving the highest predictive accuracy (R2 = 0.975) and minimal error (MAE = 1.13%), followed by GAN-SVR (R2 = 0.929, MAE = 1.46%) and GAN-ANN (R2 = 0.919, MAE = 1.62%). In contrast, the VAE-based models exhibited comparatively lower performance, with VAE-SVR (R2 = 0.865, MAE = 2.00%), VAE-RF (R2 = 0.879, MAE = 2.01%), and VAE-ANN (R2 = 0.861, MAE = 2.46%) demonstrating reduced generalization capability. These results highlight the superior robustness of GAN-enhanced models in fatigue life prediction, particularly GAN-RF, which delivered the highest R2 and lowest MAE, while the VAE-driven models exhibited larger error margins, suggesting limitations in their ability to generalize across unseen datasets.
The results showed that the VAE-ML models outperformed the GAN-ML models on the validation set, but on the test set, the VAE-ANN (validation set R2 = 0.923 vs. test set R2 = 0.861) and VAE-RF (validation set R2 = 0.968 VS test set R2 = 0.879) models showed overfitting. Overfitting occurs when a model achieves high performance on training data but significantly underperforms on new data, primarily due to excessively capturing noise or localized patterns in the data, thereby weakening its generalization capability. The GAN-ML models demonstrated better performance than the VAE-ML models in terms of R2 and MAE, indicating a better fit.
The GAN-RF (test set R2 = 0.864) models experienced underfitting. Underfitting occurs when a model exhibits poor performance on both training and test data, typically stemming from an oversimplified architecture or insufficient learning capacity, which prevents it from capturing fundamental patterns in the dataset. The GAN-SVR, GAN-ANN, and VAE-SVR models might be better for predicting fatigue life.
Furthermore, the purely data-driven models employed in this study lack physics-constrained interpretability, limiting their ability to align predictions with established mechanical principles or to generalize across untested operational conditions. Therefore, developing hybrid physics-informed ML frameworks that explicitly couple defect propagation laws with data-driven models could improve interpretability.

Author Contributions

Z.S.: Conceptualization, Data curation, Software, Validation, Investigation, Writing—original draft. J.P.: Formal analysis, Conceptualization, Funding acquisition. Writing—Reviewing and Editing. L.Z.: Data curation, Formal analysis. C.D.: Conceptualization, Investigation. Y.Z.: Conceptualization, Investigation. Q.G.: Data curation, Formal analysis. A.Z.: Conceptualization, Investigation. All authors have read and agreed to the published version of the manuscript.

Funding

The authors thank the Science & Technology Development Fund of Tianjin Education Commission for Higher Education (2021KJ098).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Gong, G.; Ye, J.; Chi, Y.; Zhao, Z.; Wang, Z.; Xia, G.; Du, X.; Tian, H.; Yu, H.; Chen, C. Research status of laser additive manufacturing for metal: A review. J. Mater. Res. Technol. 2021, 15, 855–884. [Google Scholar] [CrossRef]
  2. Armstrong, M.; Mehrabi, H.; Naveed, N. An overview of modern metal additive manufacturing technology. J. Manuf. Process. 2022, 84, 1001–1029. [Google Scholar] [CrossRef]
  3. Lu, J.; Zhuo, L. Additive manufacturing of titanium alloys via selective laser melting: Fabrication, microstructure, post-processing, performance and prospect. Int. J. Refract. Met. Hard Mater. 2023, 111, 106110. [Google Scholar] [CrossRef]
  4. Kanishka, K.; Acherjee, B. Revolutionizing manufacturing: A comprehensive overview of additive manufacturing processes, materials, developments, and challenges. J. Manuf. Process. 2023, 107, 574–619. [Google Scholar] [CrossRef]
  5. Qin, Z.; Li, B.; Chen, C.; Chen, T.; Chen, R.; Zhang, H.; Xue, H.; Yao, C.; Tan, L. Crack initiation mechanisms and life prediction of GH4169 superalloy in the high cycle and very high cycle fatigue regime. cycle and very high cycle fatigue regime. J. Mater. Res. Technol. 2023, 26, 720–736. [Google Scholar] [CrossRef]
  6. Derrick, C.; Fatemi, A. Correlations of fatigue strength of additively manufactured metals with hardness and defect size. Int. J. Fatigue 2022, 162, 106920. [Google Scholar] [CrossRef]
  7. Sanaei, N.; Fatemi, A. Defects in additive manufactured metals and their effect on fatigue performance: A state-of-the-art review. Prog. Mater. Sci. 2021, 117, 100724. [Google Scholar] [CrossRef]
  8. Mayer, H.; Haydn, W.; Schuller, R.; Issler, S.; Furtner, B.; Bacher-Höchst, M. Very high cycle fatigue properties of bainitic high carbon-chromium steel. Int. J. Fatigue 2009, 31, 242–249. [Google Scholar] [CrossRef]
  9. Mayer, H.; Haydn, W.; Schuller, R.; Issler, S.; Bacher-Höchst, M. Very high cycle fatigue properties of bainitic high carbon–chromium steel under variable amplitude conditions. Int. J. Fatigue 2009, 31, 1300–1308. [Google Scholar] [CrossRef]
  10. Zhu, M.-L.; Jin, L.; Xuan, F.-Z. Fatigue life and mechanistic modeling of interior micro-defect induced cracking in high cycle and very high cycle regimes. Acta Mater. 2018, 157, 259–275. [Google Scholar] [CrossRef]
  11. Wang, Q.Y.; Bathias, C.; Kawagoishi, N.; Chen, Q. Effect of inclusion on subsurface crack initiation and gigacycle fatigue strength. Int. J. Fatigue 2002, 24, 1269–1274. [Google Scholar] [CrossRef]
  12. Murakami, Y.; Kodama, S.; Konuma, S. Quantitative evaluation of effects of non-metallic inclusions on fatigue strength of high strength steels. I: Basic fatigue mechanism and evaluation of correlation between the fatigue fracture stress and the size and location of non-metallic inclusions. Int. J. Fatigue 1989, 11, 291–298. [Google Scholar] [CrossRef]
  13. Fan, J.-L.; Zhu, G.; Zhu, M.-L.; Xuan, F.-Z. A data-physics integrated approach to life prediction in very high cycle fatigue regime. Int. J. Fatigue 2023, 176, 107917. [Google Scholar] [CrossRef]
  14. Peng, X.; Wu, S.; Qian, W.; Bao, J.; Hu, Y.; Zhan, Z.; Guo, G.; Withers, P.J. The potency of defects on fatigue of additively manufactured metals. Int. J. Mech. Sci. 2022, 221, 107185. [Google Scholar] [CrossRef]
  15. Luo, Y.W.; Zhang, B.; Feng, X.; Song, Z.M.; Qi, X.B.; Li, C.P.; Chen, G.F.; Zhang, G.P. Pore-affected fatigue life scattering and prediction of additively manufactured Inconel 718: An investigation based on miniature specimen testing and machine learning approach. Mater. Sci. Eng. A 2021, 802, 140693. [Google Scholar] [CrossRef]
  16. Liu, Y.-K.; Fan, J.-L.; Zhu, G.; Zhu, M.-L.; Xuan, F.-Z. Data-driven approach to very high cycle fatigue life prediction. Eng. Fract. Mech. 2023, 292, 109630. [Google Scholar] [CrossRef]
  17. Niu, S.; Peng, Y.; Li, B.; Wang, X. A transformed-feature-space data augmentation method for defect segmentation. Comput. Ind. 2023, 147, 103860. [Google Scholar] [CrossRef]
  18. Shi, L.; Yang, X.; Chang, X.; Wu, J.; Sun, H. An improved density peaks clustering algorithm based on k nearest neighbors and turning point for evaluating the severity of railway accidents. Reliab. Eng. Syst. Saf. 2023, 233, 109132. [Google Scholar] [CrossRef]
  19. Zhan, Z.; Hu, W.; Meng, Q. Data-driven fatigue life prediction in additive manufactured titanium alloy: A damage mechanics based machine learning framework. Eng. Fract. Mech. 2021, 252, 107850. [Google Scholar] [CrossRef]
  20. Tran, L.; Yin, X.; Liu, X. Representation Learning by Rotating Your Faces. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 3007–3021. [Google Scholar] [CrossRef]
  21. Qiao, T.; Zhang, J.; Xu, D.; Tao, D. MirrorGAN: Learning Text-To-Image Generation by Redescription. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15 June 2019; pp. 1505–1514. [Google Scholar]
  22. Yu, J.; Guo, Z. Remaining useful life prediction of planet bearings based on conditional deep recurrent generative adversarial network and action discovery. J. Mech. Sci. Technol. 2021, 35, 21–30. [Google Scholar] [CrossRef]
  23. Wang, X.; Xu, L.; Zhao, L.; Ren, W.; Li, Q.; Han, Y. Machine learning method for estimating the defect-related mechanical properties of additive manufactured alloys. Eng. Fract. Mech. 2023, 291, 109559. [Google Scholar] [CrossRef]
  24. He, G.; Zhao, Y.; Yan, C. Application of tabular data synthesis using generative adversarial networks on machine learning-based multiaxial fatigue life prediction. Int. J. Press. Vessel. Pip. 2022, 199, 104779. [Google Scholar] [CrossRef]
  25. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Bangkok, Thailand, 23–27 November 2020; MIT Press: Montreal, QC, Canada, 2014; Volume 2, pp. 2672–2680. [Google Scholar]
  26. Xu, X.; Zhang, T.; Su, Y. Temporal variations and trend of ground-level ozone based on long-term measurements in Windsor, Canada. Atmos. Chem. Phys. 2019, 19, 7335–7345. [Google Scholar] [CrossRef]
  27. Patki, R.N.; Veeramachaneni, K. The Synthetic Data Vault, 2016. Available online: https://ieeexplore.ieee.org/document/7796926 (accessed on 21 May 2022).
  28. Choi, E.; Biswal, S.; Malin, B.; Duke, J.; Stewart, W.; Sun, J. Generating Multi-Label Discrete Patient Records Using Generative Adversarial Networks. arXiv 2017, arXiv:1703.06490. [Google Scholar]
  29. Xu, L.; Veeramachaneni, K. Synthesising Tabular Data Using Generative Adversarial Networks. arXiv 2018, arXiv:1811.11264. [Google Scholar]
  30. Kingma, D.P. Max Welling. Auto-Encoding Variational Bayes. arXiv 2022, arXiv:1312.6114. [Google Scholar]
  31. Yan, N.; Han, X.; Yu, Y.; You, M.; Peng, W. Ultra-high circumferential fatigue properties of Inconel718 nickel-based high-temperature alloy. Mech. Eng. Mater. 2016, 40, 9–12. [Google Scholar]
  32. Wang, L.; Liu, Y.; Chen, G.; Liu, J. High-temperature low-week fatigue damage mechanism of Inconel718 nickel-based high-temperature alloy. Mech. Eng. Mater. 2019, 43, 45–49. [Google Scholar]
  33. Yao, L.; Zhang, X.; Liu, F.; Tu, M.C. High-temperature low-frequency fatigue properties of Inconel718 nickel-based high-temperature alloy. Mech. Eng. Mater. 2016, 40, 25–29, 64. [Google Scholar]
  34. Song, Z. Study on Ultra-High Circumferential Fatigue Fracture Mechanism of Inconel 718 Nickel-Based High Temperature Alloy Based on SLM Forming; Taiyuan University of Science and Technology: Taiyuan, China, 2021. [Google Scholar]
  35. Horňas, J.; Běhal, J.; Homola, P.; Doubrava, R.; Holzleitner, M.; Senck, S. A machine learning based approach with an augmented dataset for fatigue life prediction of additively manufactured Ti-6Al-4V samples. Eng. Fract. Mech. 2023, 293, 109709. [Google Scholar] [CrossRef]
  36. Suykens, J.A.K.; Vandewalle, J.; De Moor, B. Optimal control by least squares support vector machines. Neural Netw. 2001, 14, 23–35. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, H.; Li, B.; Xuan, F.-Z. Fatigue-life prediction of additively manufactured metals by continuous damage mechanics (CDM)-informed machine learning with sensitive features. Int. J. Fatigue 2022, 164, 107147. [Google Scholar] [CrossRef]
  38. Zhang, G.; Su, B.; Liao, W.; Wang, X.; Zou, J.; Yuan, R.; Li, J. Prediction of fatigue life of powder metallurgy superalloy disk via machine learning. Foundry Technol. 2022, 43, 519–524. [Google Scholar]
  39. Rodriguez-Galiano, V.; Sanchez-Castillo, M.; Chica-Olmo, M.; Chica-Rivas, M. Machine learning predictive models for mineral prospectivity: An evaluation of neural networks, random forest, regression trees and support vector machines. Ore Geol. Rev. 2015, 71, 804–818. [Google Scholar] [CrossRef]
  40. Zhang, X.-C.; Gong, J.-G.; Xuan, F.-Z. A deep learning based life prediction method for components under creep, fatigue and creep-fatigue conditions. Int. J. Fatigue 2021, 148, 106236. [Google Scholar] [CrossRef]
  41. Fathalla, E.; Tanaka, Y.; Maekawa, K. Remaining fatigue life assessment of in-service road bridge decks based upon artificial neural networks. Eng. Struct. 2018, 171, 602–616. [Google Scholar] [CrossRef]
  42. Kang, S.; Cho, S. Approximating support vector machine with artificial neural network for fast prediction. Expert Syst. Appl. 2014, 41, 4989–4995. [Google Scholar] [CrossRef]
Figure 1. Definition of defect parameters: (a) ratio of crack source extension area to specimen area; (b,d) defect location; (c) defect parameters.
Figure 1. Definition of defect parameters: (a) ratio of crack source extension area to specimen area; (b,d) defect location; (c) defect parameters.
Materials 18 02604 g001
Figure 2. Structure of GAN model.
Figure 2. Structure of GAN model.
Materials 18 02604 g002
Figure 3. Structure of VAE model.
Figure 3. Structure of VAE model.
Materials 18 02604 g003
Figure 4. Distribution of GAN-generated data.
Figure 4. Distribution of GAN-generated data.
Materials 18 02604 g004
Figure 5. Distribution of VAE-generated data.
Figure 5. Distribution of VAE-generated data.
Materials 18 02604 g005
Figure 6. Pearson correlation plot.
Figure 6. Pearson correlation plot.
Materials 18 02604 g006aMaterials 18 02604 g006b
Figure 7. Framework of the modeling process for fatigue life prediction.
Figure 7. Framework of the modeling process for fatigue life prediction.
Materials 18 02604 g007
Figure 8. (a) GAN models’ predictive accuracy; (b) VAE models’ predictive accuracy.
Figure 8. (a) GAN models’ predictive accuracy; (b) VAE models’ predictive accuracy.
Materials 18 02604 g008
Figure 9. (a) GAN models’ predictive accuracy; (b) variation in GAN models’ predicted life compared to experimental data.
Figure 9. (a) GAN models’ predictive accuracy; (b) variation in GAN models’ predicted life compared to experimental data.
Materials 18 02604 g009
Figure 10. (a) VAE models’ predictive accuracy; (b) variation in VAE models’ predicted life compared to experimental data.
Figure 10. (a) VAE models’ predictive accuracy; (b) variation in VAE models’ predicted life compared to experimental data.
Materials 18 02604 g010
Table 1. Chemical composition of Inconel 718 alloy (wt.%).
Table 1. Chemical composition of Inconel 718 alloy (wt.%).
NiNbMoTiAlCrCFe
53.005.303.001.000.5019.000.05Balance
Table 2. Fatigue experiment data for Inconel 718.
Table 2. Fatigue experiment data for Inconel 718.
σ a / M p a d / μ m d inc / μ m area / μ m S i / 1 N f / C y c l e
52221.008.0018.610.345.4 × 107
49321.003.0018.610.294.4 × 108
47417.003.0015.060.296.3 × 108
4746.006.005.310.272 × 108
44432.85128.5729.110.342.3 × 108
42025.4515.9022.550.312.87 × 108
41528.81141.3025.530.279.35 × 108
452200.00430.76177.240.291.85 × 108
49658.33333.3251.690.253.8 × 106
Table 3. Hyperparameters and parameters of the considered ML models.
Table 3. Hyperparameters and parameters of the considered ML models.
ML ModelTuning EntityRangeNo. of Values
SVRK[‘linear’, ‘poly’, ‘rbf’, ‘sigmoid’]4
C[10, 20, ……, 140, 150]15
ε[0.001, 0.01, 0.1, 1, 10]5
γ[0.001, 0.01, 0.1, 1, 10]5
RFn_estimators[50, 100, 150, 200, 500]5
max_depth[None, 10, 20, 30, 40, 50]6
min_samples_split[1, 2, 4, 6, 8, 10, 16]6
min_samples_leaf[1, 2, 4]6
ANNhidden_layer_sizes [(i, j, k)]
i ϵ [20, 30, …, 70, 80]
j ϵ [0, 10, 20, 40]
k ϵ [0, 10, 15, 20]
(7, 5, 4)
activation[identity, RELU, sigmoid, tanh]4
solver[adam, lbfgs, sgd]3
max_iter[10, 50, 100, 500, 1000, 5000]6
Table 4. Values of hyperparameters of GAN-ML models.
Table 4. Values of hyperparameters of GAN-ML models.
ML ModelTuning EntityValue
GAN-SVRKRBF
C100
ε0.01
γ1
GAN-RF n _ e s t i m a t o r s 100
m a x _ d e p t h 20
m i n _ s a m p l e s _ s p l i t 1
m i n _ s a m p l e s _ l e a f 2
GAN-ANNhidden_layer_sizes. (20, 20, 5)
activation.[tanh]
solver.[lbfgs]
max_iter.500
Table 5. Values of hyperparameters of VAE-ML model.
Table 5. Values of hyperparameters of VAE-ML model.
ML ModelTuning EntityValue
VAE-SVRKRBF
C20
ε0.01
γ1
VAE-RF n _ e s t i m a t o r s 100
m a x _ d e p t h 20
m i n _ s a m p l e s _ s p l i t 1
m i n _ s a m p l e s _ l e a f 2
VAE-ANNhidden_layer_sizes. (80, 10, 20)
activation.[tanh]
solver.[lbfgs]
max_iter.100
Table 6. Prediction accuracy of the optimized ML model for the validation set.
Table 6. Prediction accuracy of the optimized ML model for the validation set.
ML ModelR2
GAN-SVR0.958
GAN-RF0.864
GAN-ANN0.965
VAE-SVR0.979
VAE-RF0.968
VAE-ANN0.923
Table 7. Prediction accuracy of the optimized ML model on the test set.
Table 7. Prediction accuracy of the optimized ML model on the test set.
ML ModelR2MAE (Per Cent)
GAN-SVR0.9291.46
GAN-RF0.9751.13
GAN-ANN0.9191.62
VAE-SVR0.865 2.00
VAE-RF0.879 2.01
VAE-ANN0.861 2.46
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, Z.; Peng, J.; Zhu, L.; Deng, C.; Zhao, Y.; Guo, Q.; Zhu, A. High-Cycle Fatigue Life Prediction of Additive Manufacturing Inconel 718 Alloy via Machine Learning. Materials 2025, 18, 2604. https://doi.org/10.3390/ma18112604

AMA Style

Song Z, Peng J, Zhu L, Deng C, Zhao Y, Guo Q, Zhu A. High-Cycle Fatigue Life Prediction of Additive Manufacturing Inconel 718 Alloy via Machine Learning. Materials. 2025; 18(11):2604. https://doi.org/10.3390/ma18112604

Chicago/Turabian Style

Song, Zongxian, Jinling Peng, Lina Zhu, Caiyan Deng, Yangyang Zhao, Qingya Guo, and Angran Zhu. 2025. "High-Cycle Fatigue Life Prediction of Additive Manufacturing Inconel 718 Alloy via Machine Learning" Materials 18, no. 11: 2604. https://doi.org/10.3390/ma18112604

APA Style

Song, Z., Peng, J., Zhu, L., Deng, C., Zhao, Y., Guo, Q., & Zhu, A. (2025). High-Cycle Fatigue Life Prediction of Additive Manufacturing Inconel 718 Alloy via Machine Learning. Materials, 18(11), 2604. https://doi.org/10.3390/ma18112604

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop