1. Introduction
Gaseous fuels are often used in industrial combustion systems because they produce fewer emissions, are easier to manage when they burn, and are cheaper than liquid and solid fuels. The focus of previous studies has been to cut down on emissions and make combustion more efficient. Many studies are focused on developing advanced flame stabilization ways to help with this. Non-premixed turbulent flames have gotten a lot of attention since they are very efficient and may be used in many ways. It is a widely used method to enhance flame stability in these systems by including swirl, which generates recirculating zones that enhance fuel–air mingling and facilitate flame stabilization [
1].
Swirl-stabilized flames are present in gas turbines, industrial furnaces, and internal combustion engines. The recirculation zones enhance combustion stability and lower emissions [
2].
The complex behavior of swirl flames is still being studied despite their extensive use. Challenges arise from phenomena such as flow instabilities and vortex breakdown [
3].
These dynamics can be effectively captured by Large Eddy Simulation (LES), but it comes at a high computational cost. An alternative that strikes a good balance between accuracy and efficiency is the combination of probability density function (PDF) combustion models and Reynolds-Averaged Navier–Stokes (RANS) models [
4].
Kalt et al. found that the SST
k-
ω turbulence model works well with experiments. Their results were similar to LES for temperature and species fields, but they cost a lot less [
5]. Boke et al. looked at turbulence models for swirl flow in another study and compared how well the SST k-ω, RNG k-ε, and LES models work. The SST k-ω model was very similar to actual data and gave findings that were similar to those of LES for combustion variables like temperature and species mass fractions, but at a lower cost [
6].
The SM1 flame is a well-established benchmark case representing a swirl-stabilized turbulent non-premixed flame. It is included in the Sydney swirl flame dataset, which was developed as part of the Turbulent Non-Premixed Flames (TNF) workshop. This flame configuration is widely used for validating turbulence-chemistry interaction models due to its clearly defined boundary conditions, systematic experimental setup, and strong relevance to real-world combustion applications.
In the SM1 case, methane is injected through a central nozzle surrounded by annular air flow. A swirl number of 0.5 is imposed by tangential air injection. This creates two main recirculation zones: a central vortex breakdown and an outer corner vortex. These structures improve flame stabilization by enhancing fuel–air mixing [
7].
Recurrent Neural Networks (RNNs) and Convolutional Neural Networks (CNNs) are both prevalent deep learning architectures that depend on extensive datasets to efficiently acquire intricate feature representations. CNNs have done very well in recognizing patterns and images. Their layered structure uses convolutional kernels to find local spatial patterns. This helps cut down on the number of parameters and solves the curse of dimensionality, which is a problem where adding more parameters can make errors worse. CNNs were first created for analyzing images, and they have also been used in combustion research. Some of these are modeling unresolved flame surface wrinkling, forecasting scalar variances, and calculating chemical rate constants from shock-tube studies [
8].
Convolutional Neural Networks (CNNs) have been instrumental in the revolutionization of computational fluid dynamics (CFD) by deep learning in recent years. There has been significant progress in this area. These networks excel in capturing intricate flow patterns, managing complex structures, and efficiently processing large datasets. Consequently, CNNs have proven effective in forecasting fluid dynamics. Classical CFD has some problems, like high processing costs and long modeling times [
9].
Deep learning improvements in image processing and pattern recognition have shown that methods based on data can be used to create complicated flame structures. Convolutional Neural Networks (CNNs) have emerged as a pivotal instrument in this domain due to their exceptional capability in feature extraction. Zhang et al.’s study demonstrates that CNN architectures may effectively capture significant flow characteristics, highlighting their potential utility in fluid dynamics research [
10].
Zhang et al. combined a super resolution CNN (SRCNN) and U-Net and built a model named SRUNet to reconstruct flow images. They successfully performed an increase in the model accuracy rather than U-Net and SRCNN [
11].
ANN applications of combustion have generally focused on acceleration of chemical kinetics, combustion kinetics uncertainty, discovery of unknown reaction paths, and building surrogate solvers for simulations [
12].
Ding et al. developed the hybrid flamelet/random data—multiple multilayer perception (HFRD-MMP) method to build chemical kinetic tables. This method improves chemical calculations 12× faster. This model only utilizes the CPU 19.9% [
13]. In addition, Ding et al. made some improvements on the model for NOx formation. The Multiple Multilayer Perceptron—II” (MMLP-II) model calculates the chemical calculations 15× faster [
14]. Nguyen et al. built an ANN model to calculate chemical calculations rapidly. Before training, a stochastic micro-mixing model diluted with thermal losses and burnt gases was simulated. Neural networks were trained by processing chemical composition data using clustering methods. The results show that chemistry analysis with ANN requires only 60% more CPU power than traditional simulations, and the results are consistent with experimental data [
15].
Li et al. used machine learning (ML) techniques to solve the memory requirements of the Flamelet Generated Manifold (FGM) method. Four different ML models, namely two Artificial Neural Networks (ANN), Random Forest (RF), and Gradient Boosted Trees (GBT), were trained and compared to predict the source term and transport properties of the propagation variable. Data preprocessing played an important role in improving model performance. RF and GBT models exhibited high training efficiency and acceptable accuracy, while ANN models had lower error rates but longer training times [
16].
An et al. proposed a methodology to represent complex hydrocarbon chemistry with artificial neural networks (ANNs). These networks were trained on a comprehensive dataset generated by the Latin hypercube sampling (LHS) method. Chemical kinetic mechanisms were represented by thermochemical sample data, and the model was built to cover the entire pressure/temperature/species space in different turbulent flames. The methodology was used to represent 30-species methane chemical mechanisms and validated with non-mixed turbulent flame (DLR_A) and partially mixed turbulent flame (Flame D) simulations. The results showed that ANN-based chemical kinetics do not compromise accuracy while reducing the computational cost by more than two orders of magnitude. This approach holds great potential for complex hydrocarbon fuels [
17].
U-Net is a fully convolutional neural network that is specifically engineered for image segmentation tasks. It converts input photos into the same sized output images immediately. Assembled in a symmetric encoder-decoder design, the encoder shrinks the picture, makes the receptive field bigger, and downsamples to bring out low-frequency details. The decoder subsequently upsamples the input and employs skip connections to maintain information from all encoding steps, maintaining critical visual details. Traditional convolutional networks use fully connected layers to turn extracted features into outputs. U-Net, on the other hand, directly regresses from 2D inputs to 2D outputs, which makes it better at generalization. The Computer Science Department at the University of Freiburg created U-Net to help with biological image segmentation. Since then, it has been used in many other image segmentation tasks because it works so well [
18].
Li et al. introduced a U-Net architecture for forecasting temperature and CO
2 concentration distributions in flames. This approach provides multiple significant benefits for the reconstruction of flame scalar fields. Initially, it necessitates solely spectral data obtained along the axial centerline of the flame, so considerably diminishing the requirement for comprehensive experimental observations. Second, prior knowledge of flame characteristics can be effectively incorporated into the training dataset, allowing the U-Net to achieve high reconstruction accuracy through offline training. Finally, the U-Net performs direct image-to-image mapping between the measured spectral optical thickness and the two-dimensional scalar fields of the flame, preserving spatial continuity more effectively than traditional line-of-sight reconstruction methods. These features demonstrate the U-Net’s strong capability as a reliable tool for flame reconstruction [
19].
In the study by An et al., a deep learning-based turbulent combustion simulation framework is proposed to overcome the high costs of high-resolution CFD simulations. CFDNN, an optimized deep convolutional neural network (CNN) inspired by the U-Net architecture and inception module, is trained to simulate hydrogen combustion. The CFDNN solver provides more than two orders of magnitude speedup compared to traditional CFD solvers, while excellent agreement is achieved in spatial and temporal dynamics. This method offers new possibilities for low-cost, high-accuracy simulations and real-time control of combustion systems [
20].
Maged et al. used deep learning methods to estimate combustion pressure from flame images that provide more information than traditional pressure sensors. Five different models, namely EfficientNetB4, ResNet50, Ensemble Adversarial Inception ResNet, CNN, and CNN-XGBoost, were trained using flame images from a single-cylinder optical gasoline direct injection (GDI) engine. EfficientNetB4 model showed the best performance with R2 of 0.94 and RMSE of 0.70. Furthermore, the deep learning approach achieved higher accuracy than pressure sensors for tracking cycle-to-cycle variations [
21].
Artificial neural networks also can be used for prediction for combustion instabilities.
The study by Zhou et al. investigated the application of deep learning methods to monitor combustion instabilities based on time-averaged flame images. In the experiments conducted on a BASIS burner, a Convolutional Neural Network (CNN) called BIM (BASIS Image Monitor) was designed to extract features from flame shapes and predict thermoacoustic states. The BIM, trained in images of 112 different operating conditions, achieved 99% accuracy after a short training period. The features visualized by the CAM method reveal the connections between flame images and instabilities. The BIM has been shown to be a potential leading model for monitoring and controlling combustion instabilities by providing accurate results under unknown operating conditions [
22].
In the study conducted by Li et al., a convolutional neural network (CNN)-based method supported by long short-term memory (LSTM) and attention mechanisms is proposed for thermoacoustic instability (TAI) detection. Thermoacoustic modes are classified into five different regimes, namely low and high frequency and low and high amplitude. While CNN extracts spatial features from flame images, LSTM captures temporal dynamics, and the attention mechanism focuses on important time steps. The model successfully detects dynamic instabilities of flames in both spatial and temporal dimensions. In addition, the model provides an effective solution for real-time detection and classification of thermoacoustic modes by accurately determining regime transitions at different time scales [
23].
Pan et al. addressed the multimodality detection of combustion instabilities in swirl flames using convolutional neural networks (CNNs). One hundred twenty-nine sets of flame images obtained under different operating conditions were classified by Proper Orthogonal Decomposition (POD), Fast Fourier Transform (FFT), and Phase Space Reconstruction (PSR). The model was trained with K-fold cross-validation, and the best performing model was selected. The results show that ResNet18 model exhibits the best performance under unknown operating conditions. The focus of the model is visualized with Class Activation Mapping (CAM) and it is demonstrated that the model accurately captures not only the flame structure but also the flow field structures and thermo-flow interactions [
24].
Deep learning modeling is also used in flame imaging technologies. In some studies, deep learning methodology has been used to generate flame front structures without using laser. In the study conducted by Han et al., a deep neural network-based method has been proposed to generate flame front structures without using laser. CH-PLIF and chemiluminescence images of turbulent premixed methane/air flames were recorded simultaneously and training was done with conditional generator adversarial network (CGAN). Two different generators, Resnet and U-net, were evaluated, and it was determined that Resnet performed better. The trained model can generate CH-PLIF images from chemiluminescence images with 91% accuracy and can effectively estimate flame surface density at high Reynolds numbers [
25].
The research indicates that artificial neural networks (ANNs) present numerous applications. The improvement of chemical reaction calculations and their applications in tabular chemistry are highly significant. Moreover, numerous works examine AI-assisted instruction of time-dependent flame imagery via targeted preprocessing. Nonetheless, a practical methodology for instructing CFD outcomes for natural gas burners in industrial contexts is absent. The work we have conducted addresses this deficiency. The main objectives of this study can be shown as follows:
A model based on input parameters for temperature distribution image prediction from input parameters
Temperature distribution image prediction from cold image-based RGB fusion.
To assess and compare the accuracy, efficiency, and size of DL models.
To provide a practical method that supports faster and reliable combustion temperature predictions.
The primary aim is to create models that depict general flame behavior and high-temperature areas in temperature forecasts and to evaluate them against existing artificial intelligence models. In literature, investigations have predominantly utilized Res-Net and U-Net, while comparative analyses involving Efficient Net and InceptionV3 Net have not been observed. The RGB Fusion approach produced integrated graphics from cold flow (none-reacting) CFD visuals, although their incorporation into industrial-level artificial intelligence remains unaddressed in the literature. The study also analyzed the file sizes and prediction periods of the models regarding their suitability for industrial application.
4. ANN Architecture
Two different approaches were developed to predict temperature fields in the SM1 flame configuration. In order to better compare the performance of the models created in both approaches, a comparison was made with Res-Net, Efficient Net, and Inception V3 models, and the efficiency of the models was determined by giving comparative results in both approaches.
The first approach employs a fully connected dense neural network that maps scalar input parameters—fuel velocity, swirl ratio, and equivalence ratio—directly to temperature contour images. This model is trained on labeled CFD data, enabling temperature field prediction based solely on these physical input variables.
The second approach uses a convolutional neural network (CNN) with an encoder-decoder architecture for image-to-image translation. It takes as input composite RGB images generated by fusing cold flow scalar fields—velocity magnitude, methane mass fraction, and turbulence time scale—using an RGB Fusion technique. The network outputs predicted temperature contour images. This method leverages the computational efficiency of cold flow CFD simulations and spatial feature extraction capabilities of CNNs to estimate temperature fields.
Both models were trained using an 80/20 train-test split of the CFD dataset. While the first model relies on explicit physical parameters and dense layers, the second utilizes hierarchical convolutional layers for improved spatial feature learning and reduced model size.
4.1. Input Parameters to Image Learning Application (First Approach)
In the first approach of the study, a deep learning model was developed to predict temperature contours directly from three input parameters: fuel velocity, swirl ratio, and equivalence ratio (
Table 2). In this approach, four models were performed. These models are Dense Model, Res-Net, Efficient Net, and Inception V3.
Dense Model was implemented using the TensorFlow and Keras libraries. Its architecture, shown in
Figure 9, consists of fully connected (dense) layers that process the input vector and generate high-resolution temperature contour images. The input layer receives a three-dimensional vector corresponding to the physical parameters. This is followed by three dense layers, each with 256 neurons activated by the ReLU function to capture nonlinear relationships. A dropout layer with a 10% rate is included after the first dense layer to reduce overfitting and improve generalization. The final dense layer outputs a vector reshaped into a 256 × 256 × 3 tensor representing the predicted temperature image in RGB format. A sigmoid activation function is applied to normalize pixel values between 0 and 1. The model was compiled with the Adam optimizer, employing Mean Squared Error (MSE) as the loss function. Mean Absolute Error (MAE) was used as an additional metric to monitor training and validation performance. Training was performed using mini batches, with an early stopping criterion applied to halt training if validation loss did not improve over five consecutive epochs, preventing overfitting and reducing computation time. The dataset was partitioned into training and testing subsets using an 80/20 split ratio. Specifically, 80% of the available data was allocated for model training to enable the neural network to learn the underlying relationships between input parameters and temperature contours. The remaining 20% was reserved as an independent test set to rigorously evaluate the model’s generalization capability on unseen data. This standard data partitioning approach helps prevent overfitting and provides an unbiased assessment of model performance, ensuring the robustness and reliability of the predictive results.
The Res-Net model utilizes a framework that incorporates deep learning attributes trained on ImageNet. The Res-Net layers have been immobilized, with just the last layers of the model undergoing training. The model’s advantages include the utilization of residual connections to mitigate gradient loss. The EfficientNet network has been evaluated to provide a more efficient network design for industrial applications. EfficientNet has undergone training on ImageNet. It has been designed for mobile platform utilization. The Inception V3 model, trained on ImageNet, employs smaller filters and multi-scale learning techniques. The performance of multi-learning has been evaluated.
4.2. Cold Flow Image to Output Image Learning Application (Second Approach)
In the study’s second approach, four distinct ANN models were employed, similar to the first approach. The models include U-Net, ResNet, EfficientNet, and InceptionNet.
Firstly, it was developed to predict temperature contours using input derived from cold flow CFD simulations. Steady-state simulations were performed using the SST k-ω turbulence model and the species transport model, with combustion reactions neglected.
Equations of the SST k-
ω turbulence model are given as follows Equation (9) [
29]:
In Equation (9), ρ represents the density of the mixture (kg/m3), D is the diffusion coefficient, and Y is the mass fraction of species i.
Scalar fields such as velocity magnitude, methane mass fraction, and turbulent time scale (
k/
ε) were extracted from the non-reacting CFD results [
36].
These scalar fields were converted into grayscale images by normalizing their values between 0 and 1. The grayscale images were then assigned to the red, green, and blue channels, respectively, creating composite RGB images that encode multiple physical fields in a single representation. This preprocessing technique, referred to as RGB Fusion, facilitates the integration of diverse flow information into a unified format suitable for deep learning. A schematic diagram of the image preprocessing pipeline is presented in
Figure 10.
The resulting 256 × 256 × 3 RGB images were used as input to the CNN model. The U-Net architecture employed consists of an encoder-bottleneck-decoder structure. The encoder includes two downsampling levels, each containing convolutional layers with ReLU activation functions and batch normalization, followed by max pooling layers. The bottleneck comprises convolutional layers with 512 filters. The decoder mirrors the encoder with upsampling via transposed convolutions and includes skip connections to retain spatial context from earlier layers. A final 1 × 1 convolution with a sigmoid activation function produces the normalized RGB output image representing the predicted temperature field. Model architecture is represented in
Figure 11The model was compiled using the Adam optimizer and trained using Mean Squared Error (MSE) as the loss function. Mean Absolute Error (MAE) was also monitored to assess model accuracy. The dataset was partitioned into 80% training and 20% testing subsets to evaluate the model’s generalization to unseen data. Training was conducted with a batch size of 16 for up to 200 epochs.
To prevent overfitting and ensure training efficiency, an early stopping mechanism was applied. If the validation loss did not improve for 10 consecutive epochs, training was halted, and the model weights corresponding to the lowest validation loss were restored. This training strategy enabled efficient convergence while maintaining robust performance.
6. Conclusions
This study presents a novel integration of deep learning techniques with CFD simulations for the prediction of temperature fields in the SM1 swirl-stabilized turbulent flame. Initial CFD analyses using the SST k-ω turbulence model coupled with a Steady Laminar Flamelet combustion model provided a robust database for training.
Two distinct deep learning frameworks were developed and evaluated. The first model directly maps scalar input parameters—fuel velocity, swirl ratio, and equivalence ratio—to high-resolution temperature contours, enabling efficient parametric studies. In the first approach four models were performed. These are Dense Model, Res-Net, Efficient Net, and InceptionV3 model.
The preliminary analysis revealed that the InceptionV3 model and our Dense Model offered a superior solution. Due to its multi-layered architecture, the InceptionV3 model has surpassed the Dense Model in predictive performance. Upon examining the model dimensions, it is evident that all models possess comparable sizes. The Dense Model, Res-Net, and Inception V3 models have excelled at predicting contour borders.
In the second approach, cold flow (non-reacting) computational fluid dynamics images underwent pre-processing through the RGB fusion technique to generate composite images. The images were processed using U-Net, Res-Net, Efficient Net, and Inception V3 networks, and the prediction outcomes were disseminated. In the second technique, U-Net excels in both file size and predictive performance. Efficient Net exhibited the poorest model performance. The Efficient Net model has demonstrated notable variations within the temperature contour borders. It is inadequate for ascertaining overall conduct. The Res-Net model is the most appropriate choice following U-Net.
The model tends to produce reliable results only within the range of parameters presented during training, and due to its nature as a non-physics-based model, the ability to directly interpret the physical consistency of the results may be limited. By design, the model focuses more on steady-state conditions rather than time-dependent transient behaviors. Future work will focus on extending the deep learning framework to predict pollutants and emissions. In addition, the turbulence model’s effect on flame behavior is investigated. The presented approach represents a significant step toward data-driven combustion modeling, promising substantial benefits in terms of computational cost reduction and design efficiency.