Next Article in Journal
Innovative Approaches to RC Deep Beam Strengthening: Evaluating Low-Cost Glass Fiber Wraps Against Traditional CFRP Solutions
Next Article in Special Issue
Mitigation of Membrane Fouling in Lignin Recovery from Black Liquor via Surface-Patterned Ceramic Membrane
Previous Article in Journal
Evaluation of Selected Fire Properties of Recycled Particleboards
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring the Feasibility of Deep Learning for Predicting Lignin GC-MS Analysis Results Using TGA and FT-IR

1
Department of Computer Engineering, Kyungnam University, Changwon 51767, Gyeongsangnam-do, Republic of Korea
2
Carbon-Neutral Resources Research Center, Hankyong National University, 327, Jungang-ro, Anseong 17579, Gyeonggi-do, Republic of Korea
3
School of Food Biotechnology and Chemical Engineering, Hankyong National University, 327, Jungang-ro, Anseong 17579, Gyeonggi-do, Republic of Korea
4
Fivenode, Seoul 07549, Republic of Korea
5
Department of Artificial Intelligence, Kyungnam University, Changwon 51767, Gyeongsangnam-do, Republic of Korea
*
Author to whom correspondence should be addressed.
Polymers 2025, 17(6), 806; https://doi.org/10.3390/polym17060806
Submission received: 10 February 2025 / Revised: 8 March 2025 / Accepted: 13 March 2025 / Published: 18 March 2025
(This article belongs to the Special Issue Lignin Isolation, Characterization and Application)

Abstract

:
Lignin is a complex biopolymer extracted from plant cell walls, playing a crucial role in structural integrity. As the second most abundant biopolymer after cellulose, lignin has significant industrial value in bioenergy, the chemical industry, and agriculture, gaining attention as a sustainable alternative to fossil fuels. Its composition changes during degradation, affecting its applications, making accurate analysis essential. Common lignin analysis methods include Thermogravimetric Analysis (TGA), Fourier-transform Infrared Spectroscopy (FT-IR), and Gas Chromatography–Mass Spectrometry (GC-MS). While GC-MS enables precise chemical identification, its high cost and time requirements limit frequent use in budget-constrained studies. To address this challenge, this study explores the feasibility of an artificial intelligence model that predicts the GC-MS analysis results of depolymerized lignin using data obtained from TGA and FT-IR analyses. The proposed model demonstrates potential but requires further validation across various lignin substrates for generalizability. Additionally, collaboration with organic chemists is essential to assess its practical applicability in real-world lignin and biomass research.

1. Introduction

Lignin is a major component of the plant cell wall, along with cellulose and hemicellulose. Traditionally considered a by-product of cellulose production, it has often been discarded or used as fuel. However, with increasing emphasis on sustainability, lignin is gaining attention as a potential fossil fuel alternative across various industries and remains an active research focus in the chemical industry, bioenergy, and agriculture [1,2,3].
Composed of monomeric units such as p-hydroxyphenyl, guaiacyl, and syringyl, lignin forms a complex three-dimensional structure stabilized by various linkages ( β -O-4, β - β , and 5-5) and functional groups, including methoxy, carboxyl, and carbonyl moieties [4]. The heterogeneous nature of lignin complicates its utilization as a uniform chemical feedstock in industrial applications, particularly in bio-based material synthesis and biorefinery processes [5].
To enhance lignin utilization, depolymerization is essential. Methods include thermal decomposition, chemical treatments with oxidizing or reducing agents, and catalytic processes. Among these, hydrothermal liquefaction (HTL) is widely applied, converting high-molecular-weight biomass into low-molecular-weight aromatic compounds, such as bio-oil, gas, and solid residues [6].
Since the physicochemical properties of depolymerized lignin are strongly influenced by the processing method, precise compositional analysis is essential for optimizing downstream applications such as bio-oil production and polymer synthesis [7,8]. Common techniques include Thermogravimetric Analysis (TGA), Fourier-transform Infrared Spectroscopy (FT-IR), and Gas Chromatography–Mass Spectrometry (GC-MS). This study utilizes these methods to analyze depolymerized lignin.
TGA assesses thermal stability by monitoring decomposition under controlled heating, while FT-IR identifies functional groups based on their infrared absorption profiles. GC-MS delivers comprehensive qualitative and quantitative analysis of complex mixtures and remains the method of choice in many research settings [9,10,11,12,13].
However, the high capital and operational costs of GC-MS instruments (typically USD 50,000–USD 150,000) often limit in-house analyses, forcing researchers to outsource sample testing at significant per-sample costs (USD 100–USD 500, compared to USD 50–USD 150 for TGA and USD 20–USD 100 for FT-IR [14]), thereby restricting the scale and frequency of GC-MS applications in resource-limited studies. These economic constraints pose significant challenges, particularly in studies involving extensive sample analysis.
Recent advancements in artificial intelligence (AI) have facilitated more efficient lignin analysis, particularly in response to financial and logistical constraints associated with conventional chemical techniques. Wen et al. (2024) developed a predictive model leveraging FT-IR and machine learning (XGBoost, LightGBM) to estimate lignin content, demonstrating that this approach significantly accelerates analysis compared to traditional methods [15]. Ge et al. (2023) employed machine learning to model lignin structural transformations during DES pretreatment, confirming that delignification dynamics and structural modifications can be reliably predicted with minimal experimental data [16]. Diment et al. (2024) utilized Bayesian optimization to enhance the production of lignin–carbohydrate complexes (LCCs), demonstrating that machine learning can optimize biorefinery conditions, reducing the need for extensive experimental trials while improving yield and structural control [17]. Löfgren et al. (2022) applied Bayesian optimization to infer lignin properties without direct experimentation, facilitating the derivation of optimal compositions [18].
Existing studies utilizing machine learning for lignin characterization predominantly focus on macroscopic property estimation, such as structural changes and functional group analysis. In contrast, this study specifically aims to predict the individual chemical composition of lignin depolymerization products by leveraging deep learning models trained on TGA and FT-IR data. To address this limitation, this study introduces a deep learning-based framework that utilizes cost-effective TGA and FT-IR data to predict GC-MS analytical results. The key contributions of this research are as follows:
  • Decoder-Based Generative Models: A deep learning model incorporating the distinct features of TGA, FT-IR, and GC-MS data to generate synthetic datasets.
  • Predictive Modeling in Limited Data Scenarios: Synthetic data augmentation to enable training under data-scarce conditions.
  • GC-MS Predictability Using TGA and FT-IR Data: Development and evaluation of a GC-MS prediction model based on TGA and FT-IR inputs.
This paper is organized as follows: Section 2 provides details on dataset preprocessing methods and the generative dataset model. Section 3 introduces the proposed predictive model. Section 4 analyzes the results, highlights key findings, and suggests future research directions.

2. Proposed Method

This section describes the process of developing a deep learning-based GC-MS prediction model using data preprocessing methods and data generation deep learning models.
Figure 1 illustrates the overall research workflow. Due to the limitations of the available data, we employed a data generation strategy that encapsulates the inherent characteristics of the data, facilitating the generation of supplementary samples. This study investigates the potential of predicting GC-MS data using TGA and FT-IR datasets.
Recent research highlights the prevalence of Convolutional Neural Network (CNN)-based data generation and augmentation methodologies across various domains. These approaches effectively mitigate the constraints posed by limited experimental data and enhance the efficacy of predictive models [19,20,21]. Consequently, we implemented CNN-based data augmentation techniques to expand both TGA and FT-IR datasets. The augmented TGA and FT-IR datasets are then utilized as inputs for the GC-MS prediction model, which aims to estimate the corresponding GC-MS data. The predicted GC-MS results are subsequently compared with the augmented GC-MS data.
Through this approach, this study examines the feasibility of the deep learning-based GC-MS prediction model.

2.1. Dataset Overview

The datasets analyzed in this study were collected in a laboratory environment. The provided data include analysis results from three methods: TGA, FT-IR, and GC-MS. These methods were used to study depolymerized lignin heated at four different temperatures (250 °C, 300 °C, 350 °C and 400 °C) during the HTL process. Each method contains experimental results for all four temperatures.
TGA analysis was performed on depolymerized lignin under HTL temperature conditions of 250 °C, 300 °C, 350 °C and 400 °C. The initial temperatures of the TGA data were 39.65 °C, 40.42 °C, 39.62 °C, and 39.61 °C, while the final temperatures were 787.58 °C, 785.94 °C, 787.52 °C, and 786.58 °C. The dataset contains approximately 22,000 data points.
FT-IR analysis was performed three times for each temperature, resulting in a total of 12 datasets. The recorded wavelength range of the data is from 649.8934 cm−1 to 4000.6047 cm−1. Characteristic chemical bonds, such as C=O at 1730 cm−1 and O-H at 3200–3600 cm−1, are identified in the spectrum, where peaks can be observed in the graph.
The compounds detected through GC-MS analysis are classified into nine groups, each comprising multiple derivatives, totaling 114 chemical compounds. The analysis results indicate the relative proportions of depolymerized lignin components represented by each derivative. Detailed information on the identified compounds is provided in Appendix A.

2.1.1. Tga Data Preprocessing and Generation

Figure 2 illustrates the proposed TGA data preprocessing pipeline, which consists of two stages. Initially, temperature clipping was conducted. The depolymerized lignin was analyzed via TGA at HTL temperatures of 250 °C, 300 °C, 350 °C, and 400 °C. The initial temperatures of the TGA data recorded were 39.65 °C, 40.42 °C, 39.62 °C, and 39.61 °C, with final temperatures measuring 787.58 °C, 785.94 °C, 787.52 °C, and 786.58 °C. However, a notable discrepancy in the initial and final temperatures was observed. This variance can be attributed to factors such as the operational conditions of the TGA apparatus, the initial status of the sample, and the automatic shutdown parameters. Beyond 700 °C, the stabilization of residual carbon and the thermal changes progressively decrease. In particular, data collected above 760 °C may be influenced by noise arising from the thermal constraints of the TGA instrument at elevated temperatures, thereby compromising data reliability. To mitigate this concern, the data range was confined to an initial temperature of 40 °C and a final temperature of 760 °C, ensuring consistency in the temperature gradients.
Furthermore, a resampling process was applied. The dataset comprised over 22,000 data points, which, if used for training, could lead to overfitting and require long training durations. To address this challenge, the weight change data were averaged at 1 °C intervals for each temperature, thereby reducing the dataset size. This methodological approach facilitated the elimination of redundancy and minimized the risk of overfitting [22].
Table 1 shows the proposed TGA data generation model architecture. The TGA data are a one-dimensional time series dataset that records the weight of a sample as the temperature increases. This characteristic was considered in the model design, leading to the adoption of the CNN architecture. To enable the model to learn the causal relationship between weight changes and temperature variations, two sets of Conv+ReLU layers were placed to extract local patterns. Subsequently, a Fully Connected layer (FC) integrates global features to generate the final TGA data. This structure captures the sequential nature of TGA data, and its decoder-based design effectively reflects TGA characteristics, allowing data generation under various conditions.

2.1.2. FT-IR Data Preprocessing and Generation

Figure 3 illustrates the FT-IR data preprocessing pipeline. An analysis of the provided FT-IR datasets revealed a slight exceedance of the 100% threshold due to noise and outliers. Since FT-IR data are used to assess the spectral absorbance or transmittance of samples, values exceeding 100% pose inherent issues, necessitating a preprocessing step to prevent the generation of extraneous or distorted data during model training. The dataset comprises results from three analyses conducted under uniform conditions, resulting in a total of 12 distinct datasets. Figure 3 illustrates the process of integrating these datasets, addressing the aforementioned issues, and transforming them into a format optimized for efficient model training. This integration can be mathematically articulated as follows.
D preprocessed = log max 101 min 1 N i Group D FTIR , i , 100 , 10 6
In Equation (1), D FTIR , i represents the FT-IR spectrum data of the i-th sample, and preprocessing is carried out through the following procedure.
The data are first integrated by calculating the arithmetic mean for each temperature. Subsequently, values exceeding 100 are constrained using the minimum function to ensure they remain within this threshold. The clipped mean values are then subtracted from 101 to adjust for offset and stabilization.
To prevent errors during logarithmic transformation due to excessively small values, a stabilization constant of 10 6 is applied using the maximum function. Finally, a logarithmic transformation is performed to reduce the dynamic range of the data, making it more suitable for analysis.
This preprocessing enhances the reliability of FT-IR data by minimizing the influence of extreme values while improving its usability for subsequent modeling and analysis.
Table 2 shows the proposed FT-IR data generation model architecture. FT-IR data encapsulate the absorption of light by a sample across various frequency ranges, represented in the form of a spectrum. Given that this methodology yields pronounced peaks in specific frequency bands alongside inherent noise, a dropout technique has been integrated into CNN to address these nuances. Dropout is instrumental in mitigating overfitting by randomly deactivating a subset of neurons during the training process. Initially, three convolutional layers with ReLU activation extract features from distinct frequency regions, while the FC aggregates the overall spectral distribution. Subsequently, the dropout layer diminishes the propensity of the model to become biased toward noise or dominant peak values in particular frequency bands during training. This framework is conceived as a decoder-based model that faithfully represents the spectral characteristics intrinsic to FT-IR data, thereby ensuring robust data generation, even in the presence of complex FT-IR spectra.

2.1.3. GC-MS Data Preprocessing and Generation

GC-MS analysis is a technique that integrates Gas Chromatography (GC) and Mass Spectrometry (MS). GC serves as the initial step, facilitating the separation of the components within a complex mixture. The sample is heated to vaporize the constituents, which are then introduced into a chromatographic column. As the gaseous components traverse the column, they interact variably with the stationary phase, resulting in differential retention times that enable the respective components to be distinguished. Following this separation, Mass Spectrometry is employed to further analyze the isolated components. The samples are directed into a mass spectrometer, where they undergo ionization. The resulting ionized molecules are subsequently separated according to their mass-to-charge ratio ( m / z ), with their velocities varying in correlation to their mass. This intricate process culminates in the generation of a mass spectrum, which provides critical insights into the molecular weight, structural characteristics, and elemental composition of each identified component.
During the analysis of the acquired GC-MS data, certain components were not fully quantified. For instance, several compounds within the Syringyl group were not detected in the dataset utilized for this study. These compounds include 4-Propylsyringol (2,6-Dimethoxy-4-propylphenol), Propiosyringone (1-Propanone, 1-(4-hydroxy-3,5-dimethoxyphenyl)-), and Dihydrosyringenin (3-Syringylpropanol). Consequently, the total proportion of detected components at each temperature condition (250 °C, 300 °C, 350 °C, and 400 °C) was 91.16%, 91.23%, 86.11%, and 88.07%, respectively, falling short of the expected sum of 100%.
This phenomenon can be attributed to multiple factors, including insufficient volatility that hinders the migration of the compound through the GC column, leading to its residual presence, thermal degradation at elevated temperatures which results in the loss of its original signal, and inadequate ionization within the MS process, culminating in the absence of detectable signals.
To resolve this issue, data preprocessing is performed. Initially, instead of using the proportion of each derivative within a group, the overall group proportion was considered. This methodology preserves critical information essential for model training while enhancing computational efficiency. Furthermore, an “Other” classification was introduced to account for undetectable compounds within the GC-MS analysis, thereby ensuring that the total proportion sums to 100%. This modification prevents distortions arising from incomplete data and enhances the accuracy and generalizability of the resultant findings.
Table 3 shows the proposed GC-MS data generation model architecture. This model is designed as a decoder-based CNN architecture, generating new GC-MS data using the provided input. First, the Conv+ReLU layers learn the local correlations between component ratios. Then, the Global Average Pooling (GAP) layer compresses the feature maps extracted from multiple convolutional layers into a single average value [23]. The GAP layer reduces the number of parameters, prevents overfitting, and efficiently represents the global characteristics of the input data. The global features obtained from the GAP layer are then passed to the FC to learn the overall correlations between the input component ratios and GC-MS data. Finally, the output layer, which includes a Softmax activation function, converts the probabilistic distribution of the GC-MS data into component values.

2.2. Proposed GC-MS Prediction Model

This section proposes a model that predicts GC-MS data using generated data from the model described in Section 2.1 as input data.

2.2.1. Model Architecture

In this study, a model integrating a Multimodal Variational Auto Encoder (MVAE) [24] with a Mixture of Experts (MoE) [25] to predict GC-MS data is presented. The model encompasses an encoder that assimilates TGA and FT-IR spectroscopic data as inputs, a decoder that reconstructs data within the latent space, an expert network that forecasts GC-MS data, and a gating network that determines the weighting of the expert network.
Figure 4 presents the proposed GC-MS data prediction model architecture, which integrates encoders, latent space representation, decoders, and expert-based prediction using an MoE.
In the encoders, 761-dimensional TGA data and 3476-dimensional FT-IR data are processed through separate encoders. Each encoder extracts meaningful features from its respective input data, computing the mean ( μ ) and logarithmic variance ( log σ ) to define a probabilistic distribution. These encoded features are then integrated into a shared latent space, where the intrinsic characteristics of TGA and FT-IR data are fused into a unified latent representation z. The latent vector is sampled using the reparameterization trick, which enables the model to maintain probabilistic properties while ensuring efficient training.
The latent vector z serves as a central representation that facilitates both data reconstruction and predictive modeling. Two decoders independently reconstruct the original spectral data from z, with the TGA decoder and FT-IR decoder minimizing the reconstruction loss of the Variational Auto Encoder (VAE) [26]. Concurrently, the same latent vector is utilized for GC-MS prediction through the MoE framework.
Within the MoE structure, multiple expert networks analyze the latent vector from different perspectives to generate GC-MS predictions. A gating network dynamically assigns optimal contributions to each expert, ensuring adaptive weighting based on input characteristics. By leveraging these mechanisms, the MVAE-MoE model effectively integrates TGA and FT-IR data to improve the accuracy and generalizability of GC-MS predictions while enabling robust estimations under varying temperature conditions in the HTL process.

2.2.2. Loss Function

To optimize the model training, the loss function was newly defined as follows.
L Total = L Recon + L Pred + L KL , where L Recon = L Recon TGA + L Recon FT IR
In Equation (2), L Total signifies the comprehensive loss value. L Recon TGA and L Recon FT - IR denote the reconstruction losses associated with the TGA and FT-IR datasets, respectively, which are restored through the TGA decoder and FT-IR decoder within the model architecture.
In a VAE, the input data are encoded into a latent distribution, and the reparameterized latent vector z is subsequently reconstructed to closely approximate the original input. This procedure assesses the fidelity with which the latent space represents the input data. Furthermore, L Pred quantifies the difference between the predicted GC-MS data and the actual GC-MS data, using the Mean Squared Error (MSE) as the loss function. The reconstruction and prediction loss is quantified using MSE, which can be formulated as follows:
{ L Recon TGA , L Recon FT IR , L Pred } = 1 n i = 1 n y i y ^ i 2
In Equation (3), y i represents the actual data, while y ^ i denotes the reconstructed or predicted data. The variable n indicates the data index. The MSE is computed as the squared difference between the actual and predicted values, imposing greater penalties on larger errors. This mechanism guides the latent space to effectively learn the intrinsic characteristics of TGA and FT-IR data, thereby enhancing the model’s representation capability.
L KL = 1 2 1 + log σ 2 μ 2 σ 2
Equation (4) defines K KL as the KL divergence loss function in Equation (2), which quantifies how closely the latent vector z approximates the standard normal distribution N ( 0 ,   1 ) [27]. KL divergence loss plays a crucial role in regulating the structure of the latent space by encouraging the mean ( μ ) to be close to 0 and the variance ( σ 2 ) to be close to 1. Ensuring that μ 0 prevents the latent variables from shifting excessively in one direction, maintaining a balanced and well-structured distribution. Similarly, enforcing σ 2 1 prevents the latent space from collapsing into a single point or becoming overly dispersed, both of which can degrade the quality of learned representations. If σ 2 is too small, the latent space becomes too concentrated, limiting the diversity of generated outputs, while if it is too large, important structural information may be lost. By maintaining these constraints, the model ensures that the latent variables remain consistent with the standard normal distribution, allowing for stable training and generalization. Ultimately, KL divergence loss functions as a mechanism that prevents the latent space from becoming unstructured, thereby improving the model’s ability to generate meaningful and diverse representations.

3. Experimental Result

This section performs a comprehensive evaluation of each data generation model and the GC-MS prediction model introduced in the preceding section. The model training settings are as follows: the Adam optimizer is utilized with an initial learning rate of 0.001. To adjust the learning rate dynamically, the StepLR scheduler reduces it by a factor of 0.1 after 75 epochs. The batch size is set to 32, and training is conducted for 100 epochs. During training, the loss value is calculated at each epoch, followed by backpropagation to minimize the loss. In addition, loss is monitored at each epoch to track training progress.

3.1. Evaluate Each Data Generation Model

The evaluation methodology involves generating results across different temperature settings for each model, which are subsequently analyzed to assess model performance. However, the target temperatures that the models are intended to predict are not included in the provided dataset. Consequently, actual data are not available for direct comparison with the model predictions.
To address this limitation, interpolation was applied based on the relationships between existing data to generate a comparative dataset. Recently, data augmentation techniques based on interpolation have been utilized in various studies. This approach has led to studies aimed at overcoming the limitations of the dataset [28,29,30,31,32].
Given the limited availability of labeled experimental data, interpolation provides a viable approach to synthetically enhance data diversity while preserving key statistical properties, thereby improving the robustness of predictive models. In this study, linear interpolation was applied for comparative analysis, as defined below:
y = y 1 + x x 1 x 2 x 1 ( y 2 y 1 )
In Equation (5), x 1 and x 2 denote the specified temperature conditions, while y 1 and y 2 represent the GC-MS data corresponding to these temperature conditions.
Linear interpolation is an effective method when data exhibit linear characteristics; however, if nonlinear patterns are present, the interpolated values may not accurately reflect the underlying data properties. Nonetheless, this study adopted linear interpolation for its simplicity and ease of application with a limited number of data samples.
The model evaluation was performed using metrics such as the Mean Absolute Error (MAE), R 2 , and the Pearson correlation coefficient. Botchkarev elucidated that distinct evaluation metrics possess unique characteristics, and reliance on a solitary metric may only capture specific dimensions of model performance [33]. Therefore, by concurrently analyzing MAE, R 2 , and the correlation coefficient, we attain a more nuanced assessment of the efficacy and reliability of the model.
MAE = 1 n i = 1 n | y i y ^ i |
Equation (6) defines the MAE [34]. The MAE quantifies the average absolute deviation between the predicted values and the interpolated values, indicating the extent to which the model predictions align with the reference values. A lower MAE value, approaching zero, signifies a stronger correspondence between the predicted and reference values.
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
Equation (7) defines the R 2 [35], which quantifies how well the model captures the variability in the interpolated values. A value closer to 1 indicates that the model accurately reflects the changes in the interpolated values. In some cases, the R 2 value may be negative, implying that the model’s predictions perform worse than the interpolated values.
r = Cov y , y ^ σ y · σ y ^
Equation (8) defines the correlation coefficient [36], a statistical measure that quantifies the linear relationship between predicted and actual values. This coefficient evaluates the degree to which the two datasets co-vary, with values approaching 1 indicating a strong positive linear correlation, while values near 0 suggest the absence of a discernible linear relationship.

3.1.1. TGA Data Generation Model

To quantify the data loss introduced during the TGA data preprocessing, the Wasserstein distance is utilized.
W 1 ( P , Q ) = inf γ Π ( P , Q ) R d × R d | x y | d γ ( x , y )
Equation (9) defines the Wasserstein distance, which is a metric that measures the optimal transport cost between two probability distributions [37] and is used to quantitatively evaluate the differences between data distributions. In this study, the Wasserstein distance was utilized to quantify the information loss between data before and after preprocessing.
Table 4 presents the Wasserstein distance between preprocessed and original data. In Table 4, it can be observed that information loss during TGA data preprocessing was minimal. This indicates that the preprocessing method effectively preserved the intrinsic characteristics of the data while also reducing its size. Furthermore, the results suggest that the preprocessed data maintain a high degree of similarity to the original data, supporting their reliability as input for subsequent modeling. These findings demonstrate that the preprocessing approach contributes to the stability and predictive performance of the model.
Table 5 presents the evaluation of the TGA data generation model at different temperatures. The results indicate that the model performs consistently well under varying temperature conditions.

3.1.2. FT-IR Data Generation Model

Table 6 presents a comparison of the FT-IR data generated by the data generation model at various temperatures and the interpolated data obtained by interpolation.
The comparison results generally show high performance. However, the MAE value at 345 °C is relatively higher than at 260 °C. This issue can be addressed through model optimization or by acquiring additional data.

3.1.3. GC-MS Data Generation Model

Table 7 shows the comparison of the GC-MS data generated by the data generation model at various temperatures and the interpolated data obtained by interpolation.
The model demonstrated strong performance across most temperature ranges. However, at 280 °C, the MAE was relatively high (0.00228), while R 2 and the correlation coefficient dropped significantly to 0.64181 and 0.82605, respectively. The performance degradation at a specific temperature can be addressed by acquiring additional data for that temperature range.
The analysis of the evaluation tables for each data generation model shows that all models exhibit satisfactory performance. This confirms that the proposed data generation models can address data scarcity in situations with limited data and can be effectively utilized as training data for the final GC-MS prediction model.

3.2. Evaluate GC-MS Prediction Model

The trained model was used to predict the GC-MS analysis results for lignin depolymerized at various temperatures during the HTL process. Table 8 presents the predicted results.
The GC-MS prediction results in Table 8 demonstrate that the model effectively captures the compositional changes in lignin depolymerization across different temperatures. Other aromatic compounds (0.2984) had the highest predicted mean proportion, followed by glycerol-derived compounds (0.1479), syringyl (0.1384), and guaiacyl (0.1210). At 280 °C, syringyl (0.0585) and guaiacyl (0.1075) sharply declined, while other aromatics increased (0.3308), indicating that the model recognizes the decomposition of major aromatic monomers and the formation of low-molecular-weight compounds.
Variance analysis revealed that syringyl (0.0038) and guaiacyl (0.0027) showed significant fluctuations, while polycyclic aromatics (0.0013), alkanes (0.0018), and cyclic compounds (0.0021) remained relatively stable. The model also predicted a peak in polycyclic aromatics at 345 °C (0.0528), followed by a decline at 390 °C (0.0062), suggesting high-temperature decomposition. Additionally, alkanes increased at 390 °C (0.0883), fatty acids peaked at 345 °C (0.0729), and glycerol-derived compounds peaked at 365 °C (0.2303) before declining (0.1077 at 390 °C), reflecting thermal degradation trends.
Although the model captures general depolymerization patterns, fluctuations at 280 °C and 345 °C indicate potential limitations in learning nonlinear transformations. Further data collection and model refinement are necessary to improve predictive accuracy.
Table 9 shows the evaluation results assessing the consistency between the interpolated data and the reference data. The table presents the group ratios, MAE, R 2 , and correlation coefficients for each temperature condition.
Notably, at 260 °C and 365 °C, the data indicate low MAE values, high correlation coefficients, and robust R 2 values, thereby confirming the model’s superior predictive performance under these conditions. In contrast, the results at 280 °C and 345 °C reveal comparatively higher MAE, lower R 2 , and weaker correlation coefficients, clearly demonstrating a decline in performance. These findings not only attest to the interpolation method’s effectiveness in preserving the original data characteristics but also highlight the necessity for further validation using empirical data.

4. Conclusions

This study investigated the feasibility of an AI-based model for predicting GC-MS analysis results of depolymerized lignin using FT-IR and TGA data. The primary objective was to establish a cost-effective and efficient analytical methodology that could serve as a viable alternative to conventional GC-MS analysis, which is often associated with high costs and complex sample preparation procedures.
The experimental results showed promising predictive performance, particularly at temperatures of 260 °C ( R 2 = 0.6916 , correlation coefficient = 0.89697) and 365 °C ( R 2 = 0.76062 , correlation coefficient = 0.92887), indicating that the model accurately captures subtle variations in lignin composition under these conditions. However, variability in model performance at 280 °C and 345 °C highlights nonlinear characteristics of lignin decomposition that remain incompletely addressed, necessitating further data acquisition and model refinement.
Compared to conventional GC-MS methods, the proposed approach significantly reduces analytical costs and complexity due to lower initial investments and per-sample processing costs. This advantage is particularly beneficial for large-scale lignin characterization and real-time monitoring in industrial processes, such as biomass conversion and polymer development.
Overall, this study provides compelling evidence supporting the viability and potential impact of AI-driven predictive modeling using FT-IR and TGA data in lignin analysis. To further validate and enhance the generalizability of the proposed model, future research will expand datasets to include diverse lignin substrates and refine feature extraction methods.
In addition, systematic experimental validation through controlled laboratory-scale depolymerization experiments will be conducted, comparing predicted GC-MS profiles against actual measurements. Collaboration with organic chemists will specifically target the optimization of sample preparation methods, structural validation of predicted outcomes, and detailed chemical interpretation of observed nonlinearities at specific temperature ranges. These steps are expected to substantially strengthen the practical utility and robustness of the proposed analytical methodology.

Author Contributions

Conceptualization, M.P., S.-H.P. and D.-Y.K.; methodology, M.P., S.-H.P. and D.-Y.K.; software, M.P. and D.-Y.K.; validation, S.-H.P. and D.-Y.K.; formal analysis, M.P., S.-H.P. and D.-Y.K.; investigation, M.P., S.-H.P. and D.-Y.K.; resources, B.H.U.; data curation, B.H.U.; writing—original draft, M.P.; writing—review and editing, D.-Y.K.; visualization, M.P.; supervision, D.-Y.K.; project administration, S.-H.P. and D.-Y.K.; funding acquisition, S.-H.P. All authors have read and agreed to the published version of the manuscript.

Funding

This study was carried out with the support of ‘R&D Program for Forest Science Technology (Project No. 2023473C10-2425-EE02)’ provided by Korea Forest Service (Korea Forestry Promotion Institute).

Institutional Review Board Statement

Not applicable. This study did not involve human or animal subjects; therefore, no institutional review board approval was required.

Data Availability Statement

Data are available upon request to the corresponding authors.

Conflicts of Interest

Author SeungHyun Park was employed by the company Fivenode. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflicts of interest.

Appendix A

Table A1. Compounds identified through GC-MS analysis in lignin samples.
Table A1. Compounds identified through GC-MS analysis in lignin samples.
GroupsDerivativeChemical Formula
SyringylPhenol, 2,6-dimethoxy-C8H10O3
4-methylsyringol; 3,5-Dimethoxy-4-hydroxytolueneC9H12O3
4-Ethylsyringol; 4-Ethyl-2,6-dimethoxyphenolC10H14O3
4-Propylsyringol; 2,6-Dimethoxy-4-propylphenolC11H16O3
Syringaldehyde; Benzaldehyde, 4-hydroxy-3,5-dimethoxy-C9H10O4
(E)-4-Propenylsyringol (E)-2,6-Dimethoxy-4-(prop-1-en-1-yl) phenolC11H14O3
4-Acetylsyringol; Acetosyringon; Ethanone, 1-(4-hydroxy-3,5-dimethoxyphenyl)-C10H12O4
SyringylacetoneC11H14O4
Syringyl alcohol; 3,5-Dimethoxy-4-hydroxybenzeneethanolC10H14O4
ButylsyringoneC12H16O4
Acetyl syringic acid, ethyl esterC13H16O6
Propiosyringone; 1-Propanone, 1-(4-hydroxy-3,5-dimethoxyphenyl)-C11H14O4
Dihydrosyringenin; 3-SyringylpropanolC11H16O4
GuaiacylGuaiacol; Phenol, 2-methoxy-C7H8O2
5-Methylguaiacol; m-Creosol; 2-Methoxy-5-methylphenolC8H10O2
4-Ethylguaiacol; Phenol, 4-ethyl-2-methoxy-C9H12O2
4-Propylguaiacol; Phenol, 2-methoxy-4-propyl-C10H14O2
Benzaldehyde, 3-hydroxy-4-methoxy-C8H8O3
Allylguaiacol; EugenolC10H12O2
Guaiacylacetone; 2-Propanone, 1-(4-hydroxy-3-methoxyphenyl)-C10H12O3
4-(2-Hydroxyethyl)guaiacol; Homovanillyl alcoholC9H12O3
3-(4-guaiacyl)propanol; Benzenepropanol, 4-hydroxy-3-methoxy-C10H14O3
Poly aromaticsNaphthaleneC10H8
7-Methoxy-1-naphtholC11H10O2
2-Naphthalenol, 3-methoxy-C11H10O2
1,6-DimethoxynaphthaleneC12H12O2
Naphthalene, 2,3-dimethoxy-C12H12O2
1,6-DimethoxynaphthaleneC12H12O2
ReteneC18H18
2-Isopropyl-10-methylphenanthreneC18H18
Methyl dehydroabietateC21H30O2
8-Isopropyl-1,3-dimethylphenanthreneC19H20
Other aromaticsPhenolC6H6O
p-CresolC7H8O
o-Cresol; Phenol, 2-methyl-C7H8O
CreosolC8H10O2
CatecholC6H6O2
1-Propanone, 1-(5-methyl-2-thienyl)-C8H10OS
2-Acetyl-4-methylphenol; o-Acetyl-p-cresolC9H10O2
3-methoxycatechol; 1,2-Benzenediol, 3-methoxy-C7H8O3
HydroquinoneC6H6O2
4-Methylcatechol; 1,2-Benzenediol, 4-methyl-C7H8O2
3-Methylcatechol; 1,2-Benzenediol, 3-methyl-C7H8O2
Phenol, 4-methoxy-3-methyl-C8H10O2
2,3-DimethoxyphenolC8H10O3
Phenol, 3,4-dimethoxy-C8H10O3
5-Methoxy-m-cresol; 3-Methoxy-5-methylphenolC8H10O2
2,6-DimethoxyhydroquinoneC8H10O4
1,4-Benzenedicarboxaldehyde, 2-methyl- 2-MethylterephthalaldehydeC9H8O2
Ethanone, 1-(2-hydroxy-5-methylphenyl)-C9H10O2
Ethanone, 1-(2-hydroxy-6-methoxyphenyl)-C9H10O3
1,2,3-TrimethoxybenzeneC9H12O3
4-EthylcatecholC8H10O2
1,4-Benzenediol, 2,3,5-trimethyl- TrimethylhydroquinoneC9H12O2
Ethanone, 1-(2,3,4-trihydroxyphenyl)-C8H8O4
VanillinC8H8O3
3-Ethoxy-4-methoxyphenolC9H12O3
Phenol, 2-methoxy-4-(2-propenyl)-, acetate; Aceto eugenolC12H14O3
3-Acetylphenol; Ethanone, 1-(3-hydroxyphenyl)-C8H8O2
2-methoxy-5-acetylphenol; Ethanone, 1-(3-hydroxy-4-methoxyphenyl)-C9H10O3
ApocyninC9H10O3
Benzene, 1,2,3-trimethoxy-5-methyl-C10H14O3
2-Propanone, 1-(4-hydroxy-3-methoxyphenyl)-C10H12O3
3-Hydroxy-4-methoxybenzoic acidC8H8O4
FlopropioneC9H10O4
3,4-DimethoxyphenylacetoneC11H14O3
1-Propanone, 1-(4-hydroxy-3-methoxyphenyl)-C10H12O3
ButyrovanilloneC11H14O3
Homovanillic acidC9H10O4
Benzenepropanol, 4-hydroxy-3-methoxy-C10H14O3
Phenol, 2-methoxy-4-methyl-6-[propenyl]-C11H14O2
2,3-Dimethoxy-5-aminocinnamonitrileC11H12N2O2
5-(3-Hydroxypropyl)-2,3-dimethoxyphenolC11H16O4
AsaroneC12H16O3
Benzene, 1,2,3-trimethoxy-5-(2-propenyl)-C12H16O3
3,4-DivanillyltetrahydrofuranC20H24O5
1-(2,4-Dihydroxyphenyl)-2-(3,4-dimethoxyphenyl)ethan oneC16H16O5
1-(2,4-Dihydroxyphenyl)-2-(3,5-C17H18O5
DehydroabietateC20H28O2
3,4-Dimethoxyphenol, 2- methylpropionate-
Alkanes (Paraffins)Propane, 1,1-diethoxy-C7H16O2
1,3,5-TrioxaneC3H6O3
Propanal ethyl isopentyl acetal 1-(1-Ethoxypropoxy)-3-methylbutaneC10H22O2
CyclicOxazolidin-2-oneC3H5NO2
ButyrolactoneC4H6O2
2-Cyclopenten-1-one, 3-methyl-C6H8O
1,2-Cyclopentanedione, 3-methyl-C6H8O2
2-Cyclopenten-1-one, 2-hydroxy-3-methyl-C6H8O2
2-Cyclopenten-1-one, 2,3-dimethyl-C7H10O
Fatty AcidsPropanoic acidC3H6O2
Butanoic acid, 4-hydroxy-C4H8O3
Methyltartronic acidC4H6O5
Lactic acid; Propanoic acid, 2-hydroxy-, ethyl esterC5H10O3
Pentanoic acid, 4-oxo-C5H8O3
Pentanoic acid, 4-oxo-, ethyl esterC7H12O3
Butanoic acid, anhydrideC8H14O3
Butanoic acid, 2-methylpropyl esterC8H16O2
Propanoic acid, 2-methyl-, anhydrideC8H14O3
Pentanoic acid, 4-oxo-, 2-methylpropyl esterC9H16O3
Dodecanoic acid, methyl ester Pentanoic acid, 2-methyl-4-oxo-C13H26O2
Alcohols1,3-PropanediolC3H8O2
Ethanol, 2,2’-oxybis-C4H10O3
1,2-Propanediol, 3-methoxy-C4H10O3
1-Propanol, 2-(2-hydroxypropoxy)-C6H14O3
Glycerol-derived3-Ethoxy-1,2-propanediol; Glycerol 1-ethyl etherC5H12O3
Glycerol triethyl etherC9H20O3
1,3-Dioxolane-4-methanol, 2-ethyl-C6H12O3
GlycerinC3H8O3
1,2,3-Propanetriol, 1-acetateC5H10O4
Glycerol 1,2-diacetateC7H12O5
Alpha-monopropioninC6H12O4
Hydroxyacetone; 2-Propanone, 1-hydroxy-C3H6O2
Ethylene glycol Formate IsobutyrateC7H12O4
2,3-dihydroxypropyl isobutyrateC7H14O4

References

  1. Beaucamp, A.; Muddasar, M.; Amiinu, I.S.; Leite, M.M.; Culebras, M.; Latha, K.; Gutiérrez, M.C.; Rodriguez-Padron, D.; del Monte, F.; Kennedy, T.; et al. Lignin for energy applications—State of the art, life cycle, technoeconomic analysis and future trends. Green Chem. 2022, 24, 8193–8226. [Google Scholar] [CrossRef]
  2. Yu, O.; Kim, K.H. Lignin to materials: A focused review on recent novel lignin applications. Appl. Sci. 2020, 10, 4626. [Google Scholar] [CrossRef]
  3. Huang, Z.; Zhang, Y.; Zhang, C.; Yuan, F.; Gao, H.; Li, Q. Lignin-based composite film and its application for agricultural mulching. Polymers 2024, 16, 2488. [Google Scholar] [CrossRef]
  4. Mun, J.S.; Mun, S.P. Structural and thermal characterization of milled wood lignin from bamboo (Phyllostachys pubescens) grown in Korea. Molecules 2024, 29, 183. [Google Scholar] [CrossRef]
  5. Ragauskas, A.J.; Beckham, G.T.; Biddy, M.J.; Chandra, R.; Chen, F.; Davis, M.F.; Davison, B.H.; Dixon, R.A.; Gilna, P.; Keller, M.; et al. Lignin Valorization: Improving Lignin Processing in the Biorefinery. Science 2014, 344, 1246843. [Google Scholar] [CrossRef]
  6. Elliott, D.C.; Biller, P.; Ross, A.B.; Schmidt, A.J.; Jones, S.B. Hydrothermal liquefaction of biomass: Developments from batch to continuous process. Bioresour. Technol. 2015, 178, 147–156. [Google Scholar] [CrossRef]
  7. Zakzeski, J.; Bruijnincx, P.C.A.; Jongerius, A.L.; Weckhuysen, B.M. The Catalytic Valorization of Lignin for the Production of Renewable Chemicals. Chem. Rev. 2010, 110, 3552–3599. [Google Scholar] [CrossRef]
  8. Wang, Y.-Y.; Meng, X.; Pu, Y.; Ragauskas, A.J. Recent Advances in the Application of Functionalized Lignin in Value-Added Polymeric Materials. Polymers 2020, 12, 2277. [Google Scholar] [CrossRef]
  9. Blindheim, F.H.; Ruwoldt, J. The effect of sample preparation techniques on lignin Fourier transform infrared spectroscopy. Polymers 2023, 15, 2901. [Google Scholar] [CrossRef]
  10. Sebio-Puñal, T.; Naya, S.; López-Beceiro, J.; Tarrío-Saavedra, J.; Artiaga, R. Thermogravimetric analysis of wood, holocellulose, and lignin from five wood species. J. Therm. Anal. Calorim. 2012, 109, 1163–1167. [Google Scholar] [CrossRef]
  11. Takada, D.; Ehara, K.; Saka, S. Gas chromatographic and mass spectrometric (GC-MS) analysis of lignin-derived products from Cryptomeria japonica treated in supercritical water. J. Wood Sci. 2004, 50, 253–259. [Google Scholar] [CrossRef]
  12. Ehara, K.; Takada, D.; Saka, S. GC-MS and IR spectroscopic analyses of the lignin-derived products from softwood and hardwood treated in supercritical water. J. Wood Sci. 2005, 51, 256–261. [Google Scholar] [CrossRef]
  13. Guillén, M.D.; Ibargoitia, M.L. GC/MS analysis of lignin monomers, dimers and trimers in liquid smoke flavourings. J. Sci. Food Agric. 1999, 79, 1889–1903. [Google Scholar] [CrossRef]
  14. Washington State University. Analytical Chemistry Service Center–Equipment Rates. Available online: https://bsyse.wsu.edu/acsc/rates/ (accessed on 1 February 2025).
  15. Wen, Y.; Liu, X.; He, F.; Shi, Y.; Chen, F.; Li, W.; Song, Y.; Li, L.; Jiang, H.; Zhou, L.; et al. Machine Learning Prediction of Stalk Lignin Content Using Fourier Transform Infrared Spectroscopy in Large Scale Maize Germplasm. Int. J. Biol. Macromol. 2024, 280, 136140. [Google Scholar] [CrossRef]
  16. Ge, H.; Liu, Y.; Zhu, B.; Xu, Y.; Zhou, R.; Xu, H.; Li, B. Machine Learning Prediction of Delignification and Lignin Structure Regulation of Deep Eutectic Solvents Pretreatment Processes. Ind. Crops Prod. 2023, 203, 117138. [Google Scholar] [CrossRef]
  17. Diment, D.; Löfgren, J.; Alopaeus, M.; Stosiek, M.; Cho, M.; Xu, C.; Hummel, M.; Rigo, D.; Rinke, P.; Balakshin, M. Enhancing Lignin-Carbohydrate Complexes Production and Properties with Machine Learning. ChemSusChem 2024, e202401711. [Google Scholar] [CrossRef] [PubMed]
  18. Löfgren, J.; Tarasov, D.; Koitto, T.; Rinke, P.; Balakshin, M.; Todorovic, M. Machine Learning Optimization of Lignin Properties in Green Biorefineries. ACS Sustain. Chem. Eng. 2022, 10, 9469–9479. [Google Scholar] [CrossRef]
  19. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  20. Luo, W.; Yang, M.; Zheng, W. Weakly-supervised semantic segmentation with saliency and incremental supervision updating. Pattern Recognit. 2021, 115, 107858. [Google Scholar] [CrossRef]
  21. Lemley, J.; Bazrafkan, S.; Corcoran, P. Smart augmentation: Learning an optimal data augmentation strategy. IEEE Access 2017, 5, 5858–5869. [Google Scholar] [CrossRef]
  22. Montesinos López, O.A.; Montesinos López, A.; Crossa, J. (Eds.) Overfitting, model tuning, and evaluation of prediction performance. In Multivariate Statistical Machine Learning Methods for Genomic Prediction; Springer International Publishing: Berlin/Heidelberg, Germany, 2022; pp. 109–139. [Google Scholar] [CrossRef]
  23. Lin, M.; Chen, Q.; Yan, S. Network in Network. arXiv 2014, arXiv:1312.4400. [Google Scholar] [CrossRef]
  24. Suzuki, M.; Nakayama, K.; Matsuo, Y. Improving Bi-directional Generation between Different Modalities with Variational Autoencoders. arXiv 2018, arXiv:1801.08702. [Google Scholar] [CrossRef]
  25. Jacobs, R.A.; Jordan, M.I.; Nowlan, S.J.; Hinton, G.E. Adaptive Mixtures of Local Experts. Neural Comput. 1991, 3, 79–87. [Google Scholar] [CrossRef]
  26. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar] [CrossRef]
  27. Kullback, S.; Leibler, R.A. On Information and Sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  28. Aboussalah, A.M.; Kwon, M.J.; Patel, R.G.; Chi, C.; Lee, C.G. Don’t overfit the history—Recursive time series data augmentation. arXiv 2023, arXiv:2207.02891. [Google Scholar] [CrossRef]
  29. Lim, S.S.; Kang, B.O.; Kwon, O.W. Improving transformer-based speech recognition performance using data augmentation by local frame rate changes. J. Acoust Soc. Korea 2022, 41, 122–129. [Google Scholar] [CrossRef]
  30. Tian, L.; Wang, Z.; Liu, W.; Cheng, Y.; Alsaadi, F.E.; Liu, X. A new GAN-based approach to data augmentation and image segmentation for crack detection in thermal imaging tests. Cogn. Comput. 2021, 13, 1263–1273. [Google Scholar] [CrossRef]
  31. Dunphy, K.; Fekri, M.N.; Grolinger, K.; Sadhu, A. Data augmentation for deep-learning-based multiclass structural damage detection using limited information. Sensors 2022, 22, 6193. [Google Scholar] [CrossRef]
  32. Oh, C.; Han, S.; Jeong, J. Time-series data augmentation based on interpolation. Procedia Comput. Sci. 2020, 175, 64–71. [Google Scholar] [CrossRef]
  33. Botchkarev, A. Performance metrics (error measures) in machine learning regression, forecasting and prognostics: Properties and typology. arXiv 2018, arXiv:1809.03006. [Google Scholar] [CrossRef]
  34. Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
  35. Draper, N.R.; Smith, H. Applied Regression Analysis, 3rd ed.; Wiley-Interscience: New York, NY, USA, 1998; ISBN 978-0-471-17082-2. [Google Scholar] [CrossRef]
  36. Asuero, A.G.; Sayago, A.; González, A.G. The Correlation Coefficient: An Overview. Crit. Rev. Anal. Chem. 2006, 36, 41–59. [Google Scholar] [CrossRef]
  37. Villani, C. Topics in Optimal Transportation; American Mathematical Society: Providence, RI, USA, 2003. [Google Scholar] [CrossRef]
Figure 1. Overall research workflow.
Figure 1. Overall research workflow.
Polymers 17 00806 g001
Figure 2. TGA data preprocessing pipeline.
Figure 2. TGA data preprocessing pipeline.
Polymers 17 00806 g002
Figure 3. FT-IR data preprocessing pipeline.
Figure 3. FT-IR data preprocessing pipeline.
Polymers 17 00806 g003
Figure 4. The proposed GC-MS data prediction model architecture.
Figure 4. The proposed GC-MS data prediction model architecture.
Polymers 17 00806 g004
Table 1. The proposed TGA data generation model architecture.
Table 1. The proposed TGA data generation model architecture.
ConfigurationOutput Shape
Feature Extractor (CNN)
Conv1D(1, 16, kernel=3, stride=1, padding=1)→ ReLU(batch, 16, input_size)
Conv1D(16, 32, kernel=3, stride=1, padding=1)→ ReLU(batch, 32, input_size)
Flatten (batch, 32 × input_size)
Fully Connected Layers
Linear (32 × input_size, 1024)→ ReLU(batch, 1024)
Linear (1024, output_size) (batch, 761)
Table 2. The proposed FT-IR data generation model architecture.
Table 2. The proposed FT-IR data generation model architecture.
ConfigurationOutput Shape
Feature Extractor (CNN)
Conv1D (1, 16, kernel=3, stride=1, padding=1)→ ReLU(batch, 16, input_size)
Conv1D (16, 32, kernel=3, stride=1, padding=1)→ ReLU(batch, 32, input_size)
Conv1D (32, 64, kernel=3, stride=1, padding=1)→ ReLU(batch, 64, input_size)
Flatten (batch, 64 × input_size)
Fully Connected Layers
Linear (64 × 1, 1024)→ ReLU(batch, 1024)
Dropout (0.3) (batch, 1024)
Linear (1024, 3476) (batch, 3476)
Table 3. The proposed GC-MS data generation model architecture.
Table 3. The proposed GC-MS data generation model architecture.
ConfigurationOutput Shape
Feature Extractor (CNN)
Conv1D (1, 64, kernel=3, padding=1)→ ReLU(batch, 64, input_size)
Conv1D (64, 64, kernel=3, padding=1)→ ReLU(batch, 64, input_size)
Conv1D (64, 64, kernel=3, padding=1)→ ReLU(batch, 64, input_size)
Global Average Pooling (batch, 64)
Fully Connected Layers
Linear (64, 128)→ ReLU(batch, 128)
Linear (128, 10)→ Softmax(batch, 10)
Table 4. Wasserstein distance between preprocessed and original TGA data.
Table 4. Wasserstein distance between preprocessed and original TGA data.
TemperatureWasserstein Distance
250 °C 1.77 × 10 3
300 °C 1.406 × 10 3
350 °C 1.036 × 10 3
400 °C 1.215 × 10 3
Table 5. Evaluation of the TGA data generation model at various temperatures.
Table 5. Evaluation of the TGA data generation model at various temperatures.
TemperatureMAE R 2 Correlation
260 °C 1.0 × 10 5 0.998460.99927
280 °C 3.0 × 10 5 0.996770.99856
315 °C 1.0 × 10 5 0.999360.99974
345 °C 3.0 × 10 5 0.996950.99895
365 °C 3.0 × 10 5 0.997750.99930
390 °C 1.0 × 10 5 0.999850.99994
Table 6. Evaluation of the FT-IR generation model at various temperatures.
Table 6. Evaluation of the FT-IR generation model at various temperatures.
TemperatureMAE R 2 Correlation
260 °C 2.8 × 10 4 0.999820.99991
280 °C 4.9 × 10 4 0.999690.99986
315 °C 9.3 × 10 4 0.999420.99973
345 °C 1.4 × 10 3 0.999180.99961
365 °C 1.0 × 10 3 0.999360.99968
390 °C 4.1 × 10 4 0.999740.99987
Table 7. Evaluation of the GC-MS generation model at various temperatures.
Table 7. Evaluation of the GC-MS generation model at various temperatures.
TemperatureMAE R 2 Correlation
260 °C 1.2 × 10 4 0.984180.99341
280 °C 2.3 × 10 3 0.641810.82605
315 °C 1.3 × 10 4 0.984450.99749
345 °C 1.3 × 10 4 0.986450.99837
365 °C 4.2 × 10 4 0.963080.98587
390 °C 6.8 × 10 4 0.97560.99835
Table 8. Prediction results of the GC-MS model at various temperatures (values represent relative proportions).
Table 8. Prediction results of the GC-MS model at various temperatures (values represent relative proportions).
TemperatureSyringylGuaiacylPoly Aromatics
(C10–C21)
Other Aromatics
(C6–C20)
AlkanesCyclicFatty AcidsAlcoholGlycerol-
Derived
Other
260 °C0.214770.176460.056850.295640.015790.040390.050280.013910.074100.15986
280 °C0.058520.107540.025210.330850.003430.077560.004240.020820.192080.09062
315 °C0.169340.091270.048280.285940.008310.028450.003640.013010.166180.07224
345 °C0.077580.062170.052890.307540.010260.030290.072950.017830.116120.07224
365 °C0.150260.141060.014340.322780.003120.002990.088850.050810.230300.02578
390 °C0.141200.129630.006290.261870.088330.027040.017670.016060.107730.09065
Table 9. Evaluation of the prediction results of the GC-MS model.
Table 9. Evaluation of the prediction results of the GC-MS model.
TemperatureMAE R 2 Correlation
260 °C 2.07 × 10 3 0.69160.89697
280 °C 4.87 × 10 2 0.340320.69668
315 °C 4.00 × 10 3 0.395810.76097
345 °C 1.04 × 10 2 −0.318190.40104
365 °C 2.19 × 10 3 0.760620.92887
390 °C 9.72 × 10 3 0.518350.72818
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, M.; Um, B.H.; Park, S.-H.; Kim, D.-Y. Exploring the Feasibility of Deep Learning for Predicting Lignin GC-MS Analysis Results Using TGA and FT-IR. Polymers 2025, 17, 806. https://doi.org/10.3390/polym17060806

AMA Style

Park M, Um BH, Park S-H, Kim D-Y. Exploring the Feasibility of Deep Learning for Predicting Lignin GC-MS Analysis Results Using TGA and FT-IR. Polymers. 2025; 17(6):806. https://doi.org/10.3390/polym17060806

Chicago/Turabian Style

Park, Mingyu, Byung Hwan Um, Seung-Hyun Park, and Dae-Yeol Kim. 2025. "Exploring the Feasibility of Deep Learning for Predicting Lignin GC-MS Analysis Results Using TGA and FT-IR" Polymers 17, no. 6: 806. https://doi.org/10.3390/polym17060806

APA Style

Park, M., Um, B. H., Park, S.-H., & Kim, D.-Y. (2025). Exploring the Feasibility of Deep Learning for Predicting Lignin GC-MS Analysis Results Using TGA and FT-IR. Polymers, 17(6), 806. https://doi.org/10.3390/polym17060806

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop