Enhancing Early Lung Cancer Diagnosis: Predicting Lung Nodule Progression in Follow-Up Low-Dose CT Scan with Deep Generative Model

Simple Summary Detecting lung cancer early and initiating treatment promptly can greatly enhance patient outcomes. While low-dose computed tomography (LDCT) screening aids in identifying lung cancer at an early stage, there is a risk of diagnostic delays as patients await follow-up scans. To mitigate this challenge, we developed a deep predictive model leveraging generative AI methods to forecast nodule growth patterns in follow-up LDCT scans based on baseline LDCT scans. Our findings illustrated that utilizing the predicted follow-up nodule images generated by our model during baseline screening improved diagnostic accuracy compared to using baseline nodules alone and achieved comparable performance with using real follow-up nodules. This demonstrated the potential of employing deep generative models to forecast nodule appearance in follow-up imaging from baseline LDCT scans, thereby enhancing risk assessment during initial screening. Abstract Early diagnosis of lung cancer can significantly improve patient outcomes. We developed a Growth Predictive model based on the Wasserstein Generative Adversarial Network framework (GP-WGAN) to predict the nodule growth patterns in the follow-up LDCT scans. The GP-WGAN was trained with a training set (N = 776) containing 1121 pairs of nodule images with about 1-year intervals and deployed to an independent test set of 450 nodules on baseline LDCT scans to predict nodule images (GP-nodules) in their 1-year follow-up scans. The 450 GP-nodules were finally classified as malignant or benign by a lung cancer risk prediction (LCRP) model, achieving a test AUC of 0.827 ± 0.028, which was comparable to the AUC of 0.862 ± 0.028 achieved by the same LCRP model classifying real follow-up nodule images (p = 0.071). The net reclassification index yielded consistent outcomes (NRI = 0.04; p = 0.62). Other baseline methods, including Lung-RADS and the Brock model, achieved significantly lower performance (p < 0.05). The results demonstrated that the GP-nodules predicted by our GP-WGAN model achieved comparable performance with the nodules in the real follow-up scans for lung cancer diagnosis, indicating the potential to detect lung cancer earlier when coupled with accelerated clinical management versus the current approach of waiting until the next screening exam.


Introduction
Lung cancer is a leading cause of cancer-related death for men and women worldwide, with a poor overall 5-year survival rate of 22.9% [1].Early diagnosis and treatment of lung cancer can significantly improve patient survival.For patients diagnosed at an early stage, the five-year survival rate can reach about 61.2% [2].In contrast, curative treatments are no longer effective for advanced-stage lung cancer, resulting in a five-year survival rate of only 16.0%.
Studies have shown that screening with low-dose computed tomography (LDCT) can help detect lung cancer at an early stage and reduce lung cancer mortality in people at high risk [3].However, the LDCT screening also detects many nodules that are considered to be indeterminate.Although the nodule size, shape, and density manifested on the LDCT images are correlated with the risk of malignancy, a single LDCT scan at initial screening (baseline) may not provide definitive information, so follow-up is often required to evaluate the stability of the nodule over time [4].With follow-up LDCT scans, radiologists can visualize how the nodule changes over time and make more accurate and informed decisions regarding patient management, including the need for further evaluation, biopsies, or follow-up scans.
Besides the increased cost and additional radiation, patients may experience delays in diagnosis as they wait for follow-up LDCT scans.A potential method is to harness the power of advanced artificial intelligence (AI) techniques to analyze the lung nodule characteristics manifested on CT scans and predict their progression in subsequent followup CT scans.However, the prediction of lung nodule evolution is challenging because of the heterogeneous and multifactorial nature of lung nodules [5].Traditional mathematical models have limited predictive power as they usually use simplistic formulations, for example, linear, quadratic, power-law, and exponential models [6,7], and only take into consideration physical factors of nodules such as size, mass, and volume, and thus may not be able to adequately capture the biological complexity of nodules.Machine learning-based radiomics methods have demonstrated the potential to predict the likelihood of lung nodule growth [8,9].However, these methods relied on effective radiomic features [8,9], including manually engineered features [10,11], that demand expertise in both engineering and domain knowledge to interpret data patterns and design feature extractors [12].Leveraging the advances of AI techniques, deep-learning-based methods such as Convolutional Long Short-Term Memory (Conv-LSTM) [13], U-Nets [14], and deep spatial transformation process [15] demonstrated the ability to characterize complex image patterns of nodules detected in CT scans.While the aforementioned methods have shown promise in predicting the risk of malignancy and the appearance of nodules in future exams for lung cancer diagnosis, they required prior CT scans to detect interval changes in the nodule between current and prior CT scans.This prerequisite renders these methods useless when prior scans are unavailable, and indeterminate nodules have to undergo at least one followup scan.
Generative models have been increasingly explored in medical imaging applications for different tasks, such as image synthesis, reconstruction, noise reduction, segmentation, and classification [16][17][18][19][20].The generative adversarial network (GAN) has emerged as a powerful tool for data augmentation by generating new samples that are similar to the original data, for example, increasing the size and diversity of a training dataset for classification and image segmentation [18,19].It has also been employed for cross-modality image synthesis that can transform the image of one modality to another modality, for example, by synthesizing magnetic resonance images from CT images [21].
In this study, we developed a deep-learning-based generative model to discover patterns of nodule growth on longitudinal CT scans during the training process and predict their progression without requiring prior scans at deployment.With time serial data collected from the National Lung Screening Trial (NLST) study [3], we developed a Growth Predictive model based on the Wasserstein Generative Adversarial Network training framework, referred to as the GP-WGAN model, to learn the growth pattern from pairs of nodule images in LDCT scans acquired at baseline (T0) and follow-up screening years (T1, T2).We designed new loss functions to guide detailed pixel-wise synthesis, preserve structural similarity, and enhance the perceptual quality of the generated nodules with reference to the follow-up images in future LDCT scans.
To the best of our knowledge, we are the first to develop a deep generative model with the ability to predict lung nodule images in future follow-up LDCT scans from images of current-year LDCT scans.Through the combination of four different loss functions that were designed to measure specific errors, the follow-up nodule images predicted from baseline scans by the GP-WGAN model achieved comparable performance with the actual nodules in follow-up LDCT scans in predicting lung cancer risks.This demonstrates the feasibility of using deep generative models to improve the clinical management of screening-detected pulmonary nodules, facilitating earlier diagnosis of lung cancer.

Data Sets
This retrospective study was approved by the Institutional Review Board (IRB), and informed consent was waived.The NLST was a randomized controlled trial that randomly assigned participants to receive three annual screenings (T0, T1, and T2) with either LDCT or chest X-rays.The participants who were diagnosed with lung cancer through biopsy confirmation and required treatment were not offered further screening LDCT scans.With permission from the NLST, we collected 2500 anonymized subjects containing digital LDCT files, including all 639 NLST-reported biopsy-confirmed lung cancer cases, and 1861 randomly selected benign nodule cases based on 3 years annual LDCT exams and/or up to seven years of non-imaging follow-up.More information about the LDCT scans is described in Appendix A. Among the 2500 subjects, 1226 subjects who had at least one nodule with a size ranging from 4 mm to 30 mm found in their first-year baseline LDCT scans and had corresponding follow-up LDCT scans were included in this study.Of these 1226 subjects, 218 were diagnosed with lung cancer and 1008 were negative.We randomly split 1226 subjects into a training/validation set comprising 776 subjects (165 positive and 611 negative) and a test set containing 450 subjects (53 positive and 397 negative) (Table 1).From the 776 subjects in the training/validation set who underwent baseline and annual follow-up scans up to 2 years, a total of 1121 pairs of image patches (776 pairs between T0 and T1, 345 pairs between T1 and T2) containing nodules (223 positive and 898 negative) from LDCT scans were used to develop the GP-WGAN model.In the test set, only the baseline scans and their corresponding T1 scans were included, comprising a total of 450 (53 positive and 397 negative) pairs of nodule image patches.Among the 53 positive subjects, 42 and 11 were diagnosed with lung cancer at T1 and T2 follow-up LDCT screening, respectively.All 397 negative subjects had undergone 2 years of follow-up LDCT exams.

Study Subject Characteristics
The distribution of the NLST documented subject demographic data and clinical radiologic factors was summarized in Table 1.

Data Preparation
An experienced cardiothoracic radiologist re-examined each NLST-documented lung nodule and manually marked the corresponding nodule centers on the baseline and followup scans for each subject.In a subject with multiple nodules detected at the baseline scan, the nodule with the largest size/growth during follow-up was identified by the radiologist and used in this study [8,9].All LDCT scans were resampled to isotropic volumes with a voxel size of 0.5 × 0.5 × 0.5 mm 3 using the 3D spline interpolation method [22].An experienced cardiothoracic radiologist manually selected the slice for each nodule to ensure that the selected slice correctly manifested the nodule's characteristics.A 2D region of interest (ROI) on the slide with a side length of 32 mm centered at the radiologist's manually marked nodule center was extracted and used as the nodule image patch.

Growth Predictive Model Based on the Wasserstein Generative Adversarial Network
We developed a GP-WGAN model with a deep predictor network and a deep discriminator network to predict a nodule image in the follow-up LDCT scan from their baseline LDCT scan. Figure 1 shows the adversarial framework for the training of the GP-WGAN model.The predictor network was implemented with a modified U-Net architecture [23], and the discriminator network was constructed with a five-layer deep convolutional neural network (DCNN) structure.The structures of the predictor network and the discriminator network are described in Appendix B. The pairs of nodule ROI images and their corresponding follow-up images at 1-year intervals (T0-T1 pairs or T1-T2 pairs if available) were input to the predictor and discriminator networks to train the GP-WGAN model.The real follow-up images were used as the target to guide the predictor in generating images that "mimic" the nodules in their follow-up scans.With the goal of minimizing the Wasserstein adversarial loss, the task of the discriminator was to distinguish the real follow-up images from the predicted images, provide feedback (losses) to the predictor network, and update the weights of the networks.We trained the predictor and discriminator networks iteratively by minimizing a weighted combination of losses.The predictor aimed to synthesize more realistic images that the discriminator could not distinguish from the real follow-up images.The discriminator, on the other hand, continuously improved its ability to distinguish between the real and predicted images.This adversarial training process created a feedback loop between the predictor and discriminator.Once the GP-WGAN model was trained, only the predictor network with frozen weights was used to predict the nodule images in future LDCT scans.The discriminator network was no longer needed during deployment.

Generative Loss Function
We designed a generative loss function (L G ) in our GP-WGAN model that leveraged L 1 loss for pixel-wise synthesis accuracy, structural similarity index loss (L SSIM ) [24] for preserving structure information, a learned perceptual [25] loss (L LP ) for capturing highlevel visual quality, and an adversarial loss (L A ) to ensure realistic outputs: where λ 1 , λ 2 , λ 3 , and λ 4 were hyper-parameters that weighted the four losses.The calculation of L G involved using the pair of the real nodule image X and its follow-up X F , and the "mimic" image G(X) generated by the predictor network.More detailed descriptions of each loss are included in Appendix C.

Discriminator Loss Function
The loss of the discriminator was measured by the Wasserstein distance with gradient penalty [26] as follows: where x = tX F + (1 − t)G(X) was an interpolated image, t ∼ Unif([0, 1]), λ D was the penalty weight, and ∇ x represented the gradient with respect to x.

Performance Evaluation and Statistical Analysis
Using the independent test set consisting of 450 subjects (53 positive and 397 negative), we evaluated the potential of the GP-WGAN model in improving the early diagnosis of lung cancer by predicting follow-up nodule images from baseline LDCT nodule images.We deployed the trained GP-WGAN to the ROI images of the 450 nodules in the baseline LDCT scans to predict the GP-nodules in the 1-year follow-up LDCT scans.A lung cancer risk prediction (LCRP) model developed with serial radiomics-based reinforcement learning [27] was directly applied to the 450 GP-nodules without retraining to predict the risk of malignancy for each nodule.We compared the LCRP model to predict the risk for "virtual" GP-nodules versus the real nodules from the baseline and 1-year follow-up LDCT scans.These predicted risks were also compared to those predicted by the Lung-RADS and the Brock model estimated from the patient's risk factors [28,29], such as demographic data and radiologic descriptions of nodules provided by the NLST dataset [3].
The receiver operating characteristic (ROC) analysis [30][31][32] and reclassification benefit analysis [33] were used for performance evaluation.The ROC curves of the LCRP model and the Brock model for nodule classification were compared using the method of DeLong et al. [34].The Hochberg correction [35] was employed to adjust the p-values for multiple comparisons.The p-values were adjusted using the R software function "p.adjust", and less than 0.05 after adjustment was considered statistically significant.The net reclassification index (NRI) [34] was used to assess the net gain or loss from the correct and incorrect risk escalation or de-escalation by the proposed method (i.e., LCRP model with GP-nodules as input) in comparison with the reference models that used the baseline nodules to stratify cancer risk.More detailed information and limitations of NRI are provided in Appendix D. The groups of low-, medium-, and high-risk were stratified by thresholding the Lung-RADS scores (<3, =3, and >3) derived from previous studies [36,37] and the scores of the Brock model (<0.0117, [0.0117, 0.10], >0.10) derived from the British Thoracic Society guideline [38,39].The thresholds of the LCRP models (<0.45, [0.45, 0.81], >0.81) were determined so that the subgroup sizes by the LCRP model classifying GP-nodules would align with the subgroup sizes stratified by the Brock model [33].The statistical significance of NRI was tested by Z-statistic [40].The GP-WGAN model was developed using Python 3.6.9. and PyTorch 1.8.1.The ROC and other statistical analyses were performed by using the statistical software package ORDBM MRMC 3.0 in Java [41].

Results
Figure 2 shows examples of six lung nodules on the baseline (T0), 1-year follow-up (T1) LDCT scans, and GP-WGAN predicted GP-nodules.Examples of three benign (left) and three malignant (right) lung nodules from 6 subjects on baseline (T0) and 1-year follow-up (T1) LDCT scans.Each row presents an example nodule.For each nodule, the 1st column shows the ROI image of the nodule at T0, 2nd column shows its follow-up image at T1, and the 3rd column shows the GP-nodules generated from the T0 nodule shown in the 1st column by the GP-WGAN model.(Left): the benign nodules in T0, T1, and their predicted GP-nodules showed stability in size, attenuation, and smooth margins.(Right): the malignant nodules showed a trend of enlarged sizes in T1 and GP-nodules.
Compared with the follow-up nodule images in the independent test set, GP-nodules achieved a Mean Square Error (MSE) of 0.024 and a Structural Similarity Index (SSIM) of 0.860.Figure 3 shows the test ROC curves, and Table 2 shows the test AUC achieved by our LCRP model and Brock model for the classification of 450 nodules in the independent test set.The LCRP model classifying real follow-up (T1) nodules achieved a test AUC of 0.862 ± 0.028.In comparison, the LCRP model classifying GP-nodules that were predicted from the baseline nodules achieved a comparable test AUC of 0.827 ± 0.028 (p = 0.071).The test AUC by both the LCRP model with baseline nodules (0.805 ± 0.031; p = 0.024) and the Brock model with baseline nodules (0.754 ± 0.035; p < 0.001) were significantly lower than the LCRP model with real follow-up nodules.
For early diagnosis of lung nodules from baseline scans, the LCRP model classifying GP-nodules achieved a significantly higher performance (p = 0.043) compared to the Brock model using real baseline nodules and only achieved a comparable performance (p = 0.099) compared to the LCRP model classifying real baseline nodules.
Table 2 shows that, for the classification of three subgroups of nodules with predominantly solid attenuation, spiculated margin, or nodules with size ranging from 6 to 14 mm in diameter, the LCRP model with real follow-up nodules achieved a test AUC of 0.864 ± 0.034, 0.922 ± 0.037, and 0.826 ± 0.039, respectively.In comparison, the LCRP model with GP-nodules achieved a comparable test AUC of 0.828 ± 0.037 (p = 0.091), 0.850 ± 0.055 (p = 0.150), and 0.782 ± 0.041 (p = 0.077), respectively.The test AUCs achieved by both the LCRP model with baseline nodules and the Brock model with baseline nodules were all significantly lower than the LCRP model with real follow-up nodules.
Table 2 also shows that, for early diagnosis of the same three subgroups, the LCRP model with the predicted GP-nodules achieved a significantly higher performance (p = 0.045, p < 0.001, p = 0.048, respectively) than the Brock model with real baseline nodules, whereas the LCRP model with real baseline nodules was comparable to the Brock model for solid nodules (p = 0.249) and nodules with size from 6 to 14 mm in diameter (p = 0.201).All the differences in the test AUCs between the LCRP model with GP-nodules and the LCRP model with real baseline nodules were not found to be significant for the three subgroups.to the LCRP model using real baseline nodules (NRI = 0.20, p = 0.004).The difference in risk stratification compared to the LCRP model using real follow-up nodules did not reach statistical significance (NRI = 0.04, p = 0.62).

Discussion
Studying the evolution of lung nodules over time is essential for assessing the risk of lung cancer for patients with lung nodules detected on CT scans.Accurate and reliable risk stratification plays a pivotal role in the management of lung cancer that enables tailored approaches to screening, diagnosis, and treatment for personalized care (43).In this study, we developed a deep-learning-based generative GP-WGAN model that used lung nodule images in baseline LDCT scans to predict their growth or stability in 1-year follow-up LDCT scans.The results showed that the predicted 1-year follow-up nodule images by the GP-WGAN model achieved comparable performance with the corresponding real nodule images from 1-year follow-up scans for lung cancer diagnosis.Using GP-nodules to replace the baseline nodules, the LCRP model achieved a comparable test AUC of 0.827 with those using the real nodules either from the T0 baseline (AUC = 0.805, p = 0.099) or the T1 followup (AUC = 0.862, p = 0.071) LDCT scans, and significantly outperformed the Brock model (AUC = 0.754, p = 0.043) for early diagnosis of baseline nodules.The net benefit analysis (Table 3) showed that the reclassification of the baseline nodules by the LCRP model using GP-nodules consistently and significantly improved the risk stratification compared to the classification by Lung-RADS (NRI = 0.38, p < 0.001), the Brock model (NRI = 0.20, p = 0.03) and the LCRP model (NRI = 0.20, p = 0.004) using the real baseline nodules.
The findings demonstrated the feasibility of our proposed method in predicting the growth patterns of nodules with generated nodule images in follow-up exams.The GPnodules enhanced the capability of a lung cancer risk prediction model in risk stratification at the baseline year.When compared with the LCRP model, Brock model, and Lung-RADS using real baseline nodules, the LCRP model with GP-nodules had a net gain in the identification of high-risk cases that could potentially impact mortality through early interventions.Furthermore, the GP-nodules improved the performance of the LCRP model in the classification of indeterminate nodules that were solid, spiculated, or within the 6 to 14 mm diameter range, as illustrated in Table 2.As these nodules would likely undergo follow-up exams in current clinical practice, better risk stratification is pivotal for reducing the indeterminate decision.Another advantage of generating the GP-nodules by our proposed method is it takes one step toward an explainable AI approach that may increase the radiologist's understanding of the risk prediction by the LCRP model and enable them to make more appropriate diagnostic decisions after taking into consideration the AI recommendation and their own judgment.Therefore, our proposed approach has the potential to reduce unnecessary follow-up exams and improve early diagnosis of lung cancer at baseline screening.This is expected to have clinical significance as it may not only reduce patient anxiety and the associated healthcare costs but also aid the physician in making more informed decisions about the necessary level of aggressiveness in subsequent management steps.Moreover, the ability to generate GP-nodules may enable other computerized methods, for example, nodule segmentation and temporal analysis, to be sequentially used for specific tasks.
Designing an effective loss function is essential for developing image-generative models due to the inherent difficulties in determining image similarity, especially when predicting a future image in a stochastic process.We designed a generative loss function (L G ) for the GP-WGAN model by combining four loss functions.In the combined loss functions, the L1 loss is a widely used measurement in machine learning to minimize error by capturing low-frequency patterns that are relatively uniform with small changes.L1 has been demonstrated to improve image quality in various generative models for image generation and image-to-image translation [42,43].Given the intrinsic correlation of objects in the image, a major limitation of L1 loss is its assumption of pixel independence in an image.Relying solely on this metric without considering structural information may not be able to reveal specific aspects of nodule growth patterns over time.A recent study [42] showed that using structural similarity index (SSIM) as a loss function during network training can preserve contrast in high-frequency regions compared to the L1 loss function.Other studies further demonstrated that the combination of L1 with SSIM outperformed standalone SSIM [42,44].Besides combining L1 and SSIM in our generative loss function, we also incorporated a learned perceptual loss ( L LP ) that utilized a pre-trained Resnet- 18 model.The L LP could provide a more accurate representation of human perception, serving as a valuable metric for assessing dissimilarities in a high-level feature space that encompasses disparities in both content and style between images.The results showed that the GP-WGAN model with the newly designed loss function could generate more realistic nodule images through assessing the image quality of the generated nodules at the pixel level, structural level, and high semantic feature level.
Artificial intelligence, particularly in the form of generative models, has garnered significant interest and activity in recent years.However, owing to the lack of real physical models within the general framework, there is a growing concern about the reliability and ethical implications of "black-box" models, as well as the integrity of the data they generate.Challenges include ensuring data reliability and transparency, and developing methods to mitigate potential bias and discrimination.A considerable amount of work must be undertaken before these systems are ready for clinical applications, which includes conducting extensive external validation and prospective studies.
There are limitations in this study.Despite including all malignant nodule cases from the NLST dataset, which is relatively large compared to most studies in medical imaging [18], the subset that satisfied the requirements of the current study is still small when compared with the big data set used in conventional GAN-related research in the computer vision field.This may limit the learning of nodule progression patterns over time.For example, malignant nodules in our data set exhibited a wide variety of volume doubling times ranging from 6 months to several years.The relatively small training sample size of malignant nodules with a wide range of progression paces might result in a trained GP-WGAN model having a limited capability in predicting the nodule growth rates.Some examples are included in Appendix E. Another limitation is due to the lack of effective metrics to quantify the similarity between the virtual and real images.In this study, we adopted a task-based assessment approach by using a previously developed LCRP model as the downstream task to evaluate how the GP-nodules can improve the prediction of lung cancer risk.Further work is underway to conduct external validation, reader studies with radiologists, and prospective studies to assess the performance of the proposed model.

Conclusions
In conclusion, we developed a GP-WGAN model to predict lung nodule images in 1-year follow-up LDCT scans based on the baseline LDCT scans and studied its potential to improve the early diagnosis of lung cancer at the baseline screening year.The results demonstrated that the follow-up nodule images predicted by GP-WGAN could achieve comparable performance with the real nodules in follow-up LDCT scans, which indicates the use of the GP-WGAN model provides the opportunity to accelerate the clinical management of malignant nodules towards an earlier diagnosis of lung cancer, rather than waiting for the next screening CT.

Appendix C. Losses in the Generative Loss Function
We designed a generative loss function (L G ) in our GP-WGAN model that leveraged L 1 loss for pixel-wise synthesis accuracy, structural similarity index loss (L SSIM ) for preserving structure information, a learned perceptual loss (L LP ) for capturing high-level visual quality and an adversarial loss (L A ) to ensure realistic outputs.The calculation of L G involved using the pair of the real nodule image X and its follow-up X F , and the "mimic" image G(X) generated by the predictor network: (1) L 1 loss measured the average absolute pixel-wise difference between the predicted and the real follow-up images: where E (X, X F ) represents the mathematical expectation with respect to random variable pair (X, X F ).
(2) Structural similarity index (SSIM) loss (L SSIM ) quantified the degradation of structural information between two images (x and y) with three comparative measures: luminance l, contrast c, and structure s: where (µ x , µ y ), (σ x , σ y ), and σ xy were the mean, variance, and covariance of pixel values in images x and y, respectively, and C 1 , C 2 , C 3 were hyper-parameters.The SSIM loss was defined as a weighted combination of these comparative measures: (3) Learned perceptual loss (L LP ) was a perceptual metric for assessing dissimilarities in feature spaces that encompass disparities in content and style discrepancies between images.We utilized a pre-trained modified Resnet-18 model ( 26

Figure 1 .
Figure 1.The adversarial framework for training the GP-WGAN model.

Figure 2 .
Figure 2.Examples of three benign (left) and three malignant (right) lung nodules from 6 subjects on baseline (T0) and 1-year follow-up (T1) LDCT scans.Each row presents an example nodule.For each nodule, the 1st column shows the ROI image of the nodule at T0, 2nd column shows its follow-up image at T1, and the 3rd column shows the GP-nodules generated from the T0 nodule shown in the 1st column by the GP-WGAN model.(Left): the benign nodules in T0, T1, and their predicted GP-nodules showed stability in size, attenuation, and smooth margins.(Right): the malignant nodules showed a trend of enlarged sizes in T1 and GP-nodules.

Figure A2 .
Figure A2.The network structure of the discriminator.The numbers on the left of each layer (rectangle) represent the width and height of the tensors, and the numbers on the top of the rectangles represent the number of convolution filters (channels) for the tensors.
) (described in Section Deep Residual Neural Network (ResNet-18) for Learned Perceptual Loss (L LP )) to extract deep radiomics features for characterizing lung nodules.The L LP was calcu-de-escalation by the new model in comparison with the baseline models (Lung-RADS and the Brock model).

Table 1 .
NLST documented demographic characteristics, disease staging, and other radiologic factors.

Table 2 .
Test results achieved by the LCRP model and Brock model using virtual GP-nodules or real nodules for the classification of solid nodules, spiculated nodules, and nodules with sizes ranging from 6 to 14 mm in the independent test set.