Next Article in Journal
BcDKM: Blockchain-Based Dynamic Key Management Scheme for Crowd Sensing in Vehicular Sensor Networks
Previous Article in Journal
Target Recognition Based on Millimeter-Wave-Sensed Point Cloud Using PointNet++ Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar-Based Road Surface Classification Using Range-Fast Fourier Transform Learning Models

Convergence Reasearch Center for Disaster & Safety, Advanced Institute of Convergence Technology (AICT), Suwon 16229, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5697; https://doi.org/10.3390/s25185697
Submission received: 11 August 2025 / Revised: 9 September 2025 / Accepted: 10 September 2025 / Published: 12 September 2025
(This article belongs to the Section Radar Sensors)

Abstract

Traffic accidents caused by black ice have become a serious public safety concern due to their high fatality rates and the limitations of conventional detection systems under low visibility. Millimeter-wave (mmWave) radar, capable of operating reliably in adverse weather and lighting conditions, offers a promising alternative for road surface monitoring. In this study, six representative road surface conditions—dry, wet, thin-ice, ice, snow, and sludge—were experimentally implemented on asphalt and concrete specimens using a temperature and humidity-controlled chamber. mmWave radar data were repeatedly collected to analyze the temporal variations in reflected signals. The acquired signals were transformed into range-based spectra using Range-Fast Fourier Transform (Range-FFT) and converted into statistical features and graphical representations. These features were used to train and evaluate classification models, including Extreme Gradient Boost (XGBoost), Light Gradient-Boosting Machine (LightGBM), Convolutional Neural Networks (CNN), and Vision Transformer (ViT). While machine learning models performed well under dry and wet conditions, their accuracy declined in hazardous states. Both CNN and ViT demonstrated superior performance across all conditions, with CNN showing consistent stability and ViT exhibiting competitive accuracy with enhanced global pattern-recognition capabilities. Comprehensive robustness evaluation under various noise and blur conditions revealed distinct characteristics of each model architecture. This study demonstrates the feasibility of mmWave radar for reliable road surface condition recognition and suggests potential for improvement through multimodal sensor fusion and time-series analysis.

1. Introduction

In recent years, traffic accidents caused by black ice during winter have emerged as a serious societal concern. In South Korea, a total of 3944 black ice-related accidents occurred between 2019 and 2023, resulting in 95 fatalities—a fatality rate of 2.4%, approximately 1.7 times higher than that of general traffic accidents [1]. A similar issue has also been reported in the United States. According to the Federal Highway Administration (FHWA), around 24% of weather-related vehicle crashes occur on snowy, slushy, or icy pavement annually, leading to over 1300 deaths and approximately 116,800 injuries. During active snowfall or sleet, approximately 900 fatalities and 76,000 injuries are recorded each year [2]. This high level of risk is primarily due to the visual indistinguishability of black ice and the challenge it poses for timely driver recognition.
Such accidents tend to concentrate during early morning hours in winter, when low temperatures and reduced visibility significantly impair the performance of conventional sensor-based detection systems. Optical sensors suffer drastic performance degradation in nighttime and adverse weather conditions, while infrared and thermal cameras can only measure surface temperature, making it difficult to distinguish between actual freezing and simple cooling. Acoustic sensors also face challenges due to small reflection differences and high sensitivity to environmental noise and other interferences [3,4,5,6].
These limitations have led to increasing demand for alternative technologies capable of reliably monitoring road surface conditions regardless of lighting or weather [7,8,9,10]. Research utilizing millimeter-wave (mmWave) radar has gained momentum [11], and recent systems have shown improved accuracy in recognizing diverse materials and surface conditions [12,13,14]. By operating at high frequency bands, mmWave radar can provide robust reflected signals even in snow, fog, or low-light environments, allowing for quantitative assessment of surface conditions and making it a promising alternative for ice detection [15,16,17].
Recent studies have used polarized mmWave radar operating in the 87.5–90.5 GHz range to classify surface conditions (dry, wet, icy), while other approaches combined 24 GHz mmWave radar with statistical features and machine learning to classify 12 typical road surface types and conditions (e.g., asphalt, gravel, wet, icy) [18,19,20]. Some researchers developed models based on meteorological statistics and imagery to identify snowy and icy surfaces [21]. Other studies applied Range-FFT analysis to train models that classify road objects or enhance detection speed and accuracy by integrating multiple deep learning techniques [22]. Techniques to differentiate between wet and dry roads using reflectivity strength or threshold-based freezing detection have also been proposed using 24–79 GHz radar [23,24,25,26,27]. Additionally, the use of range-Doppler maps has expanded detection capabilities beyond basic range and velocity, enabling classification of moving objects versus static surfaces and improving classification accuracy through deep learning [28], and reliably detecting moving objects under various weather conditions [29,30,31,32].
More recently, multimodal approaches have emerged as promising alternatives for road surface classification. Multimodal transformer models that integrate time-series data from multiple sensors have demonstrated enhanced feature integration capabilities and improved classification accuracy [33]. These approaches leverage the complementary strengths of different sensor modalities, such as radar, LiDAR, and optical sensors, to achieve more robust and accurate road surface recognition. However, such multimodal systems often require complex sensor fusion algorithms and may not be suitable for cost-sensitive applications where single-sensor solutions are preferred. Despite growing efforts to analyze reflection characteristics across various materials, movement-based experiments remain limited due to practical constraints [34].
While previous research has proposed road surface-recognition methods based on mmWave radar and Range-FFT analysis, most experiments have been confined to simplified laboratory conditions and basic surface states (e.g., dry, wet, ice). Although these approaches are valuable for technical validation, few studies have systematically evaluated the temporal behavior of reflected signals under dynamically changing surface conditions. In particular, there is a lack of research that investigates time-series changes in Range-FFT signals across dry, wet, thin-ice, ice, snow, and sludge conditions through repeated experiments to analyze reproducibility and classification performance. Previous studies were confined to simplified laboratory conditions and basic surface states (dry, wet, ice), lacking comprehensive evaluation of hazardous conditions such as thin-ice, snow, and sludge. Most existing research failed to systematically evaluate the temporal behavior of reflected signals under dynamically changing surface conditions, particularly the transition phases between different states. Conventional machine learning models showed significant accuracy decline when classifying dangerous surface conditions, limiting their practical applicability for safety-critical applications. Few studies investigated time-series changes in Range-FFT signals across multiple surface conditions through repeated experiments to analyze reproducibility and classification performance.
Despite the growing body of research on mmWave radar-based road surface recognition, several critical limitations remain unaddressed. Most existing studies have been confined to simplified laboratory conditions and basic surface states (e.g., dry, wet, ice), lacking comprehensive evaluation of hazardous conditions such as thin-ice, snow, and sludge that pose the greatest safety risks. Previous research has failed to systematically evaluate the temporal behavior of reflected signals under dynamically changing surface conditions, particularly the transition phases between different states that are crucial for real-time monitoring. Conventional machine learning models have shown significant accuracy decline when classifying dangerous surface conditions, limiting their practical applicability for safety-critical applications. Few studies have investigated time-series changes in Range-FFT signals across multiple surface conditions through repeated experiments to analyze reproducibility and classification performance. Current approaches primarily rely on basic statistical features (mean, max, std, median, mode) without leveraging more sophisticated time-frequency analysis techniques that could improve classification accuracy. While traditional machine learning methods have been applied, recent advances in ensemble learning, stacking models, and deep learning architectures have not been fully explored for road surface-classification tasks [35].
To address these research gaps, this study makes several key contributions to the field of road surface condition monitoring:
  • Comprehensive surface condition analysis: Six representative road surface states (dry, wet, thin-ice, ice, snow, and sludge) were systematically implemented on both asphalt and concrete specimens using a controlled temperature and humidity chamber, enabling thorough analysis of hazardous conditions.
  • Temporal signal behavior investigation: Radar data were collected at fixed time intervals to quantitatively observe transitions in surface conditions, providing insights into the dynamic behavior of reflected signals during state changes.
  • Dual-approach feature extraction: The collected data were processed using Range-FFT to generate range-based spectra, from which both statistical features and image representations were extracted, enabling comprehensive classification strategies.
  • Comprehensive model evaluation: These features were applied to conventional machine learning algorithms (XGBoost, LightGBM, Random Forest, SVM), CNN-based deep learning models, and Vision Transformer (ViT) to comprehensively evaluate classification performance across different surface conditions and material types.
  • Advanced deep learning comparison: The study provides detailed performance analysis and comparison between CNN and ViT models, demonstrating their distinct characteristics and capabilities in road surface condition classification.
  • Robustness evaluation framework: Comprehensive robustness testing under various noise and blur conditions was conducted for both CNN and ViT models, providing insights into their degradation patterns and practical applicability in challenging environments.
The rest of this paper is organized as follows: Section 2 describes the materials and methods, including mmWave radar background, experimental setup, data processing techniques, and model architectures (CNN and ViT). Section 3 presents the experimental results and performance analysis of different classification models, including comprehensive robustness evaluation under various noise and blur conditions. Section 4 provides a comprehensive discussion of the findings, including detailed comparison between CNN and ViT models, analysis of their distinct characteristics, comparison with existing literature, uncertainty analysis, and limitations. Finally, Section 5 concludes the study and outlines future research directions.
This comprehensive approach enables reliable road surface condition recognition and suggests potential for improvement through multimodal sensor fusion and time-series analysis in real-world applications.

2. Materials and Methods

2.1. Millimeter-Wave Radar Background

Millimeter-wave (mmWave) radar is a high-frequency sensing technology that operates in the 30–300 GHz band. Due to its short wavelength, it offers high spatial resolution, making it suitable for a wide range of applications such as communication networks [36,37], autonomous vehicles [38,39], human sensing [40], and structural health monitoring [41,42]. The core components of mmWave radar include signal modulation, transmission and reception, and signal processing. Among various modulation schemes, Frequency-Modulated Continuous Wave (FMCW) is most commonly employed.

2.1.1. FMCW (Frequency-Modulated Continuous Wave)

FMCW radar transmits a continuous electromagnetic wave (chirp) whose frequency varies linearly over time and receives its reflection from the target to estimate parameters such as range, velocity, and angle of arrival [43]. As illustrated in Figure 1, by measuring the frequency difference—known as the beat frequency—between the transmitted and received signals, accurate distance estimation becomes feasible. The FMCW method offers advantages such as compactness and low power consumption, which make it highly suitable for real-time sensing systems like the one used in this study [44].
The transmitted signal of the FMCW radar can be expressed as [45]:
s tx ( t ) = A cos 2 π f 0 t + K 2 t 2
where f 0 is the starting frequency, K is the chirp rate (frequency slope), and A is the signal amplitude.
The received signal reflected from the target with a time delay τ = 2 R c due to the round-trip distance R is given by [45]:
s rx ( t ) = A cos 2 π f 0 ( t τ ) + K 2 ( t τ ) 2
By mixing the transmitted and received signals, the intermediate frequency (beat signal) is obtained as [45]:
s beat ( t ) = cos 2 π f b t + ϕ
Since the beat frequency f b is linearly proportional to the target range R, the distance can be estimated using [45]:
R = c · f b 2 K

2.1.2. Raw Data Processing

The raw data collected from the FMCW radar consists of complex-valued samples arranged in a multidimensional array composed of frames, transmit antennas, receive antennas, chirps, and ADC samples. Although initially stored as a one-dimensional array, the data is reshaped during preprocessing into a five-dimensional structure, as shown in Equation (5) [46]:
RawData C N f × N chirp × N tx × N rx × N sample
where:
  • N f : number of frames,
  • N chirp : number of chirps per frame,
  • N tx : number of transmit antennas,
  • N rx : number of receive antennas,
  • N sample : number of ADC samples per chirp.
To extract the power spectrum as a function of distance, a one-dimensional Fast Fourier Transform (FFT) is applied along the ADC sample axis for each chirp, as defined in Equation (6):
S range [ k ] = n = 0 N sample 1 x [ n ] · e j 2 π k n N sample , k = 0 , 1 , , N sample 1
This transformation converts the time-domain radar signal into the range domain, allowing the extraction of the power spectrum at each distance bin [46].

2.2. Learning Background

In this study, statistical features extracted from the Range-FFT signal were used as input variables for various machine learning algorithms to classify road surface conditions. The models evaluated include Support Vector Machine (SVM) [47], Random Forest [48], eXtreme Gradient Boosting (XGBoost) [49], and Light Gradient Boosting Machine (LightGBM) [50]. Table 1 summarizes the key characteristics of each algorithm.
The input features were defined in two groups.
(1) Range-FFT statistics. From the magnitude of the Range-FFT within each predefined range interval, we computed five descriptive statistics: mean (arithmetic average of magnitude), maximum (largest magnitude), median (50th percentile of the magnitude distribution), mode (most frequent magnitude value), and standard deviation (dispersion of magnitudes).
(2) Spectral features derived from least-squares spectral analysis (LSSA). LSSA is a parametric spectral estimation method that provides high-resolution frequency analysis by minimizing the least-squares error [51]. In this study, we applied short-time LSSA (ST-LSSA) to the slow-time sequence of each distance bin to characterize time-varying Doppler signatures and improve robustness against noise. Using a sliding-window approach with a fixed window length of 512 samples (≈0.75 s) and 75% overlap, local time–frequency spectra were obtained. From these spectra we derived eleven features:mean spectral power, maximum spectral power, spectral standard deviation, peak frequency, peak amplitude, critical value (significance threshold), normalized residual, number of significant peaks, total number of peaks, average peak frequency, and average peak amplitude. For each distance bin, features were computed for every ST-LSSA window and then averaged across windows to yield one feature vector per bin for machine learning.
In parallel, Range-FFT maps were converted into two-dimensional images and fed to a convolutional neural network (CNN), and we compared the classification performance of the statistical feature–based models against the image-based CNN. Based on these evaluations, the most effective classification algorithm was identified for the given task.

2.2.1. Random Forest (RF)

Random Forest (RF) is an ensemble learning-based classification method conceptualized by Ho in 1995 and systematized by Breiman in 2001 [48]. This approach combines the bagging strategy (Bootstrap Aggregating) with random feature selection to train multiple decision trees independently and aggregates their predictions through majority voting, thereby achieving both high predictive accuracy and robustness [52,53].
Each decision tree in RF is trained using a bootstrap sample of the training data. During node splitting, a random subset of features is selected rather than considering the full feature space, and the best split is determined within this subset. This introduction of randomness reduces correlation between trees and decreases variance in predictions, ultimately enhancing model generalization.
The final prediction y ^ of the RF model is determined by majority voting (for classification) or averaging (for regression) over B independently trained trees, as defined in Equation (7):
y ^ = majority _ vote { f 1 ( x ) , f 2 ( x ) , , f B ( x ) }
where f b ( x ) is the prediction of the b-th decision tree.
RF also provides an intrinsic out-of-bag (OOB) error estimation mechanism by using samples not included in each bootstrap iteration. Furthermore, RF supports internal feature importance evaluation via Variable Importance Measures (VIM), calculated based on either Mean Decrease in Gini impurity (MDG) or Mean Decrease in Accuracy (MDA), enabling the quantification of each feature’s contribution to model prediction [48].

2.2.2. Support Vector Machine (SVM)

Support Vector Machine (SVM), proposed by Cortes and Vapnik in 1995 [47], is a supervised learning algorithm for classification and regression tasks. It aims to find the optimal separating hyperplane that maximizes the margin between classes in the feature space. SVM is particularly effective for small, high-dimensional, and non-linear datasets and has demonstrated high accuracy across diverse applications.
For linearly separable datasets, SVM seeks to find the hyperplane defined by:
w x + b = 0
where w is the normal vector to the hyperplane and b is the bias term. The optimization objective is to maximize the margin between the support vectors and the decision boundary, formulated as the following convex quadratic programming problem:
min w , b 1 2 w 2 subject   to y i ( w x i + b ) 1 , i
To handle non-linearly separable data, SVM applies kernel functions (e.g., RBF, polynomial) to map the original data into a higher-dimensional feature space. The model performance is highly dependent on the tuning of the regularization parameter C and kernel coefficient γ . One of SVM’s major strengths is its theoretical guarantee to reach a global optimum, unlike models that rely on local search heuristics. However, due to its binary classification structure, SVM may become inefficient or complex when extended to multi-class problems [52,53].

2.2.3. XGBoost

XGBoost (eXtreme Gradient Boosting), developed by Tianqi Chen and Carlos Guestrin in 2014 [49], is a high-performance ensemble learning algorithm based on the gradient boosting framework. It enhances traditional Gradient Boosted Decision Trees (GBDT) by introducing significant improvements in computational efficiency, regularization, and scalability. XGBoost has been widely adopted across various domains for its robustness to overfitting and reliable performance on high-dimensional and imbalanced datasets [54,55].
XGBoost iteratively builds an ensemble of weak learners by minimizing a regularized objective function that incorporates both first- and second-order derivatives (i.e., gradient and Hessian) of the loss function. This allows for more stable and efficient optimization. The overall objective at the t-th iteration is defined as:
L ( t ) = i = 1 n l y i , y ^ i ( t 1 ) + f t ( x i ) + Ω ( f t )
where f t is the new tree added at iteration t, and  Ω ( f t ) is the regularization term controlling the complexity of the tree, expressed as:
Ω ( f t ) = γ T + 1 2 λ j = 1 T ω j 2
Here, T is the number of leaf nodes and ω j is the output score of the j-th leaf. The regularization terms γ and λ help penalize complex trees and prevent overfitting. Additional techniques such as column sampling, early stopping, and tree pruning further enhance training efficiency and model generalization.

2.2.4. Light Gradient Boosting Machine (LightGBM)

Light Gradient Boosting Machine (LightGBM) is a fast and high-performance gradient boosting algorithm designed to enhance the efficiency of traditional Gradient Boosted Decision Trees (GBDT) [50]. While it shares structural similarities with GBDT and XGBoost, LightGBM is more optimized in terms of training speed, memory efficiency, and predictive accuracy. It is particularly effective when handling large-scale datasets with high-dimensional feature spaces.
The prediction function of LightGBM can be represented as a linear combination of base learners, as defined in Equation (12):
F ( x ) = m = 1 M δ m f m ( x )
where f m ( x ) denotes the m-th base classifier (tree), and  δ m is the learned weight coefficient (typically adjusted via gradient information).
To enhance both computational efficiency and model performance, LightGBM incorporates two key techniques: Gradient-based One-Side Sampling (GOSS) and Exclusive Feature Bundling (EFB). GOSS selects training instances with large gradient magnitudes, prioritizing the most informative samples. EFB reduces feature dimensionality by grouping mutually exclusive sparse features, thereby accelerating training and reducing memory consumption.
Moreover, LightGBM adopts a leaf-wise tree growth strategy that expands the leaf node with the highest loss reduction. This method enables more precise splits compared to level-wise approaches and achieves better accuracy with deeper trees under the same complexity constraint [55].

2.2.5. Convolutional Neural Network (CNN)

Convolutional Neural Networks (CNNs) are deep learning architectures introduced by LeCun et al. in 1998 [56]. CNNs have demonstrated exceptional performance in computer vision tasks such as image recognition, classification, and object detection. Unlike fully connected networks, CNNs automatically extract spatially correlated features from input images while using significantly fewer parameters, thus enabling efficient training.
A typical CNN architecture consists of multiple layers, including Convolution layers, Rectified Linear Unit (ReLU) activation layers, Pooling layers, and Fully Connected (FC) layers. The Convolution layer generates a feature map by applying a learnable filter to the input, as formulated in Equation (13):
Z i , j = ( X W ) i , j + b
where X is the input image, W is the convolutional kernel (filter), b is the bias term, and  Z i , j is the output feature value at position ( i , j ) .
CNNs exploit spatial locality and weight sharing to reduce the number of parameters, enabling translation-invariant feature extraction. These properties allow CNNs to learn hierarchical and highly discriminative representations directly from raw data, without the need for handcrafted features. Consequently, CNNs have become a dominant model in visual pattern-recognition and -classification tasks [57,58].

2.2.6. Vision Transformer

Vision Transformer (ViT) has emerged as a revolutionary approach in computer vision by leveraging the transformer architecture, originally developed for natural language processing, to address image-classification challenges [59]. In contrast to convolutional neural networks (CNNs) that utilize local receptive fields, ViT processes images as sequences of patches, thereby enabling global attention mechanisms that effectively capture long-range dependencies across the entire image domain.
For our implementation, we utilize the pre-trained google/vit-base-patch16-224 model, which has exhibited outstanding performance across diverse vision applications. The architectural framework comprises three fundamental components:
The patch embedding layer processes an input image I R H × W × C by partitioning it into N = ( H / P ) × ( W / P ) non-overlapping patches of dimensions P × P , where P = 16 . Each patch undergoes flattening and subsequent linear projection into a D-dimensional embedding space:
x p i = E · Flatten ( P i ) + e p o s i
where E R D × P 2 C represents the learnable patch embedding matrix and e p o s i denotes the positional encoding for patch i.
The transformer encoder constitutes the core architecture with L transformer encoder layers, each incorporating multi-head self-attention (MSA) and multi-layer perceptron (MLP) blocks with layer normalization (LN):
z l = MSA ( LN ( z l 1 ) ) + z l 1
z l = MLP ( LN ( z l ) ) + z l
The multi-head self-attention mechanism computes attention weights through the following formulation:
Attention ( Q , K , V ) = softmax Q K T d k V
where Q, K, and V represent query, key, and value matrices, respectively, and  d k corresponds to the dimension of the key vectors.
The classification head in the final layer incorporates a learnable classification token [CLS] that aggregates information from all patches via self-attention mechanisms, followed by a linear classification layer:
y = MLPHead ( LN ( z L 0 ) )
where z L 0 represents the final state of the classification token.
The model configuration employs the following hyperparameters: patch size P = 16 , embedding dimension D = 768 , number of transformer layers L = 12 , number of attention heads h = 12 , and MLP dimension = 3072 . The model generates 6 class probability outputs corresponding to road surface conditions: dry, wet, thin-ice, ice, snow, and sludge.

2.3. Experimental Setup

2.3.1. Sensor and System

The mmWave radar-based detection system employed in this study consists of the IWR1843BOOST module and the DCA1000EVM data capture board from Texas Instruments, along with a connected personal computer (PC). The IWR1843BOOST operates within a frequency range of 76–81 GHz and supports a maximum bandwidth of 4 GHz. It is an integrated sensor incorporating three transmit (Tx) antennas and four receive (Rx) antennas, enabling high-speed signal processing through its built-in MCU and DSP. In addition, it is equipped with 2 MB of onboard memory and supports multiple communication interfaces, including UART, SPI, and CAN, allowing for high-resolution measurements of range, velocity, and angle. The DCA1000EVM functions to capture radar-acquired data in real time and transmit it to the PC for storage and processing. The system configuration is illustrated in Figure 2 [60].

2.3.2. Experimental Environment

This study conducted experiments to classify road surface conditions using millimeter-wave (mmWave) radar data. As described in Section 2.3.1, the sensor system comprised the IWR1843BOOST module and the DCA1000EVM data capture board, which were installed at the upper section inside a temperature- and humidity-controlled chamber, as illustrated in Figure 3. The sensor was positioned at a height of approximately 0.64 m within the chamber, and reflected signals were measured at a horizontal distance of r = 0.17  m from the target surface. The test specimens consisted of square blocks (15 cm × 15 cm) made of concrete and asphalt. To minimize the influence of external environmental factors, each specimen was fixed using insulation material and molds measuring 20 cm × 20 cm. Data were collected for six distinct surface conditions: dry, wet, thin-ice, ice, snow, and sludge, with the definitions provided in Figure 4. These conditions were reproduced within the temperature–humidity chamber by controlling the temperature. Notably, the thin-ice and sludge conditions were defined as two distinct types of black ice. For each condition, raw mmWave radar data were acquired in PCAP format via the DCA1000EVM. To enhance data diversity and generalization, the specimens were rotated to capture measurements from multiple angles under the same condition. Subsequently, the Range-FFT was applied to extract the power spectrum in the frequency domain, and the signal strength at the fixed distance of r = 0.17 m was analyzed.

2.4. Data Acquisition

In this study, raw signal data corresponding to various road surface conditions were collected using Texas Instruments’ IWR1843BOOST mmWave radar module in conjunction with the DCA1000EVM data capture board (Texas Instruments Incorporated, Dallas, TX, USA). Instead of employing TI’s mmWave Studio, real-time data acquisition was conducted via a Python-based custom program (Python 3.10) that directly controlled the UART and LVDS interfaces. The radar configuration was specified through a .cfg file, which contained the parameters of the FMCW signal and the frame structure.
A multiple-input multiple-output (MIMO) array comprising 12 virtual channels was constructed using three transmit (Tx) antennas and four receive (Rx) antennas. The range resolution was approximately 3.75 cm, achieved with a bandwidth of about 4.0 GHz, 928 samples per chirp, a sampling rate of 7404 ksps, a chirp duration of 30 μ s, and a slope setting of 133.33 MHz/ μ s. Each frame consisted of 16 chirps, and a total of 10 frames were recorded.
The acquired data were stored in PCAP format as complex 16-bit values and transmitted from the DCA1000EVM board to a connected PC via Ethernet. Each sample was represented as a 4-byte complex number comprising real and imaginary components, which were subsequently used for Range-FFT processing and the analysis of various surface conditions (dry, wet, thin-ice, ice, snow, sludge).
Data acquisition was performed over multiple repeated experiments, with each experiment conducted separately on asphalt and concrete specimens. A total of 1000 data samples were collected per surface condition (dry, wet, thin-ice, ice, snow, sludge) for each specimen type, resulting in 6000 data samples per specimen.

2.5. Data Processing and Feature Extraction

For the reshaped signals, a one-dimensional Fast Fourier Transform (1D-FFT) was applied along the sample axis for each chirp to generate the Range-FFT results. The analysis was conducted at a range bin corresponding to approximately 0.17 m, and the FFT power value at this distance was used to evaluate variations in reflection characteristics across different road surface conditions (dry, wet, thin-ice, ice, snow, sludge). Each condition was repeatedly measured under identical experimental settings to assess the reproducibility and discriminability of the signal patterns.
From the Range-FFT results, we extracted five summary statistics (mean, maximum, median, mode, and standard deviation) and, as detailed above, incorporated the ST-LSSA-derived time-frequency features obtained through sliding-window least-squares spectral analysis; these were used as input variables for the machine-learning classifiers. Simultaneously, the Range-FFT output was converted into power spectrum images to serve as input for a CNN-based deep learning model, enabling a comparative analysis of classification performance between the two approaches.
This data processing and feature extraction pipeline was designed to generate a training dataset based on repeated experimental measurements, and the overall workflow is illustrated in Figure 5.

2.6. Classification and Analysis

In this study, both traditional machine learning models and a deep learning-based CNN model were comparatively evaluated for the classification of road surface conditions (dry, wet, thin-ice, ice, snow, sludge). The training data were constructed from the Range-FFT-based statistical features and the power spectrum images described in Section 2.5.
We employed a 5-fold cross-validation: the dataset was split into five equal folds [61]. in each iteration, one fold served as the validation set (20%) and the remaining four folds as the training set (80%). Across the five iterations, every sample appeared once in validation and four times in training, increasing data diversity and yielding a more reliable estimate of generalization performance. In addition, experiments were conducted over three candidate range intervals (0.10–0.24 m, 0.07–0.27 m, and 0.04–0.30 m).
For each specimen type (asphalt and concrete), separate model training was performed for each of the three range intervals. This approach enabled the identification of the optimal range interval that yielded the highest classification accuracy. Based on the maximum accuracy within the selected range, the optimal classification model was determined for each specimen, and further analysis was conducted to evaluate which algorithm achieved the best performance for each surface condition. Model performance was assessed using confusion matrices and accuracy as the primary evaluation metrics.
For the statistical feature–based classification, we extracted five Range-FFT statistics—mean, maximum, median, mode, and standard deviation—for each predefined range interval; for the same intervals, we also computed the previously described ST-LSSA-derived time–frequency features, and used features as input variables. The machine learning models evaluated included Random Forest (RF), Support Vector Machine (SVM), XGBoost, and LightGBM. The parameters used for training these machine learning models are listed in Table 2.
For the deep learning-based classification, the Range-FFT power values were converted into image format and used as input to a CNN-based classifier. The CNN model was implemented using PyTorch (version 2.8.0+cu128). For data preprocessing and augmentation, the input radar images were resized to 224 × 224 RGB format [62], and various techniques were applied, including Gaussian noise (SNR 20–30 dB for normal, 10–20 dB for extreme) and motion blur (kernel 3–7 for normal, 7–15 for extreme), followed by conversion to tensors and normalization using ImageNet mean and standard deviation values. During training, the Early Stopping technique [63] and a weight decay regularization method [64] were employed to prevent overfitting. The CNN architecture was based on a pre-trained ResNet18 model, with the final fully connected layer replaced by a Softmax output layer to classify the six road surface condition classes. The parameters used for CNN model training are presented in Table 3.
For the Vision Transformer (ViT) model, we employed the pre-trained google/vit-base-patch16-224 architecture, which processes 224 × 224 radar-generated images through a transformer-based architecture with 12 layers and 12 attention heads. The ViT model was implemented using the Hugging Face Transformers library and PyTorch (version 2.8.0). Similar to the CNN approach, the input radar images were resized to 224 × 224 format and underwent radar-realistic augmentation including Gaussian noise (SNR 20–30 dB for normal, 10–20 dB for extreme) and motion blur (kernel 3–7 for normal, 7–15 for extreme), followed by tensor conversion and ImageNet normalization. During training, we utilized AdamW optimizer with a learning rate of 2 × 10 5 , weight decay of 0.01, and implemented early stopping with patience of 15 epochs to prevent overfitting. The model architecture consists of patch embedding layers, transformer encoder blocks with self-attention mechanisms, and a classification head for the six road surface condition classes. The parameters used for ViT model training are presented in Table 4.

3. Results

3.1. Example of Data

In this study, both machine learning- and deep learning-based approaches were employed for road surface condition classification. Accordingly, different types of input data were utilized, and this section presents representative examples of the datasets used in each methodology.
As noted above, we constructed two feature groups for the statistical feature–based learning. The first consists of five Range-FFT statistics (mean, maximum, standard deviation, mode, and median). The second comprises eleven ST-LSSA–derived time–frequency features (mean spectrum, maximum spectrum, standard deviation of the spectrum, peak frequency, peak amplitude, critical value [significance threshold], normalized residual, number of significant peaks, total number of peaks, average peak frequency, and average peak amplitude). Using a Random Forest classifier for its training/inference efficiency, we performed a 5-fold cross-validated ablation and sensitivity analysis across three feature sets (FFT only, FFT+ST-LSSA, ST-LSSA only) (Table 5).
Across all range intervals and both specimens, the Range-FFT–only set yielded the highest validation accuracy in every case except one (FFT: 0.875, FFT+ST-LSSA: 0.880). Under the present controlled laboratory conditions, where the signals remain relatively stationary, the addition of ST-LSSA features—or the use of ST-LSSA alone—in most cases resulted in lower accuracy compared with Range-FFT statistics. This finding suggests that in stable environments, low-dimensional Range-FFT statistics are sufficient to characterize the data. Nevertheless, since ST-LSSA is specifically designed to capture nonstationary and time-varying Doppler content, its utility may become more evident in field measurements under dynamic road and environmental conditions. From a feature-efficiency standpoint, the best accuracy was achieved with just five low-dimensional features; increasing dimensionality to 16 provided no benefit and even led to a slight degradation. Accordingly, the final model inputs are restricted to the five Range-FFT statistics. An example of the training inputs is shown in Table 6, and the full dataset is provided in Table S1.
Figure 6 presents the power spectrum image data used for CNN-based classification under the same surface conditions. These images are visualizations of the 2D Range-FFT results and were directly employed as input data for training the CNN model in this study.

3.2. Classification Performance

To determine the optimal range interval, classification experiments were conducted across multiple candidate ranges (0.10–0.24 m, 0.07–0.27 m, and 0.04–0.30 m) using validation accuracy as the evaluation criterion. Both traditional machine learning models (SVM, Random Forest, LightGBM, and XGBoost) and a deep learning-based CNN model were applied, and comparative analyses were performed separately for asphalt and concrete specimens.
Table 7 summarizes the results of the accuracy of the validation for each range interval and the specimen material, providing a clear comparison of the performance of the model. By averaging the accuracy of the asphalt and concrete specimens, the classification performance was compared across different distance intervals. The machine learning (ML) models (RF, SVM, XGBoost, LightGBM) did not show a clear trend with respect to the interval configuration, maintaining similar performance in the range of approximately 79–82% (81.69% for 0.10–0.24 m, 79.67% for 0.07–0.27 m, and 80.72% for 0.04–0.30 m). In contrast, the CNN demonstrated consistent improvement as the interval was extended, achieving accuracies of 98.05% (0.10–0.24 m), 98.75% (0.07–0.27 m), and 99.00% (0.04–0.30 m). This indicates that CNNs gain a distinct advantage when operating over a wider interval, and based on this observation, the 0.04–0.30 m range was adopted as the optimal configuration in this study. This selection is meaningful because it captures both near- and far-range reflections, effectively representing the overall distribution characteristics of the signal, and maximizes the benefits of deep learning–based analysis by achieving the highest accuracy (above 99%) in CNN classification. Consequently, this interval was applied as the standard input configuration for all subsequent analyzes.
Accordingly, this study selected 0.04–0.30 m as the optimal range interval, and the subsequent sections present an in-depth analysis of detailed evaluation metrics including confusion matrices, precision, recall, and F1-score based on this interval.

3.2.1. Asphalt Results

The results presented in Figure 7 summarize the validation performance of each model for the 0.04–0.30 m range interval, based on the data collected from asphalt specimens.
The confusion matrix on the left illustrates the prediction distribution across classes, while the heatmap on the right visualizes the precision, recall, and F1-score, providing a detailed representation of the classification performance for each model.
Overall, the CNN model exhibited the best performance among all evaluated models. It achieved a accuracy of 99.25% across all classes, with precision and recall exceeding 0.97 for every class, indicating highly stable classification capability. Notably, the F1-score for all classes was above 0.98, reflecting consistent classification results.
Among the machine learning-based models, XGBoost demonstrated the highest performance, attaining a validation accuracy of 91.13%. The confusion matrices indicate fewer off-diagonal errors for difficult classes (thin-ice, ice, snow, sludge), yielding a stronger precision–recall balance.
In contrast, SVM recorded the lowest overall accuracy of 60.11% and exhibited poor performance in all classes except wet. In particular, for the ice class, SVM showed extremely low performance, with a recall of 0.055, and F1-score of 0.103.
The class-wise F1-score analysis is as follows:
  • Dry: CNN (1.000), LightGBM (0.980), Random Forest (0.975), and XGBoost (0.975) demonstrated excellent performance;
  • Wet: Most models demonstrated excellent performance, with CNN and LightGBM achieving the highest F1-score of 0.982 and 0.961, respectively;
  • Thin-Ice: CNN and XGBoost achieved the highest F1-scores of 0.980 and 0.912, respectively;
  • Ice: CNN and LightGBM achieved the highest F1-scores of 0.985 and 0.884, respectively;
  • Snow: CNN and XGBoost achieved the highest F1-scores of 0.988 and 0.916, respectively;
  • Sludge: CNN (0.995), Random Forest (0.836), LightGBM (0.854) and XGBoost (0.847) were recorded.
Based on this analysis, the CNN model was selected as the representative classification algorithm for the asphalt specimens, as it demonstrated the highest classification performance.

3.2.2. Concrete Results

For the concrete specimens, classification performance within the 0.04–0.30 m range interval was evaluated under the same experimental conditions and data processing procedures as those used for the asphalt specimens. Figure 8 presents the validation confusion matrices and classification performance metrics (precision, recall, and F1-score) for each model.
Overall, the CNN model demonstrated the highest performance. It achieved a validation accuracy of 98.75% across all classes, with both precision and recall exceeding 0.96 for each class. Notably, for the dry, wet, ice, snow and sludge classes, the F1-score exceeded 0.98, indicating highly stable classification performance.
Among the machine learning-based models, LightGBM exhibited the best results, recording a validation accuracy of 89.24%. It delivered top or near-top accuracy with the most stable precision–recall profile across classes.
In contrast, SVM recorded the lowest overall accuracy of 54.30% and showed unstable classification performance for all classes except wet.
The class-wise F1-score analysis is as follows:
  • Dry: CNN (0.995), LightGBM (0.959), Random Forest (0.958), XGBoost (0.939);
  • Wet: All models demonstrated excellent performance, with CNN, LightGBM, Random Forest and XGBoost each achieving an F1-score of at least 0.97;
  • Thin-Ice: Overall performance was lower; however, CNN (0.967) maintained stable results;
  • Ice: CNN (0.980) and LightGBM (0.879) achieved the highest F1-scores;
  • Snow: CNN (0.985), XGBoost (0.904), and LightGBM (0.903) demonstrated strong performance;
  • Sludge: CNN (0.995), LightGBM (0.791), Random Forest (0.832), and XGBoost (0.798).
Accordingly, the CNN model was selected as the representative classification model for the concrete specimens, as it demonstrated the highest performance.

3.2.3. Vision Transformer Performance Analysis

We first compared traditional machine-learning models that use statistical features as inputs (RF, SVM, XGBoost, LightGBM) with an image-based CNN under identical conditions and confirmed that the CNN achieved the highest classification performance. To enable a fair comparison within image-based approaches, we then trained and evaluated a Vision Transformer (ViT) model using the same data and protocol. Both models took as input 224 × 224 graph images obtained by transforming the radar signals via range-FFT, and we applied identical preprocessing, normalization, and realistic augmentations. We adopted stratified 5-fold cross-validation for training and validation in each fold, fixing random seeds and fold assignments to ensure reproducibility. Training settings—optimizer, learning-rate schedule, and early-stopping criterion—were aligned to ensure fairness. Robustness was evaluated by injecting Gaussian noise (SNR = 20/25/30 dB for normal, 10/15/20 dB for extreme) and motion blur (k = 3/5/7 for normal, 7/11/15 for extreme) at test time.
For asphalt specimens, the ViT model demonstrated robust performance with strong generalization capabilities across different experimental folds. The 5-fold cross-validation analysis revealed that the model achieved consistent high accuracy, with the best performing fold (Fold 2) reaching 99.75% baseline accuracy. Notably, this fold demonstrated exceptional robustness under normal degradation conditions, with minimal performance degradation of only 5.64% under normal motion blur conditions, slightly exceeding the 5% robustness threshold.
The robustness analysis for asphalt specimens showed that Fold 2 achieved the highest baseline accuracy (99.75%) but slightly exceeded the 5% robustness threshold under normal blur conditions (5.64% degradation). Under normal noise conditions, this model showed negligible performance degradation (0.03%), indicating enhanced stability compared to baseline performance. The model maintained its high classification accuracy while demonstrating resilience against common environmental perturbations.
Detailed robustness analysis for the best performing asphalt model (Fold 2) revealed: Normal noise conditions (SNR 20 dB, 25 dB, 30 dB) showed average performance degradation of 0.03%, while normal blur conditions (Blur 3, 5, 7) resulted in 5.64% performance degradation. Under extreme conditions, the model achieved 92.44% accuracy at SNR 10 dB, 99.11% at SNR 15 dB, and 99.81% at SNR 20 dB. For extreme blur conditions, performance was 87.17% at Blur 7, 91.10% at Blur 11, and 87.17% at Blur 15.
The class-wise performance analysis for the best performing asphalt model (Fold 2) showed:
  • Dry: Precision 1.000, Recall 1.000, F1-score 1.000
  • Wet: Precision 0.995, Recall 1.000, F1-score 0.998
  • Thin-Ice: Precision 0.995, Recall 0.995, F1-score 0.995
  • Ice: Precision 1.000, Recall 1.000, F1-score 1.000
  • Snow: Precision 1.000, Recall 1.000, F1-score 1.000
  • Sludge: Precision 0.995, Recall 0.990, F1-score 0.993
For concrete specimens, the ViT model exhibited strong baseline performance but faced challenges in maintaining robustness under degradation conditions. The best performing fold (Fold 1) achieved a baseline accuracy of 98.58%, demonstrating the model’s capability to learn complex surface characteristics of concrete. However, this fold failed to meet the 5% robustness threshold under normal motion blur conditions, with performance degradation of 7.22%.
The robustness analysis revealed that concrete specimens present unique challenges for the ViT architecture, likely due to the more heterogeneous surface characteristics and complex reflection patterns compared to asphalt. Despite the robustness limitations, the model maintained high classification accuracy for most surface conditions, with particularly strong performance in distinguishing between dry and wet conditions.
Detailed robustness analysis for the best performing concrete model (Fold 1) showed: Normal noise conditions (SNR 20 dB, 25 dB, 30 dB) resulted in minimal performance degradation of 0.11%, while normal blur conditions (Blur 3, 5, 7) caused 7.22% performance degradation, exceeding the 5% threshold. Under extreme conditions, the model achieved 88.97% accuracy at SNR 10 dB, 97.25% at SNR 15 dB, and 98.33% at SNR 20 dB. For extreme blur conditions, performance was 85.89% at Blur 7, 87.47% at Blur 11, and 85.89% at Blur 15.
The class-wise performance analysis for the best performing concrete model (Fold 1) showed:
  • Dry: Precision 0.985, Recall 0.985, F1-score 0.985
  • Wet: Precision 0.980, Recall 0.980, F1-score 0.980
  • Thin-Ice: Precision 0.975, Recall 0.970, F1-score 0.972
  • Ice: Precision 0.970, Recall 0.965, F1-score 0.967
  • Snow: Precision 0.965, Recall 0.975, F1-score 0.970
  • Sludge: Precision 0.980, Recall 0.975, F1-score 0.977
Figure 9 presents the confusion matrices for the best performing ViT models on asphalt and concrete specimens, demonstrating the high classification accuracy achieved across all road surface conditions.
We conducted robustness tests under Gaussian noise and motion blur to directly compare the performance of the CNN and ViT models. Table 8 summarizes the accuracies under Gaussian noise (SNR = 20/15 dB) and motion blur ( k = 3 / 5 ), together with the absolute drop relative to the clean condition (percentage points, pp). For the CNN, on asphalt, the decreases were 0.67 pp at SNR 20 dB (99.83% to 99.17%) and 4.08 pp at SNR 15 dB (to 95.75%); for blur, k = 3 yielded 3.92 pp (95.92%) and k = 5 yielded 4.75 pp (95.08%). On concrete, the corresponding figures were 0.17 pp at SNR 20 dB (98.92% to 98.75%), 3.67 pp at SNR 15 dB (to 95.25%), 3.33 pp at k = 3 (95.58%), and 4.75 pp at k = 5 (94.17%). Overall, the CNN’s performance degradation remained within 5 pp under all perturbations, satisfying the robustness criterion. For the ViT model, on asphalt, the performance showed 0.05 pp degradation at SNR 20 dB (99.75% to 99.70%) and 8.60 pp at SNR 15 dB (to 91.15%); for blur, k = 3 yielded 3.55 pp (96.20%) and k = 5 yielded 9.73 pp (90.02%). On concrete, the corresponding figures were 0.11 pp at SNR 20 dB (98.58% to 98.47%), 9.61 pp at SNR 15 dB (to 88.97%), 7.22 pp at k = 3 (91.36%), and 12.69 pp at k = 5 (85.89%). The ViT model demonstrated competitive performance with CNN under normal conditions but showed higher sensitivity to extreme perturbations, particularly under blur conditions.
Figure 10 provides a comprehensive visual comparison of these performance patterns across all conditions.

4. Discussion

In this study, six road surface conditions—dry, wet, thin-ice, ice, snow, and sludge—were reproduced under controlled chamber conditions on two types of specimens (concrete and asphalt), and mmWave radar data were collected. After acquiring the experimental data, the classification performance of various machine learning and deep learning models was systematically compared and analyzed for both asphalt and concrete specimens. Based on validation accuracy, the optimal statistical feature-based models were identified as XGBoost and LightGBM, which were trained and evaluated on asphalt and concrete surface datasets, respectively.
For the six-class classification task, machine learning-based models such as XGBoost and LightGBM demonstrated potential for real-time applications and deployment in resource-constrained environments, owing to their lightweight architectures, fast training speeds, and specialized performance in certain classes. In practice, XGBoost achieved a accuracy of 91.13% for the asphalt specimens, and LightGBM achieved 89.24% for the concrete specimens, both showing stable classification capability. However, frequent misclassification between the ice and sludge classes was observed for asphalt, while the concrete specimens exhibited slightly lower performance overall, with particularly noticeable degradation in the thin-ice and sludge classes. Overall, statistical feature-based models showed relatively lower discriminative power for hazardous classes (thin-ice, ice, snow, sludge).
The deep learning-based CNN model achieved accuracies of 99.25% for asphalt and 98.75% for concrete specimens. Across both asphalt and concrete specimens, precision, recall, and F1-score for the dry and sludge classes were at least 0.99, indicating clearly distinguishable boundaries from other surface conditions. Even for challenging classes such as thin-ice, ice, snow, and sludge, the CNN achieved high F1-scores. Notably, for concrete specimens, where the statistical feature-based models showed particularly low recall for thin-ice and sludge, the CNN achieved a recalls of 0.965 and 1.000 for these classes, respectively, clearly outperforming the machine learning models. This result indicates that the CNN maintains robust classification capability even under complex surface boundaries and ambiguous conditions.
The findings confirm that CNNs, with their ability to effectively learn complex boundaries and high-dimensional feature distributions, are well-suited for the precise classification of hazardous classes (thin-ice, ice, snow). While statistical feature-based models provided sufficient performance for normal condition classification (dry, wet), they faced limitations in distinguishing hazardous classes.
Importantly, the findings of this study demonstrate that both CNN-based and Vision Transformer deep learning models can reliably classify diverse road surface conditions using mmWave radar signals alone, thereby overcoming the limitations of statistical feature-based models. In the robustness evaluation, the CNN achieved superior baseline accuracy (99.83% for asphalt, 98.92% for concrete), while the ViT model demonstrated competitive performance (99.75% for asphalt, 98.58% for concrete) with enhanced robustness under various degradation conditions.
The comparative analysis between CNN and ViT models reveals distinct characteristics and trade-offs. The CNN model exhibited consistent performance across all conditions, with degradation remaining within 5 percentage points under normal perturbations, making it highly suitable for stable, real-world deployment scenarios. In contrast, the ViT model showed competitive baseline performance but demonstrated higher sensitivity to extreme perturbations, particularly under blur conditions where performance degradation exceeded 7 percentage points for concrete specimens. This sensitivity can be attributed to the transformer’s global attention mechanism, which, while effective for capturing complex spatial relationships, may be more susceptible to local distortions and noise in radar signals.
The ViT model’s strength lies in its ability to capture global dependencies across the entire range-FFT spectrum, enabling effective processing of complex radar signatures where spatial relationships are crucial for accurate classification. This global perspective is particularly beneficial for distinguishing subtle differences between road surface conditions that may not be apparent through local feature analysis alone. However, this same global attention mechanism contributes to the model’s increased computational complexity and sensitivity to input perturbations, as evidenced by the robustness test results.
From a practical deployment perspective, the choice between CNN and ViT models depends on the specific application requirements. For scenarios requiring high stability and consistent performance under varying environmental conditions, the CNN model provides a more reliable baseline. Conversely, for applications where capturing complex global patterns is paramount and computational resources are sufficient, the ViT model offers enhanced feature extraction capabilities.
These results are consistent with prior studies that applied CNN, Transformer, and MobileNet approaches for road condition recognition. For example, TC–Radar combined CNN and Transformer for mmWave object detection, achieving an AP of 83.99% and demonstrating the complementary effect of local and global feature extraction [22], while the lightweight Vision Transformer LeViT achieved 99.17% accuracy in pavement image classification and improved computational efficiency compared to ResNet and DenseNet [65]. These prior studies demonstrate the effectiveness of CNN- and Transformer-based models, and our study likewise verifies that CNN and ViT models can achieve over 98% accuracy with a single mmWave radar sensor, ensuring stable performance particularly in high-risk classes such as thin-ice, ice, snow, and sludge.
Nevertheless, several uncertainties and limitations should be acknowledged. First, we did not conduct a systematic exploration of hyperparameters and model sensitivity. Although the training protocol was unified across models, we did not perform extensive sweeps of key hyperparameters or architectural settings. Consequently, the models’ sensitivity has not been fully characterized; rigorous hyperparameter optimization and sensitivity analyses should be pursued in subsequent studies. Second, the experiments were conducted under artificially controlled chamber conditions, which ensured reproducibility and comparability but did not fully capture the diversity and complexity of real-world road environments. Future work should therefore include outdoor experiments and on-road validations. In such dynamic settings, advanced time–frequency analysis methods may provide additional benefits. Although this study employed short-time LSSA (ST-LSSA) to capture nonstationary Doppler content, LSWA is inherently a time–frequency domain method rather than a purely frequency-domain approach, and thus can better estimate signals with variability in both frequency and amplitude over time [51]. As a direction for future work, LSWA could be further explored to enhance feature extraction and improve the recognition of hazardous surface classes under dynamic road and environmental conditions. In addition, optimization-based strategies can further enhance model reliability. Recent research has reported that ensemble predictions and metaheuristic-optimized models substantially improve robustness compared to single unoptimized models [66], suggesting that similar approaches could be extended to road-surface recognition tasks.
Also, the necessity of multimodal sensor fusion is an important consideration. A comprehensive review has highlighted that the fusion of mmWave radar and vision sensors significantly improves the robustness and accuracy of perception in autonomous driving systems [67]. For instance, under adverse conditions such as dense fog, fusion systems that combine camera, LiDAR, FIR (thermal), radar, and environmental sensors have shown significant AP improvements over radar-only approaches [68], while Radar+Camera fusion has demonstrated gains in NDS [69]. Furthermore, early fusion of Radar+Camera+LiDAR has also been reported to increase AP [70]. These trends suggest that extending our single-radar approach to multimodal fusion is well justified.
The present study has focused on a single mmWave radar sensor; however, in parallel, we are developing a multimodal sensing platform that integrates RGB and IR cameras, with plans to further extend the system to thermal cameras. As demonstrated in previous research, such multimodal sensor fusion has been consistently reported to enhance perception performance and robustness under adverse weather and low-light conditions, and these directions will be actively incorporated into our future work.

5. Conclusions

By reproducing six distinct road surface conditions in controlled chamber environments, this work systematically compared the classification performance of machine learning and deep learning models based on mmWave radar data. The results showed that while statistical feature-based models such as XGBoost and LightGBM offered advantages in terms of lightweight architectures and fast inference, their discriminative power was limited for hazardous classes (thin-ice, ice, snow, sludge). In contrast, the CNN model achieved high accuracy and F1-scores across both asphalt and concrete specimens, demonstrating stable performance, while the ViT model exhibited competitive accuracy and highlighted the potential of global pattern learning for complex radar signal interpretation.
These findings confirm the academic significance of CNN architectures in accurately classifying complex and high-risk road surface conditions, while also highlighting the potential of Transformer-based models through their ability to capture global dependencies. In practical terms, CNN models demonstrate applicability through consistent and stable performance, whereas ViT models suggest promising directions for scenarios involving more complex radar signals. Collectively, these results underscore the potential of mmWave radar alone to enable reliable road surface monitoring and traffic safety applications, particularly under conditions where vision-based sensors may be constrained by environmental variability.
Nevertheless, this study has limitations. The experiments were performed in controlled chamber environments, which ensured reproducibility but did not fully reflect the diversity of real-world road conditions. Additionally, the inherent noise in radar data and sensitivity to environmental changes may affect model stability. Future work will include outdoor and on-road validations to confirm generalizability, as well as the stepwise development of a multimodal sensor fusion platform integrating RGB, IR, and thermal cameras to enhance perception performance and robustness under adverse weather and low-light conditions.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25185697/s1, Table S1: Complete dataset of extracted features and labels for asphalt and concrete conditions.

Author Contributions

Conceptualization, H.L. and M.Y.; Methodology, H.L., J.K., K.K. and M.Y.; Software, H.L., J.K. and K.K.; Validation, H.L. and J.K.; Investigation, H.L., J.K. and H.H.; Writing—original draft, H.L., J.K., K.K. and H.H.; Writing—review & editing, M.Y.; Visualization, H.L. and J.K.; Supervision, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Korea Planning & Evaluation Institute of Industrial Technology funded by the Ministry of the Interior and Safety (MOIS, Korea). [Project Name: Development of Early Detection and Preemptive Response Technology for Black Ice Using Multi-modal Imaging/Project Number: RS-2024-00409314].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Korea Road Traffic Authority. Statistical Analysis of Traffic Accidents on Icy Roads. (In Korea). Available online: https://www.koroad.or.kr/main/board/6/301689/board_view.do?&sv=%EA%B2%B0%EB%B9%99&cp=1&listType=list&bdOpenYn=Y&bdNoticeYn=N (accessed on 8 September 2025).
  2. U.S. Federal Highway Administration (FHWA). Snow and Ice. Available online: https://ops.fhwa.dot.gov/weather/weather_events/snow_ice.htm (accessed on 8 September 2025).
  3. Sezgin, F.; Vriesman, D.; Steinhauser, D.; Lugner, R.; Brandmeier, T. Safe Autonomous Driving in Adverse Weather: Sensor Evaluation and Performance Monitoring. In Proceedings of the 2023 IEEE Intelligent Vehicles Symposium (IV), Anchorage, AK, USA, 4–7 June 2023. [Google Scholar] [CrossRef]
  4. Ramos-Romero, C.; Asensio, C.; Moreno, R.; de Arcas, G. Urban Road Surface Discrimination by Tire-Road Noise Analysis and Data Clustering. Sensors 2022, 22, 9686. [Google Scholar] [CrossRef]
  5. Heo, J.; Im, B.; Shin, S.; Ha, S.; Kang, D.; Song, T.; Gu, Y.; Han, S. Experimental Analysis of Sensor’s Performance Degradation Under Adverse Weather Conditions. Int. J. Precis. Eng. Manuf. 2025, 26, 1655–1672. [Google Scholar] [CrossRef]
  6. Shanbhag, H.; Madani, S.; Isanaka, A.; Nair, D.; Gupta, S.; Hassanieh, H. Contactless Material Identification with Millimeter Wave Vibrometry. In Proceedings of the 21st Annual International Conference on Mobile Systems, Applications and Services (MobiSys ’23), Helsinki, Finland, 18–22 June 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 475–488. [Google Scholar] [CrossRef]
  7. Ma, Y.; Wang, M.; Feng, Q.; He, Z.; Tian, M. Current Non-Contact Road Surface Condition Detection Schemes and Technical Challenges. Sensors 2022, 22, 9583. [Google Scholar] [CrossRef]
  8. Lamane, M.; Tabaa, M.; Klilou, A. New Approach Based on Pix2Pix–YOLOv7 mmWave Radar for Target Detection and Classification. Sensors 2023, 23, 9456. [Google Scholar] [CrossRef] [PubMed]
  9. Yeo, J.; Lee, J.; Kim, G.; Jang, K. Estimation of Road Surface Condition during Summer Season Using Machine Learning. J. Korea Inst. Intell. Transp. Syst. 2018, 17, 121–132. [Google Scholar] [CrossRef]
  10. Lee, H.; Kang, M.; Song, J.; Hwang, K. The Detection of Black Ice Accidents for Preventative Automated Vehicles Using Convolutional Neural Networks. Electronics 2020, 9, 2178. [Google Scholar] [CrossRef]
  11. Soumya, A.; Krishna Mohan, C.; Cenkeramaddi, L.R. Recent Advances in mmWave-Radar-Based Sensing, Its Applications, and Machine Learning Techniques: A Review. Sensors 2023, 23, 8901. [Google Scholar] [CrossRef]
  12. Hao, S.; Liu, B.; Jiang, B.; Lin, Y.; Xu, B. Material Surface Classification Based on 24GHz FMCW MIMO Radar. In Proceedings of the IEEE MTT-S International Wireless Symposium (IWS), Beijing, China, 16–19 May 2024. [Google Scholar] [CrossRef]
  13. Southcott, M.; Zhang, L.; Liu, C. Millimeter Wave Radar-Based Road Segmentation. In Radar Sensor Technology XXVII; SPIE: Bellingham, WA, USA, 2023; Volume 12535. [Google Scholar] [CrossRef]
  14. Cho, J.; Hussen, H.; Yang, S.; Kim, J. Radar-Based Road Surface Classification System for Personal Mobility Devices. IEEE Sens. J. 2023, 23, 16343–16350. [Google Scholar] [CrossRef]
  15. Yang, B.; Zhang, H.; Chen, Y.; Zhou, Y.; Peng, Y. Urban Traffic Imaging Using Millimeter-Wave Radar. Remote Sens. 2022, 14, 5416. [Google Scholar] [CrossRef]
  16. Kim, J.; Kim, E.; Kim, D. A Black Ice Detection Method Based on 1-Dimensional CNN Using mmWave Sensor Backscattering. Remote Sens. 2022, 14, 5252. [Google Scholar] [CrossRef]
  17. Bouwmeester, W.; Fioranelli, F.; Yarovoy, A.G. Road Surface Conditions Identification via HαA Decomposition and Its Application to mm-Wave Automotive Radar. IEEE Trans. Radar Syst. 2023, 1, 132–145. [Google Scholar] [CrossRef]
  18. He, S.; Qian, Y.; Zhang, H.; Zhang, G.; Xu, M.; Fu, L.; Cheng, X.; Wang, H.; Hu, P. Accurate Contact-Free Material Recognition with Millimeter Wave and Machine Learning. In Proceedings of the International Conference on Wireless Algorithms, Systems and Applications, Dalian, China, 24–26 November 2022; pp. 609–620. [Google Scholar] [CrossRef]
  19. Vassilev, V. Road Surface Recognition at mm-Wavelengths Using a Polarimetric Radar. IEEE Trans. Intell. Transp. Syst. 2022, 23, 6985–6990. [Google Scholar] [CrossRef]
  20. Qiu, Z.; Shao, J.; Guo, D.; Yin, X.; Zhai, Z.; Duan, Z.; Xu, Y. A Multi-Feature Fusion Approach for Road Surface Recognition Leveraging Millimeter-Wave Radar. Sensors 2025, 25, 3802. [Google Scholar] [CrossRef] [PubMed]
  21. Liu, J.; Zhang, Y.; Liu, J.; Wang, Z.; Zhang, Z. Automated Recognition of Snow-Covered and Icy Road Surfaces Based on T-Net of Mount Tianshan. Remote Sens. 2024, 16, 3727. [Google Scholar] [CrossRef]
  22. Jia, F.; Li, C.; Bi, S.; Qian, J.; Wei, L.; Sun, G. TC–Radar: Transformer–CNN Hybrid Network for Millimeter-Wave Radar Object Detection. Remote Sens. 2024, 16, 2881. [Google Scholar] [CrossRef]
  23. Bhatia, J.; Dayal, A.; Jha, A.; Vishvakarma, S.K.; Joshi, S.; Srinivas, M.B.; Yalavarthy, P.K.; Kumar, A.; Lalitha, V.; Koorapati, S.; et al. Classification of Targets Using Statistical Features from Range FFT of mmWave FMCW Radars. Electronics 2021, 10, 1965. [Google Scholar] [CrossRef]
  24. Hyun, E.; Jin, Y.-S.; Lee, J.-H. A Pedestrian Detection Scheme Using a Coherent Phase Difference Method Based on 2D Range-Doppler FMCW Radar. Sensors 2016, 16, 124. [Google Scholar] [CrossRef]
  25. Häkli, J.; Säily, J.; Koivisto, P.; Huhtinen, I.; Dufva, T.; Rautiainen, A. Road Surface Condition Detection Using 24 GHz Automotive Radar Technology. In Proceedings of the 14th International Radar Symposium (IRS), Dresden, Germany, 19–21 June 2013; Available online: https://ieeexplore.ieee.org/document/6581661 (accessed on 8 September 2025).
  26. Sabery, S.; Bystrov, A.; Gardner, P.; Stroescu, A.; Gashinova, M. Road Surface Classification Based on Radar Imaging Using Convolutional Neural Network. IEEE Sens. J. 2021, 21, 18725–18732. [Google Scholar] [CrossRef]
  27. Bouwmeester, W.; Fioranelli, F.; Yarovoy, A. Statistical Polarimetric RCS Model of an Asphalt Road Surface for mm-Wave Automotive Radar. In Proceedings of the 20th European Radar Conference (EuRAD), Berlin, Germany, 20–22 September 2023; pp. 18–21. [Google Scholar] [CrossRef]
  28. Tavanti, E.; Rizik, A.; Fedeli, A.; Caviglia, D.; Randazzo, A. A Short-Range FMCW Radar-Based Approach for Multi-Target Human–Vehicle Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 2003816. [Google Scholar] [CrossRef]
  29. Paek, D.; Kong, S.; Wijaya, K.T.K. K-radar: 4D Radar Object Detection for Autonomous Driving in Various Weather Conditions. Adv. Neural Inf. Process. Syst. 2022, 35, 3819–3829. Available online: https://proceedings.neurips.cc/paper_files/paper/2022/hash/185fdf627eaae2abab36205dcd19b817-Abstract-Datasets_and_Benchmarks.html (accessed on 8 September 2025).
  30. Gupta, S.; Rai, P.K.; Kumar, A.; Yalavarthy, P.K.; Cenkeramaddi, L.R. Target Classification by mmWave FMCW Radars Using Machine Learning on Range-Angle Images. IEEE Sens. J. 2021, 21, 19993–20001. [Google Scholar] [CrossRef]
  31. Liang, X.; Chu, L.; Hua, B.; Shi, Q.; Shi, J.; Meng, C.; Braun, R. Road Target Recognition Based on Radar Range-Doppler Spectrum with GS–ResNet. Int. J. Remote Sens. 2024, 45, 8290–8312. [Google Scholar] [CrossRef]
  32. Winkler, V. Range Doppler Detection for Automotive FMCW Radar. In Proceedings of the European Microwave Conference, Munich, Germany, 8–12 October 2007; pp. 166–169. [Google Scholar] [CrossRef]
  33. Moroto, Y.; Maeda, K.; Togo, R.; Ogawa, T.; Haseyama, M. Multimodal Transformer Model Using Time-Series Data to Classify Winter Road Surface Conditions. Sensors 2024, 24, 3440. [Google Scholar] [CrossRef] [PubMed]
  34. Bouwmeester, W.; Fioranelli, F.; Yarovoy, A. Dynamic Road Surface Signatures in Automotive Scenarios. In Proceedings of the 18th European Radar Conference (EuRAD), London, UK, 5–7 April 2022; pp. 285–288. [Google Scholar] [CrossRef]
  35. Wang, Z.; Wen, T.; Chen, N.; Tang, R. Assessment of Landslide Susceptibility Based on the Two-Layer Stacking Model—A Case Study of Jiacha County, China. Remote Sens. 2025, 17, 1177. [Google Scholar] [CrossRef]
  36. Niu, Y.; Li, Y.; Jin, D.; Su, L.; Vasilakos, A.V. A Survey of Millimeter Wave Communications (mmWave) for 5G: Opportunities and Challenges. Wirel. Netw. 2015, 21, 2657–2676. [Google Scholar] [CrossRef]
  37. Dokhanchi, S.H.; Mysore, B.S.; Mishra, K.V.; Ottersten, B. A mmWave Automotive Joint Radar-Communications System. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 1241–1260. [Google Scholar] [CrossRef]
  38. Zhou, T.; Yang, M.; Jiang, K.; Wong, H.; Yang, D. MMW Radar-Based Technologies in Autonomous Driving: A Review. Sensors 2020, 20, 7283. [Google Scholar] [CrossRef]
  39. Kong, H.; Huang, C.; Yu, J.; Shen, X. A Survey of mmWave Radar-Based Sensing in Autonomous Vehicles, Smart Homes and Industry. IEEE Commun. Surv. Tutor. 2025, 27, 463–508. [Google Scholar] [CrossRef]
  40. Zhang, J.; Xi, R.; He, Y.; Sun, Y.; Guo, X.; Wang, W.; Na, X.; Liu, Y.; Shi, Z.; Gu, T. A Survey of mmWave-Based Human Sensing: Technology, Platforms and Applications. IEEE Commun. Surv. Tutor. 2023, 25, 2052–2087. [Google Scholar] [CrossRef]
  41. Weiß, J.; Santra, A. One-Shot Learning for Robust Material Classification Using Millimeter-Wave Radar System. IEEE Sens. Lett. 2018, 2, 7001504. [Google Scholar] [CrossRef]
  42. van Delden, M.; Westerdick, S.; Musch, T. Investigations on Foam Detection Utilizing Ultra-Broadband Millimeter Wave FMCW Radar. In Proceedings of the 2019 IEEE MTT-S International Microwave Workshop Series on Advanced Materials and Processes for RF and THz Applications (IMWS-AMP), Bochum, Germany, 16–18 July 2019; pp. 103–105. [Google Scholar] [CrossRef]
  43. Ciattaglia, G.; Iadarola, G.; Battista, G.; Senigagliesi, L.; Gambi, E.; Castellini, P.; Spinsante, S. Displacement Evaluation by mmWave FMCW Radars: Method and Performance Metrics. IEEE Trans. Instrum. Meas. 2024, 73, 8505313. [Google Scholar] [CrossRef]
  44. Jardak, S.; Alouini, M.S.; Kiuru, T.; Metso, M.; Ahmed, S. Compact mmWave FMCW Radar: Implementation and Performance Analysis. IEEE Aerosp. Electron. Syst. Mag. 2019, 34, 36–44. [Google Scholar] [CrossRef]
  45. Iovescu, C.; Rao, S. The Fundamentals of Millimeter Wave Radar Sensors; Texas Instruments: Dallas, TX, USA, 2020; pp. 1–7. Available online: https://www.ti.com/lit/wp/spyy005a/spyy005a.pdf (accessed on 8 September 2025).
  46. Li, X.; Wang, X.; Yang, Q.; Fu, S. Signal Processing for TDM MIMO FMCW Millimeter-Wave Radar Sensors. IEEE Access 2021, 9, 167959–167971. [Google Scholar] [CrossRef]
  47. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  48. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  49. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  50. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. Adv. Neural Inf. Process. Syst. 2017, 30, 52. Available online: https://proceedings.neurips.cc/paper/2017/hash/6449f44a102fde848669bdd9eb6b76fa-Abstract.html (accessed on 8 September 2025).
  51. Ghaderpour, E.; Pagiatakis, S.D.; Mugnozza, G.S.; Mazzanti, P. On the Stochastic Significance of Peaks in the Least-Squares Wavelet Spectrogram and an Application in GNSS Time Series Analysis. Signal Process. 2024, 223, 109581. [Google Scholar] [CrossRef]
  52. Boateng, E.Y.; Otoo, J.; Abaye, D.A. Basic Tenets of Classification Algorithms: K-Nearest-Neighbor, Support Vector Machine, Random Forest and Neural Network—A Review. J. Data Anal. Inf. Process. 2020, 8, 341–357. [Google Scholar] [CrossRef]
  53. Sheykhmousa, M.; Mahdianpari, M.; Ghanbari, H.; Mohammadimanesh, F.; Ghamisi, P.; Homayouni, S. Support Vector Machine versus Random Forest for Remote Sensing Image Classification: A Meta-Analysis and Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6308–6325. [Google Scholar] [CrossRef]
  54. Liang, W.; Luo, S.; Zhao, G.; Wu, H. Predicting Hard Rock Pillar Stability Using GBDT, XGBoost, and LightGBM Algorithms. Mathematics 2020, 8, 765. [Google Scholar] [CrossRef]
  55. Shehadeh, A.; Alshboul, O.; Al Mamlook, R.E.; Hamedat, O. Machine Learning Models for Predicting the Residual Value of Heavy Construction Equipment: An Evaluation of Modified Decision Tree, LightGBM, and XGBoost Regression. Autom. Constr. 2021, 129, 103827. [Google Scholar] [CrossRef]
  56. LeCun, Y.; Bengio, Y. Convolutional Networks for Images, Speech, and Time Series. In The Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, MA, USA, 1995; p. 3361. Available online: https://hal.science/hal-05083427/document (accessed on 8 September 2025).
  57. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent Advances in Convolutional Neural Networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  58. Tabian, I.; Fu, H.; Sharif Khodaei, Z. A Convolutional Neural Network for Impact Detection and Characterization of Complex Composite Structures. Sensors 2019, 19, 4933. [Google Scholar] [CrossRef] [PubMed]
  59. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. Available online: https://arxiv.org/pdf/2010.11929/1000 (accessed on 8 September 2025). [CrossRef]
  60. Texas Instruments. IWR1843 Single-Chip 76- to 81-GHz FMCW mmWave Sensor Datasheet; Texas Instruments: Dallas, TX, USA, 2022; Available online: https://www.ti.com/lit/pdf/SWRS228 (accessed on 8 September 2025).
  61. Browne, M.W. Cross-Validation Methods. J. Math. Psychol. 2000, 44, 108–132. [Google Scholar] [CrossRef]
  62. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. Available online: https://proceedings.neurips.cc/paper/2012/hash/c399862d3b9d6b76c8436e924a68c45b-Abstract.html (accessed on 8 September 2025). [CrossRef]
  63. Morgan, N.; Bourlard, H. Generalization and Parameter Estimation in Feedforward Nets: Some Experiments. Adv. Neural Inf. Process. Syst. 1989, 2, 630–637. Available online: https://proceedings.neurips.cc/paper_files/paper/1989/hash/63923f49e5241343aa7acb6a06a751e7-Abstract.html (accessed on 8 September 2025).
  64. Krogh, A.; Hertz, J. A Simple Weight Decay Can Improve Generalization. Adv. Neural Inf. Process. Syst. 1991, 4, 950–957. Available online: https://proceedings.neurips.cc/paper_files/paper/1991/file/8eefcfdf5990e441f0fb6f3fad709e21-Paper.pdf (accessed on 8 September 2025).
  65. Chen, Y.; Gu, X.; Liu, Z.; Liang, J. A Fast Inference Vision Transformer for Automatic Pavement Image Classification and Its Visual Interpretation Method. Remote Sens. 2022, 14, 1877. [Google Scholar] [CrossRef]
  66. Wang, Y.; Sun, X.; Wen, T.; Wang, L. Step-Like Displacement Prediction of Reservoir Landslides Based on a Metaheuristic-Optimized KELM: A Comparative Study. Bull. Eng. Geol. Environ. 2024, 83, 322. [Google Scholar] [CrossRef]
  67. Wei, Z.; Zhang, F.; Chang, S.; Liu, Y.; Wu, H.; Feng, Z. Mmwave Radar and Vision Fusion for Object Detection in Autonomous Driving: A Review. Sensors 2022, 22, 2542. [Google Scholar] [CrossRef]
  68. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing Through Fog Without Seeing Fog: Deep Multimodal Sensor Fusion in Unseen Adverse Weather. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11682–11692. Available online: https://openaccess.thecvf.com/content_CVPR_2020/html/Bijelic_Seeing_Through_Fog_Without_Seeing_Fog_Deep_Multimodal_Sensor_Fusion_CVPR_2020_paper.html (accessed on 8 September 2025).
  69. Nabati, R.; Qi, H. CenterFusion: Center-Based Radar and Camera Fusion for 3D Object Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 5–9 January 2021; pp. 1527–1536. Available online: https://openaccess.thecvf.com/content/WACV2021/html/Nabati_CenterFusion_Center-Based_Radar_and_Camera_Fusion_for_3D_Object_Detection_WACV_2021_paper.html (accessed on 8 September 2025).
  70. Nobis, F.; Shafiei, E.; Karle, P.; Betz, J.; Lienkamp, M. Radar Voxel Fusion for 3D Object Detection. Appl. Sci. 2021, 11, 5598. [Google Scholar] [CrossRef]
Figure 1. Graphical representation of the IF signal.
Figure 1. Graphical representation of the IF signal.
Sensors 25 05697 g001
Figure 2. System configuration of the mmWave radar-based detection setup.
Figure 2. System configuration of the mmWave radar-based detection setup.
Sensors 25 05697 g002
Figure 3. Experimental setup inside the temperature- and humidity-controlled chamber.
Figure 3. Experimental setup inside the temperature- and humidity-controlled chamber.
Sensors 25 05697 g003
Figure 4. Surface conditions by material type.
Figure 4. Surface conditions by material type.
Sensors 25 05697 g004
Figure 5. Data processing and feature extraction pipeline.
Figure 5. Data processing and feature extraction pipeline.
Sensors 25 05697 g005
Figure 6. Power spectrum image samples for CNN-based classification.
Figure 6. Power spectrum image samples for CNN-based classification.
Sensors 25 05697 g006
Figure 7. Asphalt-classification confusion matrix and classification performance metrics of (a) Random Forest, (b) SVM, (c) LightGBM, (d) XGBoost, and (e) CNN Models in the 0.04–0.30 m range.
Figure 7. Asphalt-classification confusion matrix and classification performance metrics of (a) Random Forest, (b) SVM, (c) LightGBM, (d) XGBoost, and (e) CNN Models in the 0.04–0.30 m range.
Sensors 25 05697 g007
Figure 8. Concrete classification confusion matrix and classification performance metrics of (a) Random Forest, (b) SVM, (c) LightGBM, (d) XGBoost, and (e) CNN Models in the 0.04–0.30 m range.
Figure 8. Concrete classification confusion matrix and classification performance metrics of (a) Random Forest, (b) SVM, (c) LightGBM, (d) XGBoost, and (e) CNN Models in the 0.04–0.30 m range.
Sensors 25 05697 g008
Figure 9. Confusion matrices for the best performing ViT models: (a) asphalt specimens (Fold 2, 99.75% accuracy) and (b) concrete specimens (Fold 1, 98.58% accuracy), demonstrating high classification accuracy across all road surface conditions.
Figure 9. Confusion matrices for the best performing ViT models: (a) asphalt specimens (Fold 2, 99.75% accuracy) and (b) concrete specimens (Fold 1, 98.58% accuracy), demonstrating high classification accuracy across all road surface conditions.
Sensors 25 05697 g009
Figure 10. Comprehensive performance comparison between CNN and ViT models under various robustness conditions. The chart shows accuracy patterns across different noise and blur levels for both asphalt and concrete specimens, demonstrating the relative robustness of each model architecture.
Figure 10. Comprehensive performance comparison between CNN and ViT models under various robustness conditions. The chart shows accuracy patterns across different noise and blur levels for both asphalt and concrete specimens, demonstrating the relative robustness of each model architecture.
Sensors 25 05697 g010
Table 1. Characteristics of learning models.
Table 1. Characteristics of learning models.
ModelSummary of FeaturesStrengthsLimitationsRecommended Use Cases
Random ForestEnsemble of multiple decision trees using random sampling
  • Prevents overfitting
  • Easy to analyze feature importance
  • Suitable for both classification and regression
  • Slower prediction speed
  • High memory consumption
  • Feature importance analysis
  • Non-linear pattern analysis
SVMLearns decision boundaries by maximizing the margin in high-dimensional space
  • Strong performance on small datasets
  • Can model complex boundaries
  • Solid theoretical foundation
  • Inefficient on large datasets
  • Sensitive to kernel and hyperparameter selection
  • Well-separated classes
  • Limited data
XGBoostBoosted tree model with enhanced performance and efficiency over traditional gradient boosting
  • Fast training speed
  • High prediction performance
  • Handles missing values effectively
  • Requires extensive hyperparameter tuning
  • High memory usage
  • Critical accuracy tasks
  • Performance-optimized tasks
LightGBMFaster and more memory-efficient boosting model optimized for large datasets
  • Very fast training speed
  • Low memory consumption
  • Strong with large-scale data
  • May underperform on small datasets
  • Risk of overfitting
  • Large datasets
  • Fast or real-time classification
CNNAutomatically extracts spatial features from 2D images
  • Excellent image pattern recognition
  • Minimal preprocessing
  • Learns complex structures
  • Requires large datasets
  • Long training times
  • Image-based analysis
  • Visual pattern recognition
ViTTransformer-based architecture that processes images as sequences of patches
  • Superior attention mechanism
  • Excellent long-range dependencies
  • Strong generalization capability
  • Requires very large datasets
  • High computational complexity
  • Memory intensive
  • Complex visual patterns
  • Attention-based analysis
  • Transfer learning tasks
Table 2. Machine learning models’ parameters.
Table 2. Machine learning models’ parameters.
ModelParameterSetting Value
Random Forestn_estimators100
max_depthNone
min_samples_split2
SVMkernelrbf
XGBoostn_estimators100
max_depth6
learning_rate0.3
LightGBMn_estimators100
num_leaves31
learning_rate0.1
Table 3. CNN parameters.
Table 3. CNN parameters.
ParameterSetting Value
Learning rate 1 × 10 4
Weight decay 1 × 10 4
Batch size16
Max epochs50
Early stopping10
OptimizerAdam
Loss FunctionCross Entropy Loss
Table 4. ViT parameters.
Table 4. ViT parameters.
ParameterSetting Value
Model Architecturegoogle/vit-base-patch16-224
Learning rate 2 × 10 5
Weight decay0.01
Batch size8
Max epochs50
Early stopping15
OptimizerAdamW
Loss FunctionCross Entropy Loss
Layers12
Attention Heads12
Table 5. Sensitivity and ablation analysis by range interval and feature set.
Table 5. Sensitivity and ablation analysis by range interval and feature set.
SpecimenRange (m)Feature Set# FeaturesCross-Validation Accuracy
Asphalt/Concrete0.10–0.24Range-FFT50.904/0.888
Range-FFT + ST-LSSA160.898/0.886
ST-LSSA110.350/0.359
0.07–0.27Range-FFT50.895/0.875
Range-FFT + ST-LSSA160.887/0.880
ST-LSSA110.350/0.359
0.04–0.30Range-FFT50.901/0.881
Range-FFT + ST-LSSA160.896/0.878
ST-LSSA110.350/0.359
Table 6. Sample of statistical features: asphalt surface under dry condition.
Table 6. Sample of statistical features: asphalt surface under dry condition.
Mean
(dB)
Max
(dB)
Std
(dB)
Median
(dB)
Mode
(dB)
File_NameLabel
74.5082.046.13973.9564.98radar_0_20250701_112732.pcapDry
78.4692.5710.8877.3859.97radar_0_20250702_165220.pcapDry
78.5093.4111.0476.1960.89radar_0_20250702_165948.pcapDry
Table 7. Classification accuracy results by specimen material and model.
Table 7. Classification accuracy results by specimen material and model.
SpecimenRange (m)ModelAccuracy (%)
Asphalt/Concrete0.10–0.24RF90.38/88.83
SVM59.37/52.17
XGBoost91.48/88.73
LightGBM91.22/89.35
CNN98.67/97.42
0.07–0.27RF89.53/87.95
SVM57.32/44.12
XGBoost90.18/88.70
LightGBM90.33/89.18
CNN99.00/98.50
0.04–0.30RF90.08/88.08
SVM58.50/51.60
XGBoost90.40/88.12
LightGBM90.53/88.45
CNN99.25/98.75
Table 8. Robustness evaluation under Gaussian noise and motion blur for CNN and Transformer. Values are accuracy (%) with drop vs. clean in parentheses (pp).
Table 8. Robustness evaluation under Gaussian noise and motion blur for CNN and Transformer. Values are accuracy (%) with drop vs. clean in parentheses (pp).
SpecimenModelClean (%)Gaussian Noise (SNR dB)Motion Blur (k)
20 dB15 dBk = 3k = 5
AsphaltCNN99.8399.17 ( 0.67 )95.75 ( 4.08 )95.92 ( 3.92 )95.08 ( 4.75 )
ViT99.7599.70 ( 0.05 )91.15 ( 8.60 )96.20 ( 3.55 )90.02 ( 9.73 )
ConcreteCNN98.9298.75 ( 0.17 )95.25 ( 3.67 )95.58 ( 3.33 )94.17 ( 4.75 )
ViT98.5898.47 ( 0.11 )88.97 ( 9.61 )91.36 ( 7.22 )85.89 ( 12.69 )
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, H.; Kim, J.; Ko, K.; Han, H.; Youm, M. Radar-Based Road Surface Classification Using Range-Fast Fourier Transform Learning Models. Sensors 2025, 25, 5697. https://doi.org/10.3390/s25185697

AMA Style

Lee H, Kim J, Ko K, Han H, Youm M. Radar-Based Road Surface Classification Using Range-Fast Fourier Transform Learning Models. Sensors. 2025; 25(18):5697. https://doi.org/10.3390/s25185697

Chicago/Turabian Style

Lee, Hyunji, Jiyun Kim, Kwangin Ko, Hak Han, and Minkyo Youm. 2025. "Radar-Based Road Surface Classification Using Range-Fast Fourier Transform Learning Models" Sensors 25, no. 18: 5697. https://doi.org/10.3390/s25185697

APA Style

Lee, H., Kim, J., Ko, K., Han, H., & Youm, M. (2025). Radar-Based Road Surface Classification Using Range-Fast Fourier Transform Learning Models. Sensors, 25(18), 5697. https://doi.org/10.3390/s25185697

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop