Next Article in Journal
Analysis of SD-WAN Architectures and Techniques for Efficient Traffic Control Under Transmission Constraints—Overview of Solutions
Previous Article in Journal
Robust Occupant Behavior Recognition via Multimodal Sequence Modeling: A Comparative Study for In-Vehicle Monitoring Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Radar-Based Gesture Recognition Using Adaptive Top-K Selection and Multi-Stream CNNs

Department of Electronic Engineering, Kumoh National Institute of Technology, Gumi-si 39177, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(20), 6324; https://doi.org/10.3390/s25206324
Submission received: 17 September 2025 / Revised: 9 October 2025 / Accepted: 9 October 2025 / Published: 13 October 2025
(This article belongs to the Special Issue Sensor Technologies for Radar Detection)

Abstract

With the proliferation of the Internet of Things (IoT), gesture recognition has attracted attention as a core technology in human–computer interaction (HCI). In particular, mmWave frequency-modulated continuous-wave (FMCW) radar has emerged as an alternative to vision-based approaches due to its robustness to illumination changes and advantages in privacy. However, in real-world human–machine interface (HMI) environments, hand gestures are inevitably accompanied by torso- and arm-related reflections, which can also contain gesture-relevant variations. To effectively capture these variations without discarding them, we propose a preprocessing method called Adaptive Top-K Selection, which leverages vector entropy to summarize and preserve informative signals from both hand and body reflections. In addition, we present a Multi-Stream EfficientNetV2 architecture that jointly exploits temporal range and Doppler trajectories, together with radar-specific data augmentation and a training optimization strategy. In experiments on the publicly available FMCW gesture dataset released by the Karlsruhe Institute of Technology, the proposed method achieved an average accuracy of 99.5%. These results show that the proposed approach enables accurate and reliable gesture recognition even in realistic HMI environments with co-existing body reflections.

1. Introduction

The development of the Internet of Things (IoT) has driven innovation across domains such as smart homes, healthcare, transportation, industrial environments, and defense by seamlessly connecting people, objects, and the environment [1,2]. At the same time, the growing demand for hygienic and natural system control without physical contact has drawn increasing attention to contactless user interfaces [3,4,5]. In the field of human–computer interaction (HCI), technologies such as speech recognition and gesture recognition are being developed as alternatives to traditional interfaces, including keyboards, mouses, and touch screens [6]. However, speech recognition technologies remain vulnerable to voice variations and noisy environments, while also raising concerns about privacy [7,8]. In contrast, gesture recognition offers an intuitive form of user input, is free from language barriers, and can be effectively used in silent environments, thereby establishing itself as a core technology in HCI [9,10]. The advancement of deep learning architectures and optimization strategies has substantially improved gesture recognition performance across diverse modalities [11]. In particular, progress in vision-based gesture recognition demonstrates that modern neural models can effectively capture subtle temporal and spatial variations, enabling accurate discrimination of complex motion patterns [12].
Numerous studies that combine optical vision methods—including RGB cameras [13], depth cameras [14], LiDAR [15], and thermal infrared cameras [16]—with deep learning have demonstrated excellent performance. However, such approaches are vulnerable to illumination changes and occlusion, and they process video or audio data that can be directly interpreted by humans, thereby raising privacy concerns [6]. In contrast, radar measures only reflect electromagnetic waves, so the data do not directly resemble human sensory inputs [17]. Owing to the penetrability of radio waves, radar is also unaffected by lighting or shadows and is relatively advantageous in terms of privacy [18]. In particular, mmWave frequency-modulated continuous-wave (FMCW) radar, with its wide bandwidth and fine temporal and spectral resolution, can simultaneously capture subtle variations in range and Doppler, making it highly suitable for implementing short-range gesture interfaces such as smart home control and in-vehicle human–machine interfaces (HMIs).
Previous studies on FMCW radar-based gesture recognition have shown excellent performance by combining diverse signal processing pipelines with deep learning techniques [19,20,21,22,23,24,25,26]. However, many of these works relied on restricted environments and assumed the suppression of non-hand reflections, which differs from real HMI scenarios [18]. In practice, sensors are oriented toward the user, and hand gestures are accompanied by torso- and arm-related reflections. Gestures arise from a kinematic chain involving the shoulder, arm, and hand [27]. Therefore, torso-related signals may also contain meaningful variations functionally related to gestures. Motivated by this observation, we propose a method that effectively summarizes body-related components to preserve gesture information, thereby enabling reliable gesture recognition in realistic HMI conditions. The main contributions of this study are threefold: (i) an entropy-based Adaptive Top-K Selectionpreprocessing method to mitigate information loss and attenuation, (ii) a Multi-Stream EfficientNetV2 architecture for jointly learning range and Doppler trajectories, and (iii) radar-specific data augmentation with a training optimization strategy to maximize model performance. To the best of our knowledge, no prior work has systematically compared different 1D vector compression methods for Range–Doppler Images, and this study is the first to provide such a comparative analysis along with a novel adaptive algorithm. Through these contributions, the proposed method was experimentally validated using datasets, achieving stable and accurate gesture recognition even under realistic conditions where body reflections coexist, and demonstrating improved performance compared to existing methods. The remainder of this paper is organized as follows. Section 2 reviews related works and formulates the problem statement. Section 3 introduces the proposed methodology, including Range–Doppler Image generation, the entropy-based Adaptive Top-K Selection algorithm, and the Multi-Stream EfficientNetV2 architecture with radar-specific data augmentation. Section 4 describes the experimental setup, training procedure, and evaluation results. Finally, Section 5 concludes the paper and discusses future research directions.

2. Related Works and Problem Statement

2.1. Related Works

Radar-based gesture recognition techniques have been studied using various radar sensors, sensing methods, and algorithms. Tiwari et al. [28] utilized wearable UWB antennas to classify six arm gestures based on S-parameters and achieved a high accuracy of 99.76% using Extreme Gradient Boosting (XGB). Despite their excellent performance, wearable systems require direct attachment to the body, which reduces usability in practical applications.
As a result, fully contactless approaches have become mainstream, where radar sensors directly sense the user. A representative example is Google’s Soli project [29], which employed a 60 GHz mmWave FMCW radar with a Random Forest classifier to recognize gestures, followed by many subsequent studies that developed high-accuracy systems. Wang et al. [24] proposed an end-to-end model combining a 2D CNN and a Long Short-Term Memory (LSTM) network using per-frame Range–Doppler Maps (RDMs, used interchangeably with RDIs in this paper) as input, achieving 88% accuracy in subject-independent classification of 11 gestures. Choi et al. [25] extracted one-dimensional motion profiles from RDMs as LSTM inputs and reported a high accuracy of 98.48% for 10 gestures. Hayashi et al. [26] employed the second-generation Soli chip to develop RadarNet, a multi-head architecture combining 2D CNNs and LSTMs capable of simultaneously predicting left–right and up–down swipe gestures, achieving over 99% accuracy on pre-segmented datasets.
RNN-based models have also been widely adopted. Suh et al. [19] introduced the Projected RDM (PRDM) and used it as LSTM input for real-time recognition of seven gestures, reaching over 91% accuracy. These results demonstrate that sequential spatiotemporal information is vital for robust gesture recognition. Extending this idea, Zhang et al. [21] combined a 3D CNN with an LSTM using an FMCW radar, achieving 96% accuracy over eight gestures. However, the use of 3D CNNs significantly increases memory consumption and computational complexity.
Beyond RNN-based architectures, several studies have investigated the extraction of spatiotemporal feature maps from frame-wise radar data through dimensionality reduction and their use with 2D CNNs. Chmurski et al. [22] proposed a method to derive range–time, Doppler–time, and angle–time maps from FMCW radar signals and applied a 2D CNN for gesture classification, achieving a test accuracy of 98.13% across eight gestures. In a similar direction, Ahmed et al. [20] introduced a multi-stream 2D CNN for digit recognition that jointly processes range–time, Doppler–time, and angle–time maps, reporting 94.2% accuracy. However, generating angle–time maps requires direction-of-arrival (DOA) estimation via algorithms such as MUSIC or MVDR [30], which are computationally expensive due to eigen-decomposition and matrix inversion.
Although these studies have demonstrated high accuracy under controlled conditions, most datasets were collected in limited environments (e.g., confined boxes, simple backgrounds) and mainly included isolated hand gestures [19,20,22,25]. In realistic HMI scenarios, sensors face users in open spaces. With wider detection ranges, reflections from the torso and upper body inevitably coexist with hand gestures, resulting in self-reflections (DC leakage), static background signals, non-hand body responses, and electronic noise [31]. To suppress such noise, Moving Target Indicator (MTI) techniques such as moving average and high-pass filters have been widely applied [32,33,34]. However, these techniques may remove functionally important slow gesture components and subtle torso movements, leading to potential information loss. Moreover, conventional methods typically involve computationally expensive operations such as eigen-decomposition and matrix inversion. In contrast, our method preserves gesture-relevant information by summarizing body-related components rather than simply discarding them. Furthermore, it employs a multi-stream 2D CNN architecture that avoids eigen-decomposition and matrix inversion, enabling robust and reliable gesture recognition even in realistic HMI conditions.

2.2. Dataset and Problem Statement

The dataset used in this study was introduced by Antes et al. [35] and provides raw FMCW radar data corresponding to seven gesture types plus a No Gesture class [36]. Hereafter, this dataset is referred to as the KIT radar gesture dataset.
In the original dataset paper [35], Part 1 and Part 2 were combined to increase the number of samples per class. In addition, G 5 show 2 fingers, and G 6 show 4 fingers, were excluded, as these gestures are highly similar to G 4 show 1 finger, contain relatively fewer samples, and are more prone to misclassification in practical applications. Therefore, this study also followed the same configuration and used six gesture classes for classification. After excluding seven None Value Data samples, a total of 2208 samples were used for experiments. Figure 1 summarizes the detailed information of each gesture class.
Data acquisition was conducted using the BGT60TR13C demo board from Infineon Technologies AG, which is equipped with one transmit antenna and three receive antennas [37]. Figure 2 shows the radar demo board with the transmit and receive antennas highlighted.
The FMCW radar transmits chirp waveforms with linearly modulated frequencies, and the reflected signals exhibit a frequency shift relative to the transmitted signal due to the time delay. This frequency difference generates a beat frequency containing the target’s range information. The transmit antenna periodically emits chirp signals, while the three receive antennas capture the reflected signals. These received signals are mixed with the transmitted signal to produce intermediate frequency (IF) signals, which are then digitized by an analog-to-digital converter (ADC) and stored as raw FMCW radar data for each channel.
According to the original paper introducing the KIT radar gesture dataset, the subject’s torso is included within the radar sensing range, as illustrated in Figure 3, providing conditions similar to real HMI environments. In such a measurement setting, gesture-related information may exist not only in the hands and arms but also in the torso; therefore, removing torso-related reflections carries the risk of discarding useful signals associated with gestures.
In this context, the main challenge of this study is to propose a preprocessing method that effectively preserves gesture information in radar data containing both hand gestures and body-related reflections and to improve deep learning classification performance.

3. Proposed Methodology

3.1. Range–Doppler Image Generation

Each sample collected by the radar sensor (see Figure 2) consists of multiple frames, and each frame forms a 2D real-valued array with N C chirps and N S samples. A total of three receive channels are available, and identical data structures are collected for each channel. The signal of one frame from a specific receive channel a { 0 , 1 , 2 } can be expressed as follows:
X raw ( f , a ) = x 1 , 1 ( f , a ) x 1 , N S ( f , a ) x N C , 1 ( f , a ) x N C , N S ( f , a ) R N C × N S
where x c , s ( f , a ) denotes the real-valued signal collected at the s-th sample of the c-th chirp in the f-th frame of receiving channel a. The sensor measurement parameters of the dataset are shown in Table 1, with each frame consisting of N C = 600 chirps and each chirp containing N S = 128 samples. These measurement parameters follow standard FMCW radar specifications, whose physical interpretations (e.g., the relation of N S , N C , and T C to range and Doppler resolutions) have been described in prior works [32,38]. Here, we summarize only the parameters directly used in our preprocessing pipeline. In the following description, the frame and channel indices are fixed, and the signal is simply denoted as X raw R N C × N S .
From X raw , in order to extract range and velocity components generated by multiple reflectors, a Range–Doppler Image (RDI) is generated across the three receiving channels. The RDI generation process is as follows. First, DC leakage and static clutter components are suppressed by applying 2D mean removal. X raw is processed by subtracting the column-wise mean and then the row-wise mean, resulting in a zero-mean matrix X zero . Each chirp of X zero is multiplied by a Hamming window to reduce sidelobes [39], and a 1D Fast Fourier Transform (FFT) is applied along the sample axis to extract range profiles. This 1D FFT applied along the sample axis is referred to as the Range FFT. To improve range resolution, a zero-padded FFT of N FFT = 448 points is applied to the original N S = 128 samples. Finally, only the positive frequency components are retained, yielding a spectrum with N R = 224 range bins. The result of the Range FFT is given as
X range ( c , r ) = s = 1 N S w r ( s ) · X zero ( c , s ) · exp j 2 π r · s N FFT , r 1 , N FFT 2
where w r denotes the Hamming window applied in the range direction. For each range bin r, a Hamming window w d is applied in the chirp direction, followed by a 1D FFT to extract Doppler components. This process is called the Doppler FFT. Since the FFT output is complex-valued, its magnitude is taken to obtain a matrix representing Doppler intensity at each range bin. The result of the Doppler FFT is expressed as
X doppler ( d , r ) = c = 1 N C w d ( c ) · X range ( c , r ) · exp j 2 π d · c N C
Subsequently, to emphasize target reflection signals while suppressing background noise and ghost targets, the Cell-Averaging Constant False Alarm Rate (CA-CFAR) algorithm is applied [40]. In this study, CA-CFAR is implemented on the Doppler spectrum X doppler of the Range–Doppler plane by sliding a 2D window-based kernel. Figure 4 illustrates the core structure of the 2D CA-CFAR window configuration.
At this stage, for each Cell Under Test (CUT), a binary mask M ( d , r ) is generated by comparing the cell value with a dynamic threshold. The dynamic threshold is obtained by multiplying a scaling factor α with the local noise level μ ( d , r ) estimated from surrounding reference cells.
M ( d , r ) = 1 , if X doppler ( d , r ) > α · μ ( d , r ) 0 , otherwise
where μ ( d , r ) denotes the mean amplitude calculated over ( 2 T + 2 G + 1 ) 2 ( 2 G + 1 ) 2 training cells, excluding both the guard cells and the CUT. In this study, the CA-CFAR parameters were empirically set to T = 20 , G = 2 , and α = 3.0 . Then, the Doppler magnitude spectrum X doppler is multiplied by the binary mask and transformed through logarithmic scaling to produce the final RDI R N D × N R :
RDI = log 1 + X doppler M = RDI ( 1 , 1 ) RDI ( 1 , N R ) RDI ( N D , 1 ) RDI ( N D , N R ) R N D × N R
where ⊙ represents the Hadamard (element-wise) product between two matrices of the same dimension. The term log ( 1 + · ) avoids undefined values when the input is zero and prevents low-amplitude components from being completely suppressed during logarithmic scaling.

3.2. Adaptive Top-K Selection Based RTM and DTM Generation

The RDI is a 2D array of size N D × N R that represents the energy distribution of gestures and background reflections within a single frame. A data sample from one channel has a 3D structure in which RDIs are arranged sequentially across frames. Each frame-level RDI is compressed into a 1D vector along either the Doppler (row) or Range (column) dimension, and these vectors are stacked along the temporal axis to generate a 2D time-series map. These time-series maps are defined as the Doppler-Time Map (DTM) and the Range-Time Map (RTM).

3.2.1. Compression Methods from RDI to the 1D Vector

Since motion-induced components from non-hand body parts were not removed in the preprocessing step described in Section 3.1, the RDI exhibits multiple peak clusters, as shown in Figure 5. In this example, the cluster on the left corresponds to reflections from the hand gesture, while the cluster on the right originates from the torso. The highlighted boxes indicate localized energy distributions, with the yellow box marking gesture-related peaks and the white box marking torso-related reflections.
Several methods can be used to compress a frame-level RDI into a 1D vector:
(a)
Summation across rows or columns:
This method utilizes all available information and is robust to noise. However, strong peaks may be diluted by background energy, leading to blurred features.
(b)
Maximum extraction from each row or column: This approach emphasizes the strongest peaks, resulting in clearer features but may incorrectly highlight noise peaks or body-related reflections.
(c)
Slicing at the maximum peak position: This method suffers from severe information loss, and if the maximum peak originates from noise rather than the gesture, feature distortion may occur.
(d)
Top-K summation after sorting by magnitude: This method preserves information around strong reflectors and mitigates noise effects. However, the balance between information preservation and noise suppression depends on the choice of K.
Figure 6 shows four compression methods using a reduced RDI for illustration. Arrows indicate the compression direction: horizontal for Doppler–time (DTM) and vertical for range–time (RTM). Figure 6a–d present the resulting 1D vectors obtained by each method. Considering their respective advantages and disadvantages, this study adopts method Figure 6d, the Top-K summation approach, to compress RDIs into 1D vectors.

3.2.2. Adaptive Top-K Selection Algorithm

Using a fixed K value to sum only the Top-K components cannot reflect the diverse energy dispersion characteristics of each row or column vector in an RDI. To address this issue, this study proposes the Adaptive Top-K Selection algorithm, which quantifies the dispersion of each vector and automatically selects an optimal K.
CA-CFAR masking converts the RDI into a sparse matrix in which most values are zero, while valid reflections remain only around peak clusters (Figure 5). In such sparse peak clusters, the most informative pixels are typically surrounded by sidelobe components, and the fewer surrounding nonzero pixels there are, the less informative the corresponding bin tends to be. Therefore, an appropriate dispersion measure is required to quantify the distribution of values in each row or column vector. For this purpose, Shannon entropy [41] was adopted, defined for a normalized vector p i = v i / ( j v j + ϵ ) as
H = i = 1 n p i log p i .
This entropy ranges from H = 0 when all energy is concentrated in a single bin (perfectly peaked vector) to H = log n when it is uniformly distributed across all bins (maximally diffuse vector).
As H decreases, the vector is more likely dominated by sidelobe-like components of lower importance, whereas a higher H indicates a vector with more evenly distributed and potentially meaningful components. Accordingly, the K-selection strategy is defined as
  • Peak distribution (low entropy): A larger K is selected to include surrounding components, thereby diluting sidelobe peaks and enhancing the relative emphasis on vectors containing richer information.
  • Diffuse distribution (high entropy): A smaller K is chosen to avoid the inclusion of unnecessary zeros, preserving only the core components and maintaining structural clarity.
Figure 7 illustrates an example of entropy-based K determination, and Table 2 summarizes the step-by-step procedure. Low-entropy vectors use a larger K to include adjacent entries, whereas high-entropy vectors use a smaller K to retain only core components.
This rule implies that, in Top-K summation, a larger K includes more zero or near-zero elements. As a result, the normalized output becomes visually blurred. In contrast, a smaller K produces a sharper result by concentrating on the strongest components. Thus, the entropy-based selection adaptively balances attenuation and emphasis according to the information content of each vector. For RDIs containing multiple peak clusters, even when entropy is high and a small K is chosen, K min should be set greater than 1 to reduce information loss. Conversely, if entropy is low and K is large, setting K max to be excessively high may also cause attenuation. In this study, we heuristically set K min = 5 and K max = 20 .
Once K is determined, the Top-K elements of each column (range bin) are summed to form an RTM vector, and the Top-K elements of each row (Doppler bin) are summed to form a DTM vector. This entire procedure—entropy-based K selection followed by summation—is hereafter referred to as the Adaptive Top-K Summation.

3.2.3. RTM and DTM Generation

Stacking the RTM and DTM vectors from all frames in chronological order yields two distinct images that represent the temporal range trajectory and velocity trajectory of the gesture. After applying Min-Max Normalization to each image and resizing them to 224 × 224 , the final RTM and DTM corresponding to a single receiving channel are obtained. Figure 8 shows examples of RTM and DTM generated from Rx0 for each gesture class.

3.3. Multi-Stream EfficientNetV2

For each receiving channel, one RTM and one DTM are generated, yielding three RTMs and three DTMs. The three RTMs are stacked into a single three-channel image for 2D CNN input, and the same is carried out for the DTMs. Thus, each sample is represented by two three-channel images, which are independently processed by 2D CNN backbones and then fused for final classification.
In this study, the EfficientNetV2-B0 was adopted as the backbone for each RTM and DTM stream, as it provides a good balance between accuracy and efficiency. EfficientNet, EfficientNetV2, and their variants are convolutional neural networks (CNNs) that have been widely applied to diverse computer vision classification tasks across various domains [42,43,44,45,46]. EfficientNetV2 builds upon its predecessor, the EfficientNet architecture [47], and provides high accuracy, strong parameter efficiency, and fast GPU training speed [48]. Pre-trained weights from the ImageNet dataset were not used; instead, only the architecture was employed. The model was implemented using TensorFlow and Keras libraries and initialized through the tf.keras.applications.EfficientNetV2B0 function. Figure 9 illustrates the architecture of the proposed Multi-EffNetV2 network. The RTM and DTM streams each take an input image of size 224 × 224 × 3 , extract features using EfficientNetV2-B0, and then perform final classification into six gesture classes through fully connected layers. The proposed Multi-EffNetV2 contains a total of 12.38 million trainable parameters and requires approximately 2.89 GFLOPs per inference. Although heavier than lightweight backbones such as MobileNet, the model is still much smaller than vision transformers or 3D CNNs, offering a practical balance between accuracy and efficiency.

3.4. Radar-Specific Data Augmentation

The RTM and DTM used as inputs to the proposed model are two-dimensional time-series images (spectrograms) obtained from the FMCW radar domain. Therefore, directly applying conventional geometric augmentation techniques commonly used in image classification (e.g., rotation, flipping, color transformation) may distort the inherent meaning embedded in the time–frequency structure of the original signals, potentially leading to degraded performance. To address this issue, inspired by the time- and Doppler-domain scaling concepts proposed by Kern et al. [49], we propose three radar-specific augmentation methods:
1.
RTM Vertical Shift: The RTM is randomly shifted along the range axis (vertical direction), with the shift magnitude defined as γ r U ( 30 , + 30 ) pixels. Empty regions created by the shift are filled with zeros, and the same shift is applied to all three receiving channels. This simulates changes in the absolute distance between the sensor and the user.
RTM ( r , t ) = RTM ( r + γ r , t )
2.
RTM and DTM Horizontal Stretch: The RTM and DTM are scaled along the time axis (horizontal direction) using a random scaling factor γ t U ( 0.7 , 1.3 ) . The transformation is centered, and empty regions are zero-padded. The same scaling is applied to all six RTM/DTM images. This simulates variations in gesture repetition cycles and overall motion speed.
RTM ( r , t ) = RTM ( r , γ t · t ) , DTM ( d , t ) = DTM ( d , γ t · t )
3.
DTM Vertical Stretch: The DTM is scaled along the Doppler axis (vertical direction) using a random scaling factor γ d U ( 0.7 , 1.3 ) . The transformation is centered, and empty regions are zero-padded. The same scaling is applied to all three receiving channels. This reflects instantaneous variations in gesture speed.
DTM ( d , t ) = DTM ( γ d · d , t )
Examples of the three augmentation methods are shown in Figure 10, illustrated using the RTM and DTM generated from Rx0 of a single sample. Each augmentation is applied to samples at the batch level during training and randomly re-sampled at each epoch, contributing to improved generalization performance of the model.

4. Experiments

4.1. Experimental Setup

This section describes the experimental environment used for training and evaluating the proposed Multi-EffNetV2 model. All experiments were conducted on a PC equipped with an Intel Core i9-13900KF CPU (24 cores, 32 threads), 128 GB DDR5 RAM, an NVIDIA GeForce RTX 4090 GPU (24 GB VRAM), and Ubuntu 20.04.6 LTS (WSL2). The software environment consisted of Python 3.10, TensorFlow 2.13, CUDA 11.8, and cuDNN 8.6. The dataset, comprising 2208 samples, was divided into training and testing sets at an 8:2 ratio using a stratified split, and evaluation was performed exclusively on the test set. To mitigate class imbalance, class weights were applied during training. All experiments were conducted with the same hyperparameter settings: the AdamW optimizer ( λ = 1 × 10 4 ) [50], an initial learning rate of 1 × 10 3 , a batch size of 32, and a total of 70 epochs. To ensure statistical reliability, model training and evaluation were repeated five times with different random seeds for each preprocessed dataset.

4.2. Training and Evaluation

This section evaluates the performance of the Adaptive Top-K Selection algorithm. To eliminate performance gains attributable to data augmentation or training optimizations, no additional strategies such as augmentation methods or learning rate schedulers were applied. Figure 11 shows the normalized confusion matrix of the test set obtained from the Multi-EffNetV2 model trained with the dataset preprocessed using Adaptive Top-K Summation, corresponding to the run that achieved the highest accuracy among five repeated experiments. The proposed method demonstrated excellent classification performance across all gesture classes, achieving the highest accuracy of 98.87% in the best run.
The evaluation accuracy was compared among datasets preprocessed using the methods in Figure 6 and those preprocessed with the proposed Adaptive Top-K Summation. Table 3 summarizes the highest, lowest, and average accuracies.
Conventional compression methods have clear drawbacks: method (a) dilutes salient peaks with background energy, while methods (b) and (c) risk information loss by retaining only local maxima. Method (d) achieved its best result at K = 10 with an average accuracy of 97.96%, but it remained limited by applying the same K uniformly to all vectors. In practice, finding such an empirically optimal K requires exhaustive experimentation and does not generalize well across different datasets or signal conditions. In contrast, the proposed Adaptive Top-K dynamically adjusts K using the Shannon entropy of each vector, preserving meaningful components within peak clusters, handling multi-peak distributions more robustly and ultimately achieving the best performance with an average accuracy of 98.60% and a highest accuracy of 98.87%.

4.3. Performance Optimization

In this section, we enhance the final performance of the Multi-EffNetV2 model by applying radar-specific data augmentation (Section 3.4) and a learning rate scheduler to the dataset preprocessed with Adaptive Top-K Summation. The scheduler reduced the learning rate by a factor of 10 at epoch 50, improving accuracy and facilitating loss convergence in the later training stage. Figure 12 illustrates the training curves of the best-performing run among five repetitions, showing that the model converged rapidly in the early epochs and stabilized after the learning rate decay. Figure 13 presents the normalized confusion matrix of the same run, which achieved the highest overall evaluation accuracy of 99.77%, demonstrating consistently high classification performance across all gesture classes, with notable improvements in the low-sample G n class.
Across the five repeated experiments, the proposed configuration achieved the highest evaluation accuracy of 99.77%, the lowest of 99.10%, and an average of 99.50%. From the best run, the macro-average accuracy across all gesture classes was 99.83%. Compared to the baseline reported in the dataset paper, where ResNet trained on a reduced subset (G0–G4, Gn) with Short-Time Fourier Transform (STFT) preprocessing achieved per-class accuracies ranging from 71.4% to 96.0% (Table 4 [35]), the proposed Multi-EffNetV2 substantially outperformed prior results. This highlights the effectiveness of the proposed preprocessing pipeline and demonstrates that, when combined with an optimized deep learning architecture, it can fully exploit the potential of the dataset.

5. Discussion and Conclusions

This study aimed to improve classification performance in FMCW radar environments by preserving gesture information despite coexistence with torso- and arm-related reflections. To address this challenge, the study proposed (i) an entropy-based Adaptive Top-K Selection algorithm to mitigate information loss and attenuation, (ii) a Multi-Stream EfficientNetV2 architecture to jointly exploit range and Doppler trajectories, and (iii) radar-specific data augmentation with a training optimization strategy to further enhance model performance. The proposed method, evaluated on the KIT radar gesture dataset, achieved the highest evaluation accuracy of 99.77% in the best run, with an average accuracy of 99.50% across five repeated experiments. From the best-performing run, the macro-average accuracy across gesture classes reached 99.83%, demonstrating its high effectiveness and notable improvements in the low-sample G n class.
Nevertheless, this study has several limitations. First, the entropy-based preprocessing requires additional computation for every frame in order to determine the adaptive K value, which increases the preprocessing overhead. Second, while the proposed Multi-Stream EfficientNetV2 achieved strong performance, it contains 12.38 million trainable parameters and requires 2.89 GFLOPs per inference, which is compact compared to state-of-the-art vision transformers or 3D CNNs but still higher than lightweight models typically deployed on resource-constrained embedded hardware. These factors indicate that further research is needed to design more efficient pipelines that reduce preprocessing cost and model complexity, thereby enabling seamless deployment on edge devices.
Moreover, the present work focused on single-user, isolated gesture recognition in general environments. Extending the approach to multi-user and multi-gesture scenarios, as well as handling continuous gesture streams, remains an open challenge. Future work will explore preprocessing algorithm acceleration together with compression strategies such as model lightweight design and quantization to minimize latency and memory usage. We also plan to investigate real-time interactive applications where gesture recognition is integrated into broader HMI systems. These directions will not only address efficiency and deployment concerns but also expand the applicability of radar-based gesture recognition to practical, real-world use cases.

Author Contributions

Conceptualization, J.J.; methodology, J.P.; software, J.P.; validation, J.P.; formal analysis, J.P.; investigation, J.P.; data curation, J.P.; writing—original draft preparation, J.P.; writing—review and editing, J.P. and J.J.; visualization, J.P.; supervision, J.J.; project administration, J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Gyeongsangbuk-do RISE (Regional Innovation System & Education) project (Specialized University unit).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The radar gesture dataset used in this study, “A Flexible Data Set for Radar-based Gesture Recognition with an FC-FMCW Radar,” is publicly available at https://doi.org/10.35097/hCNNxXlJBdFTSYse. No new data were created in this study.

Acknowledgments

We acknowledge the Karlsruhe Institute of Technology, Germany, for providing the publicly available radar gesture dataset utilized in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
1D FFTOne-Dimensional Fast Fourier Transform
2D CNNTwo-Dimensional Convolutional Neural Network
3D CNNThree-Dimensional Convolutional Neural Network
ADCAnalog-to-Digital Converter
BNBatch Normalization
CA-CFARCell Averaging Constant False Alarm Rate
CUTCell Under Test
DOADirection-of-Arrival
DTMDoppler-Time Map
GFLOPsGiga Floating-Point Operations
FMCWFrequency-Modulated Continuous-Wave
HCIHuman–Computer Interaction
HMIHuman–Machine Interface
IFIntermediate Frequency
IoTInternet of Things
LSTMLong Short-Term Memory
MTIMoving Target Indicator
MUSICMultiple Signal Classification
MVDRMinimum Variance Distortionless Response
RDIRange–Doppler Image
ReLURectified Linear Unit
RGBRed, Green, Blue
RNNRecurrent Neural Network
RTMRange-Time Map
STFTShort-Time Fourier Transform
UWBUltra-Wideband
XGBExtreme Gradient Boosting

References

  1. Kumar, S.; Tiwari, P.; Zymbler, M. Internet of Things Is a Revolutionary Approach for Future Technology Enhancement: A Review. J. Big Data 2019, 6, 111. [Google Scholar] [CrossRef]
  2. Alotaibi, B. A Survey on Industrial Internet of Things Security: Requirements, Attacks, AI-Based Solutions, and Edge Computing Opportunities. Sensors 2023, 23, 7470. [Google Scholar] [CrossRef] [PubMed]
  3. Ahmed, S.; Kallu, K.D.; Ahmed, S.; Cho, S.H. Hand Gestures Recognition Using Radar Sensors for Human-Computer-Interaction: A Review. Remote Sens. 2021, 13, 527. [Google Scholar] [CrossRef]
  4. Paravati, G.; Gatteschi, V. Human-Computer Interaction in Smart Environments. Sensors 2015, 15, 19487–19494. [Google Scholar] [CrossRef]
  5. Joseph, J.; D S, D. Hand Gesture Interface for Smart Operation Theatre Lighting. Int. J. Eng. Technol. 2018, 7, 20. [Google Scholar] [CrossRef]
  6. Dekker, B.; Jacobs, S.; Kossen, A.; Kruithof, M.; Huizing, A.; Geurts, M. Gesture Recognition with a Low Power FMCW Radar and a Deep Convolutional Neural Network. In Proceedings of the 2017 European Radar Conference (EURAD), Nuremberg, Germany, 11–13 October 2017; pp. 163–166. [Google Scholar] [CrossRef]
  7. Alexakis, G.; Panagiotakis, S.; Fragkakis, A.; Markakis, E.; Vassilakis, K. Control of Smart Home Operations Using Natural Language Processing, Voice Recognition and IoT Technologies in a Multi-Tier Architecture. Designs 2019, 3, 32. [Google Scholar] [CrossRef]
  8. Kröger, J.L.; Lutz, O.H.M.; Raschke, P. Privacy Implications of Voice and Speech Analysis – Information Disclosure by Inference. In Privacy and Identity Management. Data for Better Living: AI and Privacy; Friedewald, M., Önen, M., Lievens, E., Krenn, S., Fricker, S., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 576, pp. 242–258. [Google Scholar] [CrossRef]
  9. Yang, K.; Kim, M.; Jung, Y.; Lee, S. Hand Gesture Recognition Using FSK Radar Sensors. Sensors 2024, 24, 349. [Google Scholar] [CrossRef]
  10. Lin, P.; Li, C.; Chen, S.; Huangfu, J.; Yuan, W. Intelligent Gesture Recognition Based on Screen Reflectance Multi-Band Spectral Features. Sensors 2024, 24, 5519. [Google Scholar] [CrossRef]
  11. Shin, J.; Miah, A.S.M.; Kabir, M.H.; Rahim, M.A.; Al Shiam, A. A Methodological and Structural Review of Hand Gesture Recognition Across Diverse Data Modalities. IEEE Access 2024, 12, 142606–142639. [Google Scholar] [CrossRef]
  12. Foteinos, K.; Cani, J.; Linardakis, M.; Radoglou-Grammatikis, P.; Argyriou, V.; Sarigiannidis, P.; Varlamis, I.; Papadopoulos, G.T. Visual Hand Gesture Recognition with Deep Learning: A Comprehensive Review of Methods, Datasets, Challenges and Future Research Directions. arXiv 2025, arXiv:2507.04465. [Google Scholar] [CrossRef]
  13. Yu, J.; Qin, M.; Zhou, S. Dynamic Gesture Recognition Based on 2D Convolutional Neural Network and Feature Fusion. Sci. Rep. 2022, 12, 4345. [Google Scholar] [CrossRef] [PubMed]
  14. Xing, Z.; Meng, Z.; Zheng, G.; Ma, G.; Yang, L.; Guo, X.; Tan, L.; Jiang, Y.; Wu, H. Intelligent Rehabilitation in an Aging Population: Empowering Human-Machine Interaction for Hand Function Rehabilitation through 3D Deep Learning and Point Cloud. Front. Comput. Neurosci. 2025, 19, 1543643. [Google Scholar] [CrossRef] [PubMed]
  15. Chamorro, S.; Collier, J.; Grondin, F. Neural Network Based Lidar Gesture Recognition for Realtime Robot Teleoperation. In Proceedings of the 2021 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR), New York City, NY, USA, 25–27 October 2021; pp. 98–103. [Google Scholar] [CrossRef]
  16. Vandersteegen, M.; Reusen, W.; Beeck, K.V.; Goedeme, T. Low-Latency Hand Gesture Recognition with a Low Resolution Thermal Imager. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 440–449. [Google Scholar] [CrossRef]
  17. Stadelmayer, T. Preprocessing and Classification Techniques for FMCW Radar and Deep Learning Based Indoor Motion Classification. Ph.D. Thesis, Friedrich-Alexander-Universitaet Erlangen-Nuernberg, Erlangen, Germany, 2024. [Google Scholar]
  18. Qiu, X.; Liu, J.; Song, L.; Teng, H.; Zhang, J.; Wang, Z. A Survey of Gesture Recognition Using Frequency Modulated Continuous Wave Radar. J. Comput. Commun. 2024, 12, 115–134. [Google Scholar] [CrossRef]
  19. Suh, J.S.; Ryu, S.; Han, B.; Choi, J.; Kim, J.H.; Hong, S. 24 GHz FMCW Radar System for Real-Time Hand Gesture Recognition Using LSTM. In Proceedings of the 2018 Asia-Pacific Microwave Conference (APMC), Kyoto, Japan, 6–9 November 2018; pp. 860–862. [Google Scholar] [CrossRef]
  20. Ahmed, S.; Kim, W.; Park, J.; Cho, S.H. Radar-Based Air-Writing Gesture Recognition Using a Novel Multistream CNN Approach. IEEE Internet Things J. 2022, 9, 23869–23880. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Tian, Z.; Zhou, M. Latern: Dynamic Continuous Hand Gesture Recognition Using FMCW Radar Sensor. IEEE Sens. J. 2018, 18, 3278–3289. [Google Scholar] [CrossRef]
  22. Chmurski, M.; Mauro, G.; Santra, A.; Zubert, M.; Dagasan, G. Highly-Optimized Radar-Based Gesture Recognition System with Depthwise Expansion Module. Sensors 2021, 21, 7298. [Google Scholar] [CrossRef]
  23. Strobel, M.; Schoenfeldt, S.; Daugalas, J. Gesture Recognition for FMCW Radar on the Edge. In Proceedings of the 2024 IEEE Topical Conference on Wireless Sensors and Sensor Networks (WiSNeT), San Antonio, TX, USA, 21–24 January 2024; pp. 45–48. [Google Scholar] [CrossRef]
  24. Wang, S.; Song, J.; Lien, J.; Poupyrev, I.; Hilliges, O. Interacting with Soli: Exploring Fine-Grained Dynamic Gesture Recognition in the Radio-Frequency Spectrum. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology, Tokyo, Japan, 16–19 October 2016; pp. 851–860. [Google Scholar] [CrossRef]
  25. Choi, J.W.; Ryu, S.J.; Kim, J.H. Short-Range Radar Based Real-Time Hand Gesture Recognition Using LSTM Encoder. IEEE Access 2019, 7, 33610–33618. [Google Scholar] [CrossRef]
  26. Hayashi, E.; Lien, J.; Gillian, N.; Giusti, L.; Weber, D.; Yamanaka, J.; Bedal, L.; Poupyrev, I. RadarNet: Efficient Gesture Recognition Technique Utilizing a Miniature Radar Sensor. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; pp. 1–14. [Google Scholar] [CrossRef]
  27. Manitsaris, S.; Senteri, G.; Makrygiannis, D.; Glushkova, A. Human Movement Representation on Multivariate Time Series for Recognition of Professional Gestures and Forecasting Their Trajectories. Front. Robot. AI 2020, 7, 80. [Google Scholar] [CrossRef]
  28. Tiwari, B.; Gupta, S.H.; Balyan, V. Comparative Performance Exploration of Different Machine Learning and Deep Learning Algorithms for Classification of Hand Wrist Gestures. In Proceedings of the 2024 2nd International Conference on Disruptive Technologies (ICDT), Greater Noida, India, 15–16 March 2024; pp. 245–249. [Google Scholar] [CrossRef]
  29. Lien, J.; Gillian, N.; Karagozler, M.E.; Amihood, P.; Schwesig, C.; Olson, E.; Raja, H.; Poupyrev, I. Soli: Ubiquitous Gesture Sensing with Millimeter Wave Radar. ACM Trans. Graph. 2016, 35, 1–19. [Google Scholar] [CrossRef]
  30. Chudnikov, V.V.; Shakhtarin, B.I.; Bychkov, A.V.; Kazaryan, S.M. DOA Estimation in Radar Sensors with Colocated Antennas. In Proceedings of the 2020 Systems of Signal Synchronization, Generating and Processing in Telecommunications (SYNCHROINFO), Svetlogorsk, Russia, 1–3 July 2020; pp. 1–6. [Google Scholar] [CrossRef]
  31. Dzvonkovskaya, A.; Rohling, H. Software-Improved Range Resolution for Oceanographic HF FMCW Radar. In Proceedings of the 2013 14th International Radar Symposium (IRS), Dresden, Germany, 19–21 June 2013; Volume 1, pp. 411–416. [Google Scholar]
  32. Molchanov, P.; Gupta, S.; Kim, K.; Pulli, K. Short-Range FMCW Monopulse Radar for Hand-Gesture Sensing. In Proceedings of the 2015 IEEE Radar Conference (RadarCon), Arlington, VA, USA, 10–15 May 2015; pp. 1491–1496. [Google Scholar] [CrossRef]
  33. Ritchie, M.; Jones, A.; Brown, J.; Griffiths, H. Hand Gesture Classification Using 24 GHz FMCW Dual Polarised Radar. In Proceedings of the International Conference on Radar Systems (Radar 2017), Belfast, UK, 23–26 October 2017. [Google Scholar] [CrossRef]
  34. Ash, M.; Ritchie, M.; Chetty, K. On the Application of Digital Moving Target Indication Techniques to Short-Range FMCW Radar Data. IEEE Sens. J. 2018, 18, 4167–4175. [Google Scholar] [CrossRef]
  35. Antes, T.; Bekker, E.; Bhutani, A.; Zwick, T. A Flexible Data Set for Radar-Based Gesture Recognition. In Proceedings of the 2025 16th German Microwave Conference (GeMiC), Dresden, Germany, 17–19 March 2025; pp. 538–541. [Google Scholar] [CrossRef]
  36. Antes, T.; Bekker, E.; Bhutani, A.; Zwick, T. A Flexible Data Set for Radar-Based Gesture Recognition with an FC-FMCW Radar. 2024. Available online: https://publikationen.bibliothek.kit.edu/1000172790 (accessed on 2 October 2024).
  37. BGT60TR13C—60 GHz Radar Sensors for IoT|Infineon Technologies AG. Available online: https://www.infineon.com/part/BGT60TR13C (accessed on 26 August 2025).
  38. Stove, A.G. Linear FMCW radar techniques. In IEE Proceedings F (Radar and Signal Processing); IET: London, UK, 1992; Volume 139, pp. 343–350. [Google Scholar]
  39. Enggar, F.D.; Muthiah, A.M.; Winarko, O.D.; Samijayani, O.N.; Rahmatia, S. Performance Comparison of Various Windowing On FMCW Radar Signal Processing. In Proceedings of the 2016 International Symposium on Electronics and Smart Devices (ISESD), Bandung, Indonesia, 29–30 November 2016; pp. 326–330. [Google Scholar] [CrossRef]
  40. Nguyen, M.Q.; Feger, R.; Wagner, T.; Stelzer, A. Analysis of 2D CA-CFAR for DDMA FMCW MIMO Radar. In Proceedings of the 2023 20th European Radar Conference (EuRAD), Berlin, Germany, 20–22 September 2023; pp. 423–426. [Google Scholar] [CrossRef]
  41. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  42. Huang, M.L.; Liao, Y.C. Stacking Ensemble and ECA-EfficientNetV2 Convolutional Neural Networks on Classification of Multiple Chest Diseases Including COVID-19. Acad. Radiol. 2023, 30, 1915–1935. [Google Scholar] [CrossRef]
  43. Hoang, L.; Lee, S.H.; Lee, E.J.; Kwon, K.R. Multiclass Skin Lesion Classification Using a Novel Lightweight Deep Learning Framework for Smart Healthcare. Appl. Sci. 2022, 12, 2677. [Google Scholar] [CrossRef]
  44. Zhao, Z.; Bakar, E.B.A.; Razak, N.B.A.; Akhtar, M.N. Corrosion Image Classification Method Based on EfficientNetV2. Heliyon 2024, 10, e36754. [Google Scholar] [CrossRef]
  45. Kim, B.; Seo, S. EfficientNetV2-based Dynamic Gesture Recognition Using Transformed Scalogram from Triaxial Acceleration Signal. J. Comput. Des. Eng. 2023, 10, 1694–1706. [Google Scholar] [CrossRef]
  46. Hartanto, J.; Wijaya, S.M.; Anderies; Chowanda, A. Performance Evaluation of EfficientNetB0, EfficientNetV2, and MobileNetV3 for American Sign Language Classification. In Proceedings of the 2023 8th International Conference on Electrical, Electronics and Information Engineering (ICEEIE), Malang City, Indonesia, 28–29 September 2023; pp. 1–6. [Google Scholar] [CrossRef]
  47. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar]
  48. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10096–10106. [Google Scholar]
  49. Kern, N.; Waldschmidt, C. Data Augmentation in Time and Doppler Frequency Domain for Radar-based Gesture Recognition. In Proceedings of the 2021 18th European Radar Conference (EuRAD), London, UK, 5–7 April 2022; pp. 33–36. [Google Scholar] [CrossRef]
  50. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2019, arXiv:1711.05101. [Google Scholar] [CrossRef]
Figure 1. Dynamic hand gestures from the KIT radar gesture dataset [36]. In this study, six gestures marked with “O” in the “Use” column were selected for classification.
Figure 1. Dynamic hand gestures from the KIT radar gesture dataset [36]. In this study, six gestures marked with “O” in the “Use” column were selected for classification.
Sensors 25 06324 g001
Figure 2. BGT60TR13C radar demo board with one transmitting (Tx) and three receiving (Rx) antennas [37].
Figure 2. BGT60TR13C radar demo board with one transmitting (Tx) and three receiving (Rx) antennas [37].
Sensors 25 06324 g002
Figure 3. Measurement environment of the KIT radar gesture datase (Setup 3). This figure shows one representative configuration as an example (reproduced from [36]).
Figure 3. Measurement environment of the KIT radar gesture datase (Setup 3). This figure shows one representative configuration as an example (reproduced from [36]).
Sensors 25 06324 g003
Figure 4. Window-based 2D CA-CFAR kernel example.
Figure 4. Window-based 2D CA-CFAR kernel example.
Sensors 25 06324 g004
Figure 5. Example of a Range–Doppler Image (RDI) without removing body-related reflections. Multiple peak clusters are observed, corresponding to reflections from both the hand gesture and the torso. The highlighted regions indicate the localized energy distributions of gesture and torso components.
Figure 5. Example of a Range–Doppler Image (RDI) without removing body-related reflections. Multiple peak clusters are observed, corresponding to reflections from both the hand gesture and the torso. The highlighted regions indicate the localized energy distributions of gesture and torso components.
Sensors 25 06324 g005
Figure 6. Illustration of four compression methods for reducing RDIs into 1D vectors: (a) summation across rows or columns; (b) maximum extraction from each row or column; (c) slicing at the maximum peak position; (d) Top-K summation after sorting by magnitude.
Figure 6. Illustration of four compression methods for reducing RDIs into 1D vectors: (a) summation across rows or columns; (b) maximum extraction from each row or column; (c) slicing at the maximum peak position; (d) Top-K summation after sorting by magnitude.
Sensors 25 06324 g006
Figure 7. Example of entropy-based K determination for row vectors in a reduced RDI sample.
Figure 7. Example of entropy-based K determination for row vectors in a reduced RDI sample.
Sensors 25 06324 g007
Figure 8. Examples of RTM and DTM generated from Rx0 for each gesture class.
Figure 8. Examples of RTM and DTM generated from Rx0 for each gesture class.
Sensors 25 06324 g008
Figure 9. Framework of the proposed Multi-EffNetV2.
Figure 9. Framework of the proposed Multi-EffNetV2.
Sensors 25 06324 g009
Figure 10. Examples of the proposed radar-specific data augmentation methods applied to RTM and DTM from Rx0 of a single sample. The first column shows the original RTM and DTM, while the subsequent columns illustrate (1) RTM vertical shift, (2) RTM/DTM horizontal stretch, and (3) DTM vertical stretch. Each augmentation simulates variations in range, motion cycle, or gesture speed, thereby enhancing the diversity of the training data.
Figure 10. Examples of the proposed radar-specific data augmentation methods applied to RTM and DTM from Rx0 of a single sample. The first column shows the original RTM and DTM, while the subsequent columns illustrate (1) RTM vertical shift, (2) RTM/DTM horizontal stretch, and (3) DTM vertical stretch. Each augmentation simulates variations in range, motion cycle, or gesture speed, thereby enhancing the diversity of the training data.
Sensors 25 06324 g010
Figure 11. Normalized confusion matrix of the test set for the proposed Adaptive Top-K Summation with the Multi-EffNetV2 model, corresponding to the run with the highest accuracy among five repeated experiments.
Figure 11. Normalized confusion matrix of the test set for the proposed Adaptive Top-K Summation with the Multi-EffNetV2 model, corresponding to the run with the highest accuracy among five repeated experiments.
Sensors 25 06324 g011
Figure 12. Training curves of the Multi-EffNetV2 model with Radar-Specific Data Augmentation and learning rate scheduler applied. Results correspond to the best-performing run among five repetitions: (a) accuracy curve and (b) loss curve. The learning rate was reduced by a factor of 10 at epoch 50, leading to stabilized convergence and accuracy exceeding 99%.
Figure 12. Training curves of the Multi-EffNetV2 model with Radar-Specific Data Augmentation and learning rate scheduler applied. Results correspond to the best-performing run among five repetitions: (a) accuracy curve and (b) loss curve. The learning rate was reduced by a factor of 10 at epoch 50, leading to stabilized convergence and accuracy exceeding 99%.
Sensors 25 06324 g012
Figure 13. Normalized confusion matrix for the test set obtained from the best-performing run among five repetitions with Adaptive Top-K Summation, Radar-Specific Data Augmentation, and learning rate scheduler applied. The proposed configuration achieved the highest evaluation accuracy of 99.77%.
Figure 13. Normalized confusion matrix for the test set obtained from the best-performing run among five repetitions with Adaptive Top-K Summation, Radar-Specific Data Augmentation, and learning rate scheduler applied. The proposed configuration achieved the highest evaluation accuracy of 99.77%.
Sensors 25 06324 g013
Table 1. Radar measurement parameters of the BGT60TR13C sensor used in the KIT radar gesture dataset [36].
Table 1. Radar measurement parameters of the BGT60TR13C sensor used in the KIT radar gesture dataset [36].
ParameterValue
Frequency band f min f max 57.5–64.5 GHz
Sampling rate f S 1 MHz
Chirp repetition time T C 300 μ s
Number of samples N S 128
Number of chirps N C 600
Range resolution Δ R 2.1 cm
Maximum unambiguous range R max 1.4 m
Velocity resolution Δ v 1.5 cm/s
Maximum unambiguous velocity v max 4.3 m/s
Table 2. Adaptive Top-K Selection Algorithm.
Table 2. Adaptive Top-K Selection Algorithm.
StepDescription
InputVector v R n , K min , K max
OutputAdaptive K value
Parameter ϵ : a small positive constant to prevent division by zero
(1)Compute p i = v i v i + ϵ for i = 1 , , n .
(2)Compute entropy: H = i = 1 n p i log p i .
(3)Normalize entropy: H norm = H log n .
(4)Compute K = ( 1 H norm ) ( K max K min ) + K min .
end
Table 3. Comparison of evaluation accuracy among different compression methods over five repeated experiments. Highest, lowest, and average accuracies are reported: (a) summation across rows or columns; (b) maximum extraction from each row or column; (c) slicing at the maximum peak position; (d) Top-K summation with fixed values of K { 5 , 10 , 15 , 20 , 25 } ; (e) proposed Adaptive Top-K Summation.
Table 3. Comparison of evaluation accuracy among different compression methods over five repeated experiments. Highest, lowest, and average accuracies are reported: (a) summation across rows or columns; (b) maximum extraction from each row or column; (c) slicing at the maximum peak position; (d) Top-K summation with fixed values of K { 5 , 10 , 15 , 20 , 25 } ; (e) proposed Adaptive Top-K Summation.
Evaluation
Accuracy [%]
Compression Methods
Total
Sum
Maximum
Extraction
Peak
Slicing
Top-5
Sum
Top-10
Sum
Top-15
Sum
Top-20
Sum
Top-25
Sum
Adaptive
Top-K Sum
(a)(b)(c) (d) (e)
Highest96.3897.0692.9997.7498.1997.7497.5196.6198.87
Lowest95.7096.3890.2796.8397.7496.1596.1595.7098.19
Average96.0696.6591.3697.2997.9696.9796.7996.2098.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, J.; Jeong, J. Radar-Based Gesture Recognition Using Adaptive Top-K Selection and Multi-Stream CNNs. Sensors 2025, 25, 6324. https://doi.org/10.3390/s25206324

AMA Style

Park J, Jeong J. Radar-Based Gesture Recognition Using Adaptive Top-K Selection and Multi-Stream CNNs. Sensors. 2025; 25(20):6324. https://doi.org/10.3390/s25206324

Chicago/Turabian Style

Park, Jiseop, and Jaejin Jeong. 2025. "Radar-Based Gesture Recognition Using Adaptive Top-K Selection and Multi-Stream CNNs" Sensors 25, no. 20: 6324. https://doi.org/10.3390/s25206324

APA Style

Park, J., & Jeong, J. (2025). Radar-Based Gesture Recognition Using Adaptive Top-K Selection and Multi-Stream CNNs. Sensors, 25(20), 6324. https://doi.org/10.3390/s25206324

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop