Next Article in Journal
Squat Kinematics Analysis Using Vicon and Affordable Motion-Capture Solutions
Previous Article in Journal
Enhanced Channel Estimation for RIS-Assisted OTFS Systems by Introducing ELM Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Markerless Motion Classification Using Radar

1
Department of Physical Education, Graduate School, Pukyong National University, Busan 48513, Republic of Korea
2
Industry-Academia Cooperation Foundation, Pukyong National University, Busan 48513, Republic of Korea
3
Department of Maritime ICT & Mobility Research, Korea Institute of Ocean Science & Technology, 385, Haeyang-ro, Yeongdo-gu, Busan 49111, Republic of Korea
4
Department of Marine Sports, Pukyong National University, Busan 48513, Republic of Korea
5
Department of Electronic Engineering, Pukyong National University, Busan 48513, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(11), 3293; https://doi.org/10.3390/s25113293
Submission received: 26 March 2025 / Revised: 8 May 2025 / Accepted: 22 May 2025 / Published: 23 May 2025
(This article belongs to the Section Radar Sensors)

Abstract

:

Highlights

What are the main findings?
  • A novel markerless motion-classification method that uses radar micro-Doppler features achieves nearly 100% classification accuracy with a compact feature set.
  • The method uses cross-correlation between radar signals and basis functions derived from 3D motion-capture data, then applies PCA dimensionality reduction and nearest-neighbor classification.
What is the implication of the main finding?
  • The proposed method enables accurate classification of human motion without markers, and thereby ensures privacy and reduces the cost compared to traditional motion-capture systems.
  • The approach is accurate across varying SNR conditions, and is therefore suitable for real-world applications such as healthcare monitoring and sports analytics.

Abstract

This study proposes a novel method that uses radar for markerless motion classification by using effective features derived from micro-Doppler signals. The training phase uses three-dimensional marker coordinates captured by a motion-capture system to construct basis functions, which enable modeling of micro-motions of the human body. During the testing phase, motion classification is performed without markers, relying solely on radar signals. The feature vectors are generated by applying cross-correlation between the received radar signal and the basis functions, then compressed using principal component analysis, and classified using a simple nearest-neighbor algorithm. The proposed method achieves nearly 100% classification accuracy with a compact feature set and is accurate even at high signal-to-noise ratios. Experimental results demonstrate that to optimize training data and increase computational efficiency, the sampling duration and sampling interval must be set appropriately.

1. Introduction

Motion-capture systems for humans [1] are used in various domains, including healthcare [2] and sports [3]. Typically, motion has been captured using three-dimensional (3D) camera systems that track the coordinates of markers placed on the body [4]. However, markers interfere with natural movement, and their placement is often constrained [5]. Consequently, markerless motion capture is being evaluated. Nevertheless, they still use 3D camera systems, which are expensive and bulky [6]. Use of radar systems has been suggested as a method to overcome this problem [7].
Radar signals that reflect from targets can be measured without markers. Radar systems are significantly more affordable (a few hundred dollars) than 3D camera systems (tens of thousands of dollars). Additionally, radar’s compact size allows for flexible installation, whereas 3D motion-capture systems use fixed cameras and are therefore not easily portable. Furthermore, radar signals are less susceptible than 3D cameras to interference from optical signals, so radar systems can collect and process data faster than 3D camera systems. Radar can also observe under various conditions of weather and illumination, and by using wideband signals, can provide high-resolution data [8,9].
Accurate classification and analysis of human motion require precise modeling of micro-motions (MMs) of the human body, such as vibrations and rotations, with features derived from modeled motions serving as the basis. However, a human body is composed of several primary scatterers, each with a distinct location, speed of movement, and angular range of movement, and these traits vary significantly among individuals. Additionally, the MMs of each scatterer consists of multiple sine and cosine components with varying amplitudes, rotational velocities, and initial phases, so mathematical modeling of arbitrary MMs can be complicated. Therefore, analysis of electromagnetic data to model the MMs of the human body has not been sufficiently accurate.
The micro-Doppler (MD) effect [10] in radar technology refers to time-varying Doppler frequencies that are generated by MMs, and that represent these motions in the time–frequency domain. Analysis of the MD effect has clear applications to motion analysis. When radar observes a person performing specific movements, the motions of body parts such as the arms and legs produce corresponding MD signals. These signals are highly effective for motion recognition [11]. The radar spectrogram is a motion-analysis technique that depicts the power distribution of target returns in both the time and frequency domains [12].
Various pose-recognition techniques have been proposed; they primarily use neural networks to analyze spectrograms. Some of these classifiers can operate accurately using only limited training data [13,14,15]. To increase classification accuracy, data-augmentation methods can be used to expand training datasets [16,17,18], and transfer-learning approaches have also been explored [19,20,21,22]. However, methods that use neural networks can be accurate for predefined movements, but as the number of actions increases, the training phase requires substantial computational resources and time, and error rates often rise, so such approaches are not practical for use by simple systems.
Several methods that use empirical features in the time–frequency domain for specific MMs have been proposed for application to arbitrary motions, and in a limited capacity [23,24]. Techniques that derive the micro-Doppler image envelope of movements and compress it using the principal component analysis (PCA) for classification [25] are prone to noise interference and often struggle with envelope detection. For use when signal interference occurs, a method that uses empirical mode decomposition (EMD) to remove interfering signals has been proposed [26], but the method can yield inaccurate results if EMD fails to correctly represent the micro-Doppler features. Furthermore, the subsequent classification step requires a complex neural network structure.
This paper proposes an efficient method that uses ‘effective features’ to classify individual motions by using 3D marker coordinates obtained by a camera system only during the training phase, then classifies test data without the need for markers. The proposed features are derived from the MMs of the human body, and enable accurate classification using a minimal number of features. In experiments conducted using a continuous wave (CW) radar operating at 5.8 GHz, the proposed method achieved nearly 100% classification accuracy with a highly compact feature set and a simple classifier, and thereby demonstrates efficiency and effectiveness. The remainder of this paper is organized as follows. Section 2 introduces the signal model and the proposed method. Section 3 outlines the experimental conditions, while Section 4 presents the classification results and performance analysis. Finally, Section 5 concludes the paper and discusses potential directions for future work.

2. Signal Modeling and Proposed Method

2.1. Overview

The method constructs basis functions for MD features by using marker coordinates, and generates feature vectors by analyzing cross-correlation between the received radar signal and the basis functions. These feature vectors are then compressed using principal components analysis (PCA), and the reduced feature set is classified using a simple nearest-neighbor classifier. This approach eliminates the need for empirical analysis of motion because the basis functions are constructed using accurate marker data, so the features can clearly represent specific body parts.

2.2. Signal Modeling and the Proposed Feature

We assume a CW radar that continuously transmits and receives unmodulated carrier signals during the entire detection process. Although frequency-modulated CW (FMCW) radar offers additional information such as target range and angle, along with an increased signal-to-clutter ratio, we did not consider FMCW in this paper, because our primary focus is on developing an efficient recognition method that utilizes the frequency modulation induced by the body’s micro-motions. The radar signal transmitted at time t with a frequency f0 is expressed as
s t = e x p j 2 π f 0 t ,
where j = 1 . When the transmitted signal is reflected from a target located at r(t), the received signal sr(t) is delayed as
s r t = e x p j 2 π f 0 t t d = e x p j 2 π f 0 t 2 r t c ,
where c = 3 × 108 m/s represents the speed of light in a vacuum. The received signal Sr(t) is down-converted to the baseband as
s b t = s r t s t ) * = exp j 2 π f 0 t t d exp 2 π j f 0 t = exp ( j 2 π f 0 t d = exp j 2 π λ r t ,
where λ = c/f0 is the wavelength.
When a target consists of Nc scatterers, the received baseband signal is the sum of the reflected signals:
s T ( t ) = i = 1 N c s i ( t ) = i = 1 N c A i exp j 4 π r i ( t ) λ ,
where Ai is the amplitude of scatterer i, and ri(t) is its distance. After the translational motion of the body is removed, ri(t) represents the MMs and can be expressed as
r i ( t ) = j = 1 N i c α i , j cos ( ω i , j c t + ϕ i , j c ) + j = 1 N i s β i , j sin ( ω i , j s t + ϕ i , j s ) ,
where N i c represents the number of cosine components in scatterer i, which have amplitude α i , j , rotational velocity ω i , j c , and initial phase ϕ i , j c ; and N i s represents the number of sine components in scatterer i,, which have amplitude β i , j , rotational velocity ω i , j s , and initial phase ϕ i , j s .
After Ai of the target is fully extracted for 1 ≤ iNc, the MMs can be effectively recognized using a 1 × Nc feature vector and a simple classifier. For a scatterer k, Ak can be obtained by calculating a simple inner product between s(t) and a basis signal that eliminates the exponential component of sk(t) in (4). Therefore, assuming that rk(t) of a scatter k is completely known, the basis function to extract Ak can be modeled as
b k ( t ) = exp j 4 π r k ( t ) λ ,
and Ak can be obtained by using the inner product between s(t) and bk(t), which is
s T ( t ) , b k ( t ) = 1 T 0 T s ( t ) exp j 4 π r k ( t ) λ d t = A k T 0 T 1 d t + 1 T i = 1 , i k N c A i 0 T exp j 4 π r k ( t ) r i ( t ) λ d t = A k + R s ( t ) ,
where T is the observation duration, and Rs(t) represents the inner product between the remaining scatterers and scatterer k.
Using (5), Rs(t) can be expressed as:
R s t = 1 T i = 1 , i k N c A i 0 T exp 4 π j λ j = 1 N i c α i , j cos ω i , j c t + ϕ i , j c α k , j cos ω k , j c t + ϕ k , j c + j = 1 N i s β i , j sin ω i , j s t + ϕ i , j s β k , j sin ω k , j s t + ϕ k , j s d t = 1 T i = 1 , i k N c A i 0 T exp j M ( t ) d t = 1 T i = 1 , i k N c A i 0 T cos M ( t ) + j sin M ( t ) d t ,
where
M ( t ) = 4 π λ j = 1 N i c α i , j cos ω i , j c t + ϕ i , j c α k , j cos ω k , j c t + ϕ k , j c + j = 1 N i s β i , j sin ω i , j s t + ϕ i , j s β k , j sin ω k , j s t + ϕ k , j s .
M(t) is the phase of the complex exponential function, and obtaining the analytical integration of exp(jM(t)) is challenging because the parameters of the cosine and sine functions are generally unknown. However, for a periodic MMs and an integration duration T that is sufficiently longer than the period of the MM, then M(t) is also periodic. Consequently, the absolute value of the integration of exp(jM(t)) becomes much smaller than T, because the integral of a sinusoidal function over a complete period equals 0. Therefore, we can conclude that A k R s t .
To demonstrate A k R s t , a simulation was conducted. Assuming f0 = 5.8 GHz, T = 3 s, sampling interval dt = 0.001 s, and four scatterers with equal amplitudes = 2, which have MMs that are each a sum of three cosine or sine components randomly selected, A 1 and R s ( t ) were compared 100 times using random MMs parameters. The frequencies were chosen randomly between 0 and 2 Hz, the amplitudes between 0 and 4π/λ, and the initial phases between 0 and 2π. All cases yielded A 1 in (8) equal to 2, and R s t remained close to 0 due to the periodic nature of the sinusoidal function (Figure 1).
Therefore, considering the difference in the initial phase, Ak can be determined by finding the maximum inner product as
A k = max s ( t ) , C S b k ( t ) , τ ,   for   0 τ T ,
where C S b k ( t ) , τ is the signal of bk(t), after a circular shift by τ.
Ak is found by identifying the maximum cross-correlation between s(t) and bk(t), as
A k = max I F T F T { s ( t ) } F T { b k ( t ) } ,
where IFT and FT represent, respectively, the inverse Fourier transform and the Fourier transform. In cases where the amplitude of scatterer k is small or a dominant scatterer exists, Ak in (11) may not represent the exact amplitude of scatterer k. However, Ak can still serve as an important feature because it quantifies the amount of MMs rk(t) that is present in bk(t).

2.3. Limitations in Determining Scatterer Amplitude and the Proposed Method

One- and two-dimensional methods that use spectral-estimation theory have been proposed to extract the amplitude and location of scatterers [27,28]. However, these methods assume that targets remain stationary while being observed, and are therefore unsuitable for estimating scatterers involved in MM. Additionally, selecting the dominant scatterer of the human body for classifying MMs is challenging, because the electromagnetic characteristics of individuals vary depending on the body’s composition (e.g., fat, muscle, bone). Although methods for modeling the human body have been suggested, they are often limited in their applications [29]. Moreover, modeling rk(t) becomes highly complex for arbitrary motions. Although some existing studies describe simple MMs such as walking or crawling [10], these approaches fail to capture the arbitrary motions of individuals. Therefore, extracting Ak using (10) is a difficult task.
To address the limitations of existing methods, we propose utilizing marker data obtained by using a 3D motion-capture system only during the training phase. The use of a camera during the training phase does not invade privacy, because training datasets are generally constructed using individuals who have consented to being observed. This system, which is widely used for kinematic purposes, provides accurate 3D coordinates of body parts across sagittal, frontal, and transverse planes by using markers attached to the body, and more than two cameras. Therefore, employing this system effectively resolves the challenges in modeling MMs. The proposed approach builds on a previous simulation study [30] that demonstrated the feasibility of classification using marker coordinates and modeled radar signals with images in the time-frequency (TF) domain, despite the significant computational burden. This paper presents an improved method that eliminates the need for spectrograms, compresses the features further, and demonstrates effectiveness using measured data.
Using Nc markers, Nc basis functions can be derived using (6), and the following feature can be obtained using (11):
f = [ f 1 f 2 f 3 f N c ] T .
For P motions, each with Q feature vectors, the following training database is constructed:
F = [ f 1 f 2 f 3 f N t r ] = f 11 f 12 f 1 N t r f 21 f 22 f 2 N t r f N c 1 f N c 2 f N c N t r ,
where Ntr = P × Q represents the total number of training vectors, and F is an Nc × Ntr matrix.
Scatterers that generate large-magnitude signals may produce large elements in (11) and significantly influence the classification results, so each element should be normalized to values between zero and one by using the maximum and minimum values as [31]:
F ¯ = [ f ¯ 1 f ¯ 2 f ¯ 3 f ¯ N t r ] ,
where
f ¯ k = f ¯ k 1 f ¯ k 2 f ¯ k 3 . . . f ¯ k N c T ,
and
f ¯ k j = f k j f j , m i n f j , m a x f j , m i n   , for   1 j N c
where fj,min and fj,max are, respectively, the minimum and maximum values of the j-th feature in the training database.
In the test phase, only the radar signal is measured and used for classification, meaning the location for acquiring data is not constrained by the 3D motion system’s position. The feature vector is obtained using (12) with the measured radar signal, and the marker data rk(t) stored in the training database. The test vector is normalized using (16) to adjust the range of its elements.

2.4. Data Compression Using Principal Component Analysis and Classification Using a Simple Classifier

The exact number of scatterers in the human body cannot be precisely stated, so Nc should be sufficiently large to encompass the MMs information of the entire body. However, setting Nc excessively high can lead to redundant information, increased memory requirements for storing feature vectors, and excessive computational costs during classification, without increasing recognition accuracy. To address this problem, PCA was used [25]. This method is widely used for data compression while preserving discriminative information. The transformation matrix in PCA is designed to project the feature vector into a subspace spanned by the eigenvectors that correspond to the largest eigenvalues of the covariance matrix. This process extracts a representative of effective features while retaining most of the intrinsic information.
For PCA, the unbiased covariance matrix is calculated using the normalized feature vectors in the training database as
C o v = 1 N t r 1 i = 1 N t r f ¯ i m f ¯ i m T ,
where
m = 1 N t r 1 i = 1 N t r f ¯ i .
By the eigenvalue decomposition,
C o v = V Λ V T ,
where V is the matrix composed of the eigenvectors of Cov as its columns, and Λ is a diagonal matrix that contains the corresponding eigenvalues.
The transformation matrix P is then constructed by selecting the eigenvectors that correspond to the largest eigenvalues in descending order:
P = v 1 v 2 v d ,
where d is the number of largest eigenvalues, and v1vd are the eigenvectors that correspond to the largest d eigenvalues. F ¯ in (14) is then transformed by
X = P T F ¯ = [ x 1 x 2 x 3 . . . x N t r ] ,   x k = P T v k ,
where the resultant feature vector xk is d-dimensional, i.e., significantly smaller than Nc.
xk provides distinctive features for each MM, so a simple nearest neighbor classifier is used. This classifier uses the Euclidean norm for classification:
s ^ = m i n s g s x t e ,
with
g s x t e = x t e x s ,
where x t e is a test vector, and x s is a training vector for the s-th motion. The observation includes P MMs, so s ^ ranges from one to P.

2.5. Overall Procedure

The overall procedure proposed in this paper is summarized as follows (Figure 2):
  • Data collection: For each of the P MMs, collect Q radar signals and Q × Nc marker coordinates;
  • Signal construction: Using the marker coordinates, construct bk(t) as defined in (6);
  • Feature vector construction: For each radar signal, construct an Nc × 1 feature vector using the maximum correlation, as described in (11) and formatted in (12);
  • Training database construction: Construct the training database F as defined in (13);
  • Normalization: Normalize the feature vector in F using the maximum and minimum feature values, as described in (15);
  • Compression: Compress the normalized feature vector using PCA;
  • Test motion processing: For any randomly selected test motion, repeat steps (1)–(3) and (5)–(6). In this step, the marker coordinates from the training database are used to construct the test feature; (11) is applied to the measured test signal;
  • Classification: Classify the motion using a simple nearest neighbor classifier.

3. Experimental Conditions

3.1. Participants

The experimental subjects were six adult males and four adult females (Table 1), none of whom had injured their upper or lower extremities in the previous six months. All participants were fully informed about the purpose of the research, experimental procedures, and precautions, and data collection was conducted only after obtaining their consent to participate. The experiment was approved by the Institutional Review Board (IRB # PKNU 2024-12-002) of the university.

3.2. Motions and Camera System to Extract Marker Information

Five representative motions were used for classification: squat, lunge, front kick, front lateral raise, and arm curl. To obtain position data for the markers, eight optical motion-capture cameras (Miqus M5, Qualisys AB, Gothenburg, Sweden) with a sampling rate of 250 Hz were used. Calibration was performed using the non-linear transformation method to define 3D spatial coordinates. Each participant had a total of 30 reflective markers (12 mm in size) attached to both the upper and lower body, particularly in areas where MMs occur frequently (Figure 3). Additionally, markers were attached to the back of the body to account for penetration of electromagnetic waves. The position data of the markers were obtained using Qualisys Track Manager; Qualisys, Gothenburg, Sweden software (QTM 2024.3), and the radar was modeled using MATLAB R2023b.

3.3. Radar and Classification Parameters

In the measurement, the radar signal and the marker coordinates were obtained simultaneously for Tend = 4 s. A CW radar was used, positioned 1.5 ± 0.5 m along the y-axis from the subject. The distance between the radar and the body parts was determined using the markers attached to both the radar and the participant. The radar coordinate was set to (0, 0, 0) by subtracting the radar’s marker coordinates from the participant’s marker coordinates. The center frequency of the CW radar was 5.8 GHz, corresponding to a wavelength λ = 0.0517 m, and the sampling rate was set to 5 kHz (Table 2). For the experiments, we used MATLAB R2023 running on Windows 11 operated by AMD Ryzen 9 processor.
Simultaneously, the camera system captured marker coordinates to construct the basis function (Equation (6)). To create the training dataset for each movement, the total time-domain signal of duration Tend was segmented into signals of length Tsamp = 0.3 × Tend = 1.2 s. These segments were clipped at uniform intervals of dttr = 0.02 s, producing Ntr = (TendTsamp)/dttr = 140 training signals per participant (Figure 4). For each clipped signal, a feature vector was generated using (11). Each feature was normalized as per (16), and dimensionality reduction was performed using PCA, reducing the vector to d = 4 dimensions as specified in (21). Classification was conducted using 50 randomly clipped test signals per motion, each of the same length Tsamp. To evaluate the effect of various parameters on classification accuracy, additional experiments were performed at various signal-to-noise ratios (SNRs) by introducing additive white Gaussian noise (AWGN) to the test data (Table 2). The classification accuracy Pc was defined as
P c = N c c N s
where Ncc is the number of the correct classifications and Ns is that of the test samples.

4. Classification Results

4.1. Analysis of MMs for Each Motion

A comparison of the TF images derived from the simulated data using marker coordinates and those from the measured data revealed both similarities and differences (Figure 5 and Figure 6). Marker coordinates and radar signals were obtained simultaneously, so the time periods and Doppler-frequency magnitudes in the simulated data were strongly similar to those in the measured data. Consequently, the proposed feature in (11) produces high values when the modeled signal aligns closely with the measured one, and low values when a mismatch occurs.
The observed differences arise from the electromagnetic scattering characteristics. In the simulation, markers were assumed to be omnidirectional scatterers with consistent amplitude, and to remain visible throughout the simulation period and provide precise Doppler frequencies. However, in reality, the scatterer’s position is unknown and varies with the aspect angle, so the signal amplitude changes significantly during the observation period. As a result, (11) may yield suboptimal results when a particular motion produces weak or very strong scatterers.
The unique characteristics of each exercise were well-distinguished in a three-dimensional feature space (feature 1, feature 2, feature 3) when d = 3 (Figure 7). Features of front lateral raise and arm curl appear close or partially overlapping due to their reliance on arm movement alone, so the two movements have similar Doppler characteristics (Figure 7b–d). This overlap highlights the necessity for densely sampled training data to increase classification accuracy for these motions. On the feature 1–2 plane (Figure 7b), squat and lunge features are close, so they may be misclassified. However, in the feature 2–3 and 1–3 planes (Figure 7c,d), these motions are distinguishable due to their separation in these dimensions. Similarly, front kick features are not well-separated in the feature 2–3 plane, but are distinguishable in the feature 1–2 and 1–3 planes, in which their unique characteristics become evident. These observations suggest that use of the three-dimensional feature space has the potential to achieve high classification accuracy by using the distinct separation of features across different planes.

4.2. Classification Result for Various SNRs and Analysis on the Effect of PCA Dimensions

To evaluate classification accuracy in relation to the PCA dimensionality of training data and the SNR of test data, we assessed accuracy by varying PCA dimensions from 3 to 6 and SNR values from 0 dB to 20 dB in increments of 5 dB. The noise inherent in the measured data was not considered when adding AWGN, so the actual classification SNR was lower than the assumed SNR. Default radar and classification parameters (Table 2) were applied. Additionally, Pc for each motion was analyzed in detail.
The proposed method demonstrated consistently high Pc, even with low PCA dimensions, and increasing the dimensionality increased the accuracy. Similarly, higher SNR values corresponded to better classification accuracy (Figure 8 and Figure 9). Among the motions, the classification accuracy for arm curl was much lower for SNRs ≤ 5 dB than at other SNRs, due to its similarity to front lateral raise. The proximity of their features resulted in some arm curls being misclassified as front lateral raises (Figure 7b–d). Overall, the results indicate that achieving Pcs > 99.0% requires an SNR ≥ 10 dB and d ≥ 4. Given the sample size of the measured data = 5 kHz (sample rate) × 1.2 s (Tsamp) = 6000, d = 4 corresponds to only 0.07% of the measured data. This result demonstrates the remarkable efficiency of the proposed feature-extraction method.

4.3. Analysis of the Classification Accuracy of Each Motion

The classification accuracy of each motion was analyzed using the confusion matrix obtained at SNRs of 0, 5, and 10 dB, with d = 6 as the default parameter (Figure 10). The classification accuracy improved proportionally with the SNR, with a classification rate of 100% achieved at an SNR of 10 dB. This demonstrates that an SNR of 10 dB or higher is required, as also observed in Figure 8. Notably, the front kick was correctly classified even at low SNR values, owing to its distinctiveness from the other motions. However, the squat–lunge and front lateral raise–arm curl pairs were misclassified due to the similarity in their MDs. Specifically, as seen in the TF image (Figure 6), the front lateral raise and arm curl motions were very similar, resulting in significant misclassification of the arm curl motion, with 38 out of 50 misclassifications at SNR = 0 dB and 18 out of 50 misclassifications at SNR = 5 dB. This result indicates that further research is necessary to enhance the accuracy of the proposed method at low SNRs.

4.4. Analysis of the Training Database

4.4.1. Effect of dttr in the Training Database

Considering the results from Section 4.2, the effect of dttr for training data on classification accuracy was analyzed with d = 4 at SNR = 10 dB. Using Tsamp = 1.20 s and Tend = 4.00 s, classifications were performed with dttr = 0.02 s to 0.42 s in increments of 0.10 s. The classification accuracy remained insensitive to changes in dttr, despite a significant reduction in the number of training samples. Specifically, the number Ntr of training vectors per target decreased from 140 to just 6 (Figure 11a,b). Although Ntr was reduced to 4.29% of its original value, Pc only declined by 4.68%; this result demonstrates the efficiency of the proposed feature extraction method. Notably, an accuracy of Pc = 98.68% was achieved with dttr = 0.12 s, meaning only 23 feature vectors of dimension 4 per motion were sufficient to produce results approaching 100% accuracy. Additionally, the processing time tpr for constructing the training data dropped by 95.13% from 17.88 s at dttr = 0.02 s to just 0.87 s at dttr = 0.42 s (Figure 11c). This result demonstrates the computational efficiency of the proposed approach.

4.4.2. Effect of Tsamp

To analyze the effect of Tsamp of the TF on classification accuracy, we varied Tsamp from 1.2 s to 0.4 s in increments of 0.2 s, using SNR = 10 dB, d = 4 and dttr = 0.12 s, as determined in the previous subsections. Classification accuracy was strongly dependent on Tsamp; this result emphasizes that accurate classification of MMs requires samples of sufficient duration (Figure 12a). Although reducing Tsamp shortened tpr (Figure 12b), the classification accuracy decreased due to insufficient time representation of MM. Therefore, because the MMs typically varies over time with a characteristic period, Tsamp must be sufficiently large to allow the features to clearly represent each MMs. In our experiment, MM periods exceeded 1.2 s, so Tsamp ≥ 1.2 s was necessary to achieve Pc > 99% by ensuring distinct feature separation in feature space. Conversely, tpr increased as Tsamp increased, i.e., from 1.92 s at Tsamp = 0.4 s to 3.01 s at Tsamp = 1.2 s.

4.5. Limitations of the Proposed Method and the Solution

The proposed method assumes recognition of a single subject, so several challenges remain to be addressed. In this subsection, we analyze key factors that must be overcome to decrease the sensitivity of the model to variations in conditions.
The first challenge concerns inter-individual variability in movement speed. In our experiments, the training and test data had similar speed characteristics, so Pcs were close to 100%. However, significant differences in movement speed between individuals can lead to substantial mismatches between training and test data. To examine the impact of this variation, we temporally scaled the test data in time-domain by a factor β, ranging from 1.0 to 2.0 in increments of 0.1, under an SNR of 30 dB. As β increased, Pc dropped sharply; this result indicates that the current model is sensitive to variation in the speed of movement (Figure 13). This result underscores the need to either widen the range of speed profiles in the training data or develop an effective rescaling mechanism to adapt to individual differences in movement speed.
The second challenge lies in addressing scenarios in which multiple subjects perform various motions. The proposed method was originally designed without accounting for multiple objects, so applying its features to test vectors composed of random combinations of motions can result in significantly degraded tacking accuracy. To evaluate the effect of multiple subjects, we randomly selected and summed two test data samples at SNR = 30 dB, then applied the proposed method for classification. The resulting Pc was 21.2% (Table 3); this low accuracy emphasizes the need for a dedicated feature representation that can distinguish concurrent motions of multiple entities. This accuracy degradation may be mitigated by adopting a new protocol similar to one that was recently developed for recognizing multiple drones or jets [32,33], which constructs the training dataset by combining instances from single-target training data.
The third challenge concerns the classification of diverse populations using a limited training dataset, particularly when accounting for variations in body types, age groups, or mobility impairments. To highlight this problem, we performed a classification experiment at SNR = 30 dB, in which 50% of the test signals in the time domain were randomly selected and their amplitudes and phases were independently varied within ±50%. As in the case of multiple motion classification, the resulting Pc was degraded (77.3%); this result indicates the need for a fundamental analysis of individual RCS variability in relation to factors such as gender, body size, and other anthropometric differences (Table 3).
Moreover, real-world factors such as posture changes and non-Lambertian RCS variations can introduce significant classification errors. These factors must either be explicitly accounted for or addressed by the development of robust algorithms. Posture changes, for instance, can be treated relatively simply as anomalies during recognition, or incorporated into the training process using the proposed method. In contrast, RCS variation in humans is a particularly complex problem, because it depends heavily on individual body composition. Accurate modeling of such variations requires extensive experimentation that considers a wide range of influencing parameters including aspect angle, operating frequency, environmental conditions, and even the amount of perspiration on the skin.

5. Conclusions

This paper has presented a novel method for motion classification using training data derived from markers and radar signals. Results demonstrated the feasibility of achieving high accuracy without relying on markers in the testing phase. By using PCA for dimensionality reduction and a nearest-neighbor classifier, the method achieved excellent classification accuracy even with compact feature sets, and at various SNRs. In particular, the use of 3D marker coordinates during the training phase enabled reliable feature representation, allowing effective classification of radar signals in the testing phase without markers. The results confirm that the use of data obtained using markers can serve as an effective foundation for motion analysis radar, and that this use delivers high accuracy and efficiency. In our experiments, the parameters that achieved Pc ≥ 99.8% were d = 6, dttr = 0.02 s, and Tsamp = 1.2 s at SNR = 10 dB. However, classification accuracy was highly sensitive to these parameters, and therefore, these settings must be optimized for other motions.
However, this study utilized training and testing datasets derived from only ten subjects to verify classification accuracy. Future research should involve testing with datasets generated from additional subjects to evaluate the generalization and reliability of the radar-based motion recognition algorithm across diverse individuals. Additionally, expanding the radar’s field of view and integrating additional data sources could increase the precision of motion-recognition algorithms. This study provides a strong foundation for radar-based motion analysis technologies, by demonstrating their potential for high accuracy and efficiency in various fields such as healthcare, sports analytics, and security.

Author Contributions

Conceptualization, S.P., J.P. and C.E.; methods, S.P., J.P. and C.E.; software, S.P., S.C., S.J., M.K. and C.E.; validation, J.P., J.Y., M.K. and C.E.; formal analysis, S.P. and C.E.; investigation, S.H., J.Y., M.K. and C.E.; resources, S.P., J.P., and S.H.; data curation, S.H., S.C. and C.E.; writing—original draft preparation, C.E.; writing—review and editing, S.P. and J.P.; visualization, C.E., M.K. and J.Y.; supervision, S.P.; project administration, S.P. and J.P.; funding acquisition, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Pukyong National University (File No. 2024-12-002 2024.12.30).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS-2023-00239144). This research was funded by the project titled ‘Development of Polar Region Communication Technology and Equipment for Internet of Extreme Things (IoET)’, funded by the Ministry of Science and ICT (MSIT), Republic of Korea (No. RS-2021-KS211508).

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Poppe, R. Vision-Based Human Motion Analysis: An Overview. Comput. Vis. Image Underst. 2007, 108, 4–18. [Google Scholar] [CrossRef]
  2. Bleser, G.; Taetz, B.; Miezal, M.; Christmann, C.A.; Steffen, D.; Regenspurger, K. Development of an Inertial Motion Capture System for Clinical Application: Potentials and Challenges from the Technology and Application Perspectives. I-Com 2017, 16, 113–129. [Google Scholar] [CrossRef]
  3. Salim, M.S.; Lim, H.N.; Salim, M.; Baharuddin, M.Y. Motion Analysis of Arm Movement during Badminton Smash. In Proceedings of the IEEE EMBS Conference on Biomedical Engineering and Sciences (IECBES), Kuala Lumpur, Malaysia, 30 November–2 December 2010; pp. 111–114. [Google Scholar]
  4. Karatzas, N.; Abdelnour, P.; Corban, J.P.A.H.; Zhao, K.Y.; Veilleux, L.; Bergeron, S.G.; Fevens, T.; Rivaz, H.; Babouras, A.; Martineau, P.A. Comparing a Portable Motion Analysis System Against the Gold Standard for Potential Anterior Cruciate Ligament Injury Prevention and Screening. Sensors 2024, 24, 1970. [Google Scholar] [CrossRef] [PubMed]
  5. Pagnon, D.; Domalain, M.; Reveret, L. Pose2Sim: An End-to-End Workflow for 3D Markerless Sports Kinematics—Part 1: Robustness. Sensors 2021, 21, 6530. [Google Scholar] [CrossRef] [PubMed]
  6. Seifert, A.; Grimmer, M.; Zoubir, A.M. Doppler Radar for the Extraction of Biomechanical Parameters in Gait Analysis. IEEE J. Biomed. Health Inform. 2020, 25, 547–558. [Google Scholar] [CrossRef] [PubMed]
  7. Hadjipanayi, C.; Yin, M.; Bannon, A.; Rapeaux, A.; Banger, M.; Haar, S.; Constandinou, T.G. Remote Gait Analysis Using Ultra-Wideband Radar Technology Based on Joint Range-Doppler-Time Representation. IEEE Trans. Biomed. Eng. 2024, 71, 234–245. [Google Scholar] [CrossRef]
  8. LaHaie, I.J. Overview of an Image-Based Technique for Predicting Far-Field Radar Cross Section from Near-Field Measurements. IEEE Antennas Propag. Mag. 2003, 45, 159–169. [Google Scholar] [CrossRef]
  9. Sheeny, M.; De Pellegrin, E.; Mukherjee, S.; Ahrabian, A.; Wang, S.; Wallace, A. Radiate: A Radar Dataset for Automotive Perception in Bad Weather. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: New York, NY, USA; pp. 1–7. [Google Scholar]
  10. Chen, V.C.; Li, F.; Ho, S.; Wechsler, H. Micro-Doppler Effect in Radar: Phenomenon, Model, and Simulation Study. IEEE Trans. Aerosp. Electron. Syst. 2006, 42, 2–21. [Google Scholar] [CrossRef]
  11. Kim, Y.; Ling, H. Human Activity Classification Based on Micro-Doppler Signatures Using a Support Vector Machine. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1328–1337. [Google Scholar]
  12. Jia, Y.; Guo, Y.; Wang, G.; Song, R.; Cui, G.; Zhong, X. Multi-Frequency and Multi-Domain Human Activity Recognition Based on SFCW Radar Using Deep Learning. Neurocomputing 2021, 444, 274–287. [Google Scholar] [CrossRef]
  13. Deng, F.; Pu, S.; Chen, X.; Shi, Y.; Yuan, T.; Pu, S. Hyperspectral Image Classification with Capsule Network Using Limited Training Samples. Sensors 2018, 18, 3153. [Google Scholar] [CrossRef] [PubMed]
  14. Lin, Z.; Ji, K.; Kang, M.; Leng, X.; Zou, H. Deep Convolutional Highway Unit Network for SAR Target Classification with Limited Labeled Training Data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1091–1095. [Google Scholar] [CrossRef]
  15. Tian, S.; Wang, C.; Zhang, H.; Bhanu, B. SAR Object Classification Using the DAE with a Modified Triplet Restriction. IET Radar Sonar Navig. 2019, 13, 1081–1091. [Google Scholar] [CrossRef]
  16. Davari, A.; Özkan, H.C.; Maier, A.; Riess, C. Fast and Efficient Limited Data Hyperspectral Remote Sensing Image Classification via GMM-Based Synthetic Samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2107–2120. [Google Scholar] [CrossRef]
  17. Erol, B.; Gurbuz, S.Z.; Amin, M.G. Motion Classification Using Kinematically Sifted Acgan-Synthesized Radar Micro-Doppler Signatures. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 3197–3213. [Google Scholar] [CrossRef]
  18. Zhang, X.; Wang, Z.; Liu, D.; Ling, Q. Dada: Deep Adversarial Data Augmentation for Extremely Low Data Regime Classification. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: New York, NY, USA; pp. 2807–2811. [Google Scholar]
  19. Du, H.; He, Y.; Jin, T. Transfer Learning for Human Activities Classification Using Micro-Doppler Spectrograms. In Proceedings of the IEEE International Conference on Computational Electromagnetics (ICCEM), Chengdu, China, 26–28 March 2018; pp. 1–3. [Google Scholar]
  20. Du, H.; Jin, T.; Song, Y.; Dai, Y.; Li, M. Efficient Human Activity Classification via Sparsity-Driven Transfer Learning. IET Radar Sonar Navig. 2019, 13, 1741–1746. [Google Scholar] [CrossRef]
  21. Park, J.; Javier, R.J.; Moon, T.; Kim, Y. Micro-Doppler Based Classification of Human Aquatic Activities via Transfer Learning of Convolutional Neural Networks. Sensors 2016, 16, 1990. [Google Scholar] [CrossRef]
  22. Seyfioglu, M.S.; Erol, B.; Gurbuz, S.Z.; Amin, M.G. DNN Transfer Learning from Diversified Micro-Doppler for Motion Classification. IEEE Trans. Aerosp. Electron. Syst. 2018, 55, 2164–2180. [Google Scholar] [CrossRef]
  23. Ricci, R.; Balleri, A. Recognition of Humans Based on Radar Micro-Doppler Shape Spectrum Features. IET Radar Sonar Navig. 2015, 9, 1216–1223. [Google Scholar] [CrossRef]
  24. van Dorp, P.; Groen, F.C.A. Feature-Based Human Motion Parameter Estimation with Radar. IET Radar Sonar Navig. 2008, 2, 135–145. [Google Scholar] [CrossRef]
  25. Zeng, Z.; Amin, M.G.; Shan, T. Automatic Arm Motion Recognition Based on Radar Micro-Doppler Signature Envelopes. IEEE Sens. J. 2020, 20, 13523–13532. [Google Scholar] [CrossRef]
  26. Zhao, R.; Ma, X.; Liu, X.; Li, F. Continuous Human Motion Recognition Using Micro-Doppler Signatures in the Scenario with Micro Motion Interference. IEEE Sens. J. 2020, 21, 5022–5034. [Google Scholar] [CrossRef]
  27. Ma, D.; Shang, L.; Tang, J.; Bao, Y.; Fu, J.; Yin, J. Classifying Breast Cancer Tissue by Raman Spectroscopy with One-Dimensional Convolutional Neural Network. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2021, 256, 119732. [Google Scholar] [CrossRef]
  28. Shevgunov, T.; Efimov, E.; Kirdyashkin, V.; Kravchenko, T. The Development of the Algorithm for Estimating the Spectral Correlation Function Based on Two-Dimensional Fast Fourier Transform. In Advances in Signal Processing: Theories, Algorithms, and System Control; Springer: Berlin/Heidelberg, Germany, 2020; pp. 75–86. [Google Scholar]
  29. Tahmoush, D. Review of Micro-Doppler Signatures. IET Radar Sonar Navig. 2015, 9, 1140–1146. [Google Scholar] [CrossRef]
  30. Eom, C.; Chun, S.; Kim, B.; Joo, S.; You, J.; Yoon, J.; Han, S.; Park, J.; Park, S. Simulation Study for the Motion Classification Using the Radar and the 3D Motion Analysis System. Korean J. Appl. Biomech. 2024, 34, 240–250. [Google Scholar]
  31. Kim, K.; Seo, D.; Kim, H. Efficient Radar Target Recognition Using the MUSIC Algorithm and Invariant Features. IEEE Trans. Antennas Propag. 2002, 50, 325–337. [Google Scholar]
  32. Park, S.; Kim, M.; Jung, J.; Riggs, L.; Kim, K. Automatic Recognition of Range Profiles and ISAR Images of Targets Flying in Formation. IEEE Sens. J. 2019, 19, 515–524. [Google Scholar] [CrossRef]
  33. Yoon, S.; Kim, S.; Jung, J.; Cha, S.; Baek, Y.; Koo, B.; Choi, I.; Park, S. Efficient Protocol to Use FMCW Radar and CNN to Distinguish Micro-Doppler Signatures of Multiple Drones and Birds. IEEE Access 2022, 10, 26033–26044. [Google Scholar] [CrossRef]
Figure 1. Comparison between A1 and Rs(t).
Figure 1. Comparison between A1 and Rs(t).
Sensors 25 03293 g001
Figure 2. Overall classification procedure.
Figure 2. Overall classification procedure.
Sensors 25 03293 g002
Figure 3. Marker positions.
Figure 3. Marker positions.
Sensors 25 03293 g003
Figure 4. Construction of the training data sampled at uniform time interval.
Figure 4. Construction of the training data sampled at uniform time interval.
Sensors 25 03293 g004
Figure 5. TF image of each motion (simulated).
Figure 5. TF image of each motion (simulated).
Sensors 25 03293 g005
Figure 6. TF image of each motion (measured).
Figure 6. TF image of each motion (measured).
Sensors 25 03293 g006
Figure 7. Feature vectors of each motion (d = 3).
Figure 7. Feature vectors of each motion (d = 3).
Sensors 25 03293 g007
Figure 8. Pcs of entire motions for various PCA dimensions (d) and SNRs.
Figure 8. Pcs of entire motions for various PCA dimensions (d) and SNRs.
Sensors 25 03293 g008
Figure 9. Pcs of each motion for various PCA dimensions (d) and SNRs.
Figure 9. Pcs of each motion for various PCA dimensions (d) and SNRs.
Sensors 25 03293 g009
Figure 10. Confusion matrices for three SNRs.
Figure 10. Confusion matrices for three SNRs.
Sensors 25 03293 g010
Figure 11. Pcs, Ntrs, and tprs for various dttrs.
Figure 11. Pcs, Ntrs, and tprs for various dttrs.
Sensors 25 03293 g011
Figure 12. Pcs and tprs for various Tsamps.
Figure 12. Pcs and tprs for various Tsamps.
Sensors 25 03293 g012
Figure 13. Pcs for time-scaling.
Figure 13. Pcs for time-scaling.
Sensors 25 03293 g013
Table 1. Characteristics of subjects (mean ± s.d).
Table 1. Characteristics of subjects (mean ± s.d).
Variables (Mean ± Std)6 Males4 Females
Height (cm) 174.00 ± 5.33166.38 ± 3.82
Weight (kg)82.00 ± 11.0357.50 ± 6.45
Age (y)25.67 ± 1.2124.50 ± 2.38
Table 2. Parameters for experiment.
Table 2. Parameters for experiment.
ParameterValueParameterValue
fc5.80 GHzλ0.05 m
Tend4.00 sTsamp1.20 s
dttr0.02 sNtr140
d4SNR0–20 dB
Table 3. Result of classification for two motions and body variation.
Table 3. Result of classification for two motions and body variation.
Single Motion Two MotionsBody Variation
99.12%21. 2%77.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Eom, C.; Han, S.; Chun, S.; Joo, S.; Yoon, J.; Kim, M.; Park, J.; Park, S. Efficient Markerless Motion Classification Using Radar. Sensors 2025, 25, 3293. https://doi.org/10.3390/s25113293

AMA Style

Eom C, Han S, Chun S, Joo S, Yoon J, Kim M, Park J, Park S. Efficient Markerless Motion Classification Using Radar. Sensors. 2025; 25(11):3293. https://doi.org/10.3390/s25113293

Chicago/Turabian Style

Eom, Changhyeon, Sooji Han, Sabin Chun, Soyoung Joo, Jisu Yoon, Min Kim, Jongchul Park, and Sanghong Park. 2025. "Efficient Markerless Motion Classification Using Radar" Sensors 25, no. 11: 3293. https://doi.org/10.3390/s25113293

APA Style

Eom, C., Han, S., Chun, S., Joo, S., Yoon, J., Kim, M., Park, J., & Park, S. (2025). Efficient Markerless Motion Classification Using Radar. Sensors, 25(11), 3293. https://doi.org/10.3390/s25113293

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop