Next Article in Journal
EcoSpec: Highly Equipped Tower-Based Hyperspectral and Thermal Infrared Automatic Remote Sensing System for Investigating Plant Responses to Environmental Changes
Previous Article in Journal
Instrumental Assessment of Stepping in Place Captures Clinically Relevant Motor Symptoms of Parkinson’s Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Millimeter-Wave Array Radar-Based Human Gait Recognition Using Multi-Channel Three-Dimensional Convolutional Neural Network

College of Electronic Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(19), 5466; https://doi.org/10.3390/s20195466
Submission received: 14 August 2020 / Revised: 13 September 2020 / Accepted: 21 September 2020 / Published: 23 September 2020
(This article belongs to the Section Electronic Sensors)

Abstract

:
At present, there are two obvious problems in radar-based gait recognition. First, the traditional radar frequency band is difficult to meet the requirements of fine identification with due to its low carrier frequency and limited micro-Doppler resolution. Another significant problem is that radar signal processing is relatively complex, and the existing signal processing algorithms are poor in real-time usability, robustness and universality. This paper focuses on the two basic problems of human gait detection with radar and proposes a human gait classification and recognition method based on millimeter-wave array radar. Based on deep-learning technology, a multi-channel three-dimensional convolution neural network is proposed on the basis of improving the residual network, which completes the classification and recognition of human gait through the hierarchical extraction and fusion of multi-dimensional features. Taking the three-dimensional coordinates, motion speed and intensity of strong scattering points in the process of target motion as network inputs, multi-channel convolution is used to extract motion features, and the classification and recognition of typical daily actions are completed. The experimental results show that we have more than 92.5% recognition accuracy for common gait categories such as jogging and normal walking.

1. Introduction

Pedestrian retrieval and identification has always been an urgent need in the fields of anti-terrorism security checks, crime investigation, medical inference, etc. Due to the long distance between monitoring equipment and pedestrians, the commonly used biometric recognition methods such as face recognition and iris recognition cannot be applied to actual scenes. However, gait, a biological feature, can be effectively applied to pedestrian retrieval and recognition due to its characteristics of being unhidden, without disguise, non-invasive, stable, and remotely accessible [1,2].
Electromagnetic waves can provide long-distance detection of the human body and are not limited by time and climate conditions. Therefore, research on human body detection and feature extraction using radar has attracted more and more attention, which is a necessary supplement to traditional methods. Human gait information mainly refers to the micro-motion features of the human body, which contain effective information of the human motion state and body posture and are an important basis for human target authentication and recognition [3,4,5].
In 2018, Ann-Kathrin et al. studied motion classification based on radar micro-Doppler characteristics [6]. This paper presents a new classification method based on physical features, subspace features and harmonic models. By processing the echo data, analyzing the spectrogram (depicting Doppler and micro-Doppler signals corresponding to velocity and its time-varying properties) and the frequency–velocity gram (emphasizing periodicity and better describing the harmonic components of limbs), it is proved that different walking modes can be distinguished according to the results of feature extraction. In 2019, they used Doppler radar to monitor indoor human movements. The study found that by extracting Doppler signals, a slight difference in step size between the two legs can be detected, thus allowing the judgement of gait asymmetry and diagnosing whether the target human body has dyskinesia [7]. These related studies prove the feasibility of gait recognition using the micro-Doppler characteristics of radar signals.
In previous studies, radar-based gait recognition mainly relied on low-frequency radar with a single transmitting antenna and a single receiving antenna. A single antenna cannot carry out multi-dimensional detection and can provide less target feature information. At the same time, due to the low carrier frequency and small bandwidth, the resolution of radar is difficult to meet the requirements of high-precision identification with [8,9]. Meanwhile, traditional feature-extraction methods such as Jordan transform and principal component analysis (PCA) have poor real-time performance [10], and common classification algorithms such as invisible Markov model and support vector machine (SVM) are difficult to achieve high accuracy with [11,12,13]. Therefore, new feature-extraction and classification algorithms are urgently needed.
Millimeter-wave (MMW) radar works in the transition frequency band between microwaves and optics. It is easy to use to realize signals with large bandwidth and narrow beam, has an extremely high resolution, can obtain the fine structure and motion characteristics of targets, and has great application prospects in military and civil fields. MMW radar has two special advantages. The first advantage is high micro-Doppler sensitivity. That is, targets that cannot be distinguished by micro-Doppler differences in the traditional low-frequency band can be distinguished in the MMW band. Secondly, MMW radar can image targets with a high resolution and high frame rate, which makes it possible to retrieve and identify target characteristics with high precision based on images [14,15]. Combining MMW ultra-wideband signals with multiple-input multiple-output (MIMO) arrays can further obtain three-dimensional (3-D) spatial information on the detection area, thus obtaining a richer spatial distribution of targets [16].
The vigorous development of deep learning (DL) provides a unified framework for radar signal processing that integrates perception, processing and decision-making. Compared with the traditional pattern-recognition method, the DL method has the advantages of the automatic extraction of deep features and high recognition accuracy, and it has good universality. [17,18]. In recent years, the deep-learning (DL) technique has become a research hotspot in various fields, such as object classification and segmentation [19,20], super-resolution [21,22], image denoising [23,24], medical image reconstruction [25,26], etc. In addition to the above applications, it is also adopted in radar signal-processing applications. The typical cases include target detection [27], synthetic aperture radar(SAR)image interpretation [28] and moving-human-body classification [29,30].
Most gait recognition methods that combine radar sensors with DL mainly use convolution neural networks (CNN) to extract and recognize features of micro-Doppler signatures [6,7,31]. The birth of the 3-D spatiotemporal CNN provides a better method for processing the gait information of a time sequence [32,33,34]. It also provides more research directions for gait recognition technology. In 2018, Mingmin Zhao et al. of the Massachusetts Institute of Technology realized the through-wall identification of a target human body through radar sensors [35]. In this study, they used an optical collector and radar sensor to track and collect data of the target human body at the same time. The supervised learning framework based on the three-dimensional convolution neural network (3-D CNN) was used, and the obtained optical pictures and radar echo data were used for training. As a result, they realized the posture recognition of the human body in real time and effectively avoided the problem of recognition-rate decline caused by obstacle occlusion by using radar sensors. In 2019, this team proposed 3-D human pose estimation based on radar signals on the basis of previous research [36]. Their research results proved that the three-dimensional spatiotemporal convolution neural network had great advantages in processing spatiotemporal continuous gait data.
Based on the above analysis, this paper proposes a human gait recognition technology based on MMW array radar around the two basic problems of human gait detection by radar. After recording the echo data of array radar, we can use signal-processing methods such as linear filtering, time–frequency transformation, wavenumber domain compensation imaging and 3-D point cloud feature extraction to obtain abundant one-dimensional, two-dimensional and three-dimensional feature information, which can be used as input samples for the network. At the same time, a multi-channel three-dimensional convolution neural network (MC-3DCNN) is proposed based on improving the residual network (ResNet), which completes the classification and recognition of human gait through the hierarchical extraction and fusion of multi-dimensional features.
The rest of this paper is organized as follows. The generation of the point cloud of human gait is described in Section 2. Section 3 discusses the structure of the proposed MC-3DCNN. The training, analysis and comparison of the network are described in Section 4. The conclusions are drawn in Section 5.

2. Generation of Point Cloud of Human Gait

2.1. Frequency-Modulated Continuous Wave and Range Measurement

Frequency-modulation continuous wave (FMCW) radar, which can measure the range, velocity and angle of the target by transmitting FMCWs, plays an important role in security and intelligent driving.
By mixing the received signal with the transmitted signal, we can obtain the intermediate frequency (IF) signal. Suppose the distance between the stationary target and the radar is R , the speed of electromagnetic wave is c and the slope of the chirp signal is K ; then, the IF signal is expressed as:
S I F = e i 2 π ( K 2 R c t ) = e i 2 π ( f I F t )
The above signal is a single-frequency signal (for a single stationary target). By Fourier transformation (FT), we can find out the position of peak and obtain the IF, f I F = K 2 R c ; then, the distance of the target is expressed as:
R = f I F c 2 K
For multiple targets, the IF signal is the superposition of multiple single-frequency signals. After the fast Fourier Transformation (FFT), there will be multiple peaks in the amplitude spectrum, and the frequency point of each peak is proportional to the distance between the target and the radar.

2.2. The Principle of Velocity Estimation

Supposing the target is located at R 0 in the initial time and is far away from the radar at a speed of v , the distance between the target and radar at time t is R = R 0 + v t . The time delay τ can be expressed as τ = 2 R c = 2 ( R 0 + v t ) c , and the initial frequency of radar is f 0 ; then, the received signal S r can be expressed as:
S r = e j 2 π [ f 0 ( t 2 ( R 0 + v t ) c ) + 1 2 K ( t 2 ( R 0 + v t ) c ) 2 ]
Due to the extremely short processing time (usually approximately m s or u s ), the term of t 2 can be ignored; at the same time, the term with c 2 in the denominator can also be ignored. Thus, the IF signal of the moving target can be expressed as
S I F = e j 2 π [ ( 2 R 0 K + f 0 v c ) t + 2 R 0 f 0 c ]
Considering three practical situations under this model:
  • Time is discrete and determined by the sampling period T s .
  • A total of N data are sampled per pulse.
  • L pulses are continuously emitted.
  • The radial component of the target velocity is constant v .
Modifying the mathematical model of the above IF signal, we can obtain:
S I F = e j 2 π [ ( 2 v f 0 c + 2 K ( R 0 + v l T c ) c ) n T s + 2 ( R 0 + v l T c ) c f 0 ]
where n = 0 ,   1 ,   2 ,   ,   N 1 means a single pulse sampling point sequence. R 0 denotes the radial distance between the radar and target at time 0 (the start time of the first radar pulse). l = 0 , 1 , 2 , , L 1 represents the pulse sequence. T c denotes the pulse repetition time (the time difference between the start of two adjacent pulses).
Analysis of Equation (5) shows that for a specific pulse ( l fixed), the S I F is still a single frequency signal. Compared with IF signal of a stationary target, it has a fixed value e j 2 π [ ( 2 ( R 0 + v l T c ) c f 0 ) ] , which can be regarded as the complex envelope of the initial signal. The mathematical model of the initial signal can be expressed as:
S i n = e j 2 π [ ( 2 v f 0 c + 2 K ( R 0 + v l T c ) c ) n T s ]
Using n as the independent variable, the FFT of the signal in Equation (6) can be used to obtain the frequency component of the signal:
f I F = 2 v f 0 c + 2 K ( R 0 + v l T c ) c
Taking l as the independent variable, the FFT for different pulses is equivalent to Fourier analysis for the phase components of the above signals, and the phase information of the signals can be obtained, which includes the speed of the target.

2.3. The Principle of DOA Estimation

In the array radar system, we can estimate the direction of arrival (DOA) and obtain the spatial angle information of the target by using the spatial phase difference. As shown in the Figure 1, there is a uniform linear array with a total of M array elements, the distance between the array elements is d , and a signal (assumed to be a plane wave) is injected into the array from a direction θ away from the normal. It can be seen that the signal has to travel a further distance d sin ( θ ) to reach the second array element than to reach the first one, and so on. The signal has to travel a further distance d sin ( θ ) to reach the latter array element than the previous one.
After calculating the time difference of transmission, it can be concluded that for the same signal, the time to reach the latter array element is δ t = d sin ( θ ) c later than that of the previous array element. Assuming that the frequency of the signal is f 0 and the first array element is taken as the reference point, the time difference between each array element and the first array element is:
Δ t = [ 0 , d sin ( θ ) c , 2 d sin ( θ ) c , , ( M 1 ) d sin ( θ ) c ]
Then, the phase difference between the signals arriving at each array element and the first array element should be:
Δ ϕ = [ 0 , 2 π f 0 d sin ( θ ) c , 2 π f 0 2 d sin ( θ ) c , , 2 π f 0 ( M 1 ) d sin ( θ ) c ]
Since this phase difference is caused by different spatial positions between array elements, it is called the “spatial phase difference”.
In the time domain, the digital frequency can be extracted from the discrete time signal through FFT:
y ( n ) = [ sin ( 2 π f 0 × 0 ) , sin ( 2 π f 0 × T s ) , sin ( 2 π f 0 × 2 T s ) , , sin ( 2 π f 0 × ( N 1 ) T s ) ]
and the extracted digital frequency is:
ω = 2 π f 0 T s = 2 π f 0 f s
It is clearly seen from observation that this digital frequency is the phase difference Δ ϕ = 2 π f 0 T s = 2 π f 0 f s between adjacent sampling points.
Corresponding to the airspace, assuming that the far-field signal is s ( n ) and the first array element (numbered 0 in the above figure) of the array is taken as a reference, the signal received by the entire array is:
X ( n ) = [ s ( n ) , s ( n ) e j 2 π f 0 d sin ( θ ) c , s ( n ) e j 2 π f 0 2 d sin ( θ ) c , s ( n ) e j 2 π f 0 3 d sin ( θ ) c , , s ( n ) e j 2 π f 0 ( M 1 ) d sin ( θ ) c ]
Through simple deformation, we can obtain:
X ( n ) = [ 1 , e j 2 π f 0 d sin ( θ ) c , e j 2 π f 0 2 d sin ( θ ) c , e j 2 π f 0 3 d sin ( θ ) c , , e j 2 π f 0 ( M 1 ) d sin ( θ ) c ] s ( n )
It is clearly seen that the expression of received signal vector X ( n ) is a vector [ 1 , e j 2 π f 0 d sin ( θ ) c , e j 2 π f 0 2 d sin ( θ ) c , e j 2 π f 0 3 d sin ( θ ) c , , e j 2 π f 0 ( M 1 ) d sin ( θ ) c ] multiplied by a scalar s ( n ) , and this vector is a function of the signal incoming direction θ . Define:
a ( θ ) [ 1 , e j 2 π f 0 d sin ( θ ) c , e j 2 π f 0 2 d sin ( θ ) c , e j 2 π f 0 3 d sin ( θ ) c , , e j 2 π f 0 ( M 1 ) d sin ( θ ) c ]
The vector a ( n ) contains angle information of the signal s ( n ) . The received signal can be represented as:
X ( n ) = a ( θ ) s ( n )
When the steering vector and the received signal are written as column vectors, the received signal can be expressed as:
X ( n ) M × 1 = a ( θ ) M × 1 s ( n )
When N signals s 1 ( n ) , s 2 ( n ) , , s N ( n ) are incident on the array from θ 1 , θ 2 , , θ N respectively, the received signal can be expressed as:
X ( n ) M × 1 = a ( θ 1 ) s 1 ( n ) + a ( θ 2 ) s 2 ( n ) + + a ( θ N ) s N ( n ) = [ a ( θ 1 ) , a ( θ 2 ) , , a ( θ N ) ] M × N × [ s 1 ( n ) , s 2 ( n ) , , s N ( n ) ] 1 × N T A M × N s N × 1
Constructing a steering vector with an incoming direction of α :
a ( α ) [ 1 , e j 2 π f 0 d sin ( α ) c , e j 2 π f 0 2 d sin ( α ) c , e j 2 π f 0 3 d sin ( α ) c , , e j 2 π f 0 ( M 1 ) d sin ( α ) c ]
The vector inner product of the steering vector a ( α ) and the received signal is obtained:
y = a H ( α ) X ( n ) = a H ( α ) a ( θ ) s ( n )
The result is a scalar, and by calculating, we can obtain:
y = [ 1 + e j 2 π f 0 d sin ( α ) sin ( θ ) c + e j 2 π f 0 d 2 [ sin ( α ) sin ( θ ) ] c + + e j 2 π f 0 d ( M 1 ) [ sin ( α ) sin ( θ ) ] c ] s ( n ) M s ( n )
The equal sign holds when α = θ .
Here, we introduce a DOA estimation method (see Table 1).
According to the above method, we can accurately estimate the direction of arrival and thus determine the direction of the target.

2.4. System and Generation of Point Cloud

In this study, a 77 GHz FMCW array radar was used as the system platform, which is shown in (a) of Figure 2. The radar has three transmitting antennas and four receiving antennas, which can be equivalent to 12 virtual apertures according to the principle of equivalent phase centers, as shown in (b) of Figure 2.
The MMW array radar has an initial frequency of 77 GHz and a bandwidth of 4 GHz, which can provide better range resolution, while multiple aperture detection provides more azimuth angle data. At the same time, the radar has its own digital signal processor (DSP), which can carry out FFT on echo data in multiple dimensions, thus completing the ranging, velocity measurement and angle measurement of the target.
In this study, radar was used to quickly track and locate the human body, and point cloud data of strong scattering points of the human body were obtained in real time through algorithm processing, which is shown in Figure 3. The point cloud data include the 3-D spatial coordinates, radial velocity and intensity of each scattering point.
Through parameter setting, the radar can collect 20 frames of point cloud data per second; each frame contains 0–64 point clouds of strong scattering points of the human body, and each point cloud is a five-dimensional (5-D) array composed of the 3-D spatial coordinates, radial velocity and intensity of the point.
Generally speaking, the human gait period is 1–1.32 s, and each gait period contains a series of typical posture shifts. In the process of walking, the trunk of each part of the human body moves regularly, and the movement characteristics of each strong scattering point of the human body also show regular changes. The acquired data maintain certain continuity in time and space, and the motion characteristics of the target can be grasped by calculating and analyzing the motion data. The 3-D spatial coordinates, radial velocity and intensity of strong scattering points in the motion process are obtained, and the feature extraction of the data sequence is carried out by using spatiotemporal convolution, thus completing the classification and recognition of human gait.

3. Human Gait Recognition Based on Multi-Channel 3-D Convolutional Neural Network

The convolution neural network (CNN) has enabled great achievements in the field of image recognition. Using this model, image features can be extracted. In video classification, the traditional method is to extract the features of each key frame by using two-dimensional (2-D) CNN and then combine the features of the key frames by using relevant algorithms. There is a problem in this method: when using 2-D CNN, each frame of image is taken as a still picture for feature extraction, and the motion information in the time dimension is lost.
3-D CNN based on spatiotemporal convolution has been proved to be an effective spatiotemporal feature-extraction method. It extracts static information while retaining motion information and has achieved good results in video classification and other tasks.
In this study, we used the point cloud data of human gait as training samples. The point cloud data of human gait have certain continuity in their time and space distribution, which can be regarded as continuous frames in video classification. Therefore, the 3-D CNN can also be applied to human gait recognition.

3.1. Three-Dimensional Convolutional Neural Network

Compared with the 2-D CNNs, the 3-D CNNs have better performance in time information modeling and spatiotemporal feature learning. In the 3-D CNNs, the convolution and pooling processes are completed in the space-time dimension, while in the 2-D CNNs, the convolution and pooling processes are only completed in space. Figure 4 shows the difference between the two network structures.
The three graphs in Figure 4 show the results of different convolution processing for different data. In (a), 2-D convolution is applied to an image (2-D array), and the output is an image (2-D array) too. In (b), 2-D convolution (multiple frames are regarded as multiple channels) is applied to a video (3-D array, including time series), and the output is an image (2-D array). In (c), the application of 3-D convolution on the video produces a 3-D array. Using 2-D convolution will cause the loss of the time information of the input signal after each operation, while 3-D convolution can retain the time information of the input signal.
Therefore, in this study, the method of 3-D convolution was used to extract spatiotemporal features, which include motion features in addition to static features and can achieve better results in gait classification tasks.

3.2. Proposed Network Architecture

As shown in Figure 5, the proposed multi-channel three-dimensional convolution neural network (MC-3DCNN) consists of three channels, whose inputs are the 3-D spatial coordinates, radial velocity and intensity of the scattering points, respectively. In addition, each channel shares the same network topology, including three convolution layers and one pooling layer. The concatenation layer is used to fuse the features extracted hierarchically, and then, one convolution and one pooling layer are used to further extract and reduce the dimensions of the data and the fully-connected layer is then used to complete the classification. Finally, the classification results are given by the s o f t m a x classifier. The weights of each channel are updated independently to ensure that the characteristics of different sample data are fully learned.
Three channels are set up in the network; one channel takes 3-D spatial coordinates as input, and the other two channels take radial velocity and intensity as input, respectively. Considering the symmetry of the network structure, we will introduce it in detail with the structure of Channel 1 as the representative. First, we segment the gait sequence samples evenly. Each group of gait data contains multiple gait periods and can be divided into multiple gait segments. The gait period of the human body is usually 1–1.32 s. Here, we set the duration of each gait segment to be 2 s, including a complete gait period. Since the sampling rate is 20 frames per second, the gait data are taken as a training sample every 40 frames (2 s). Therefore, the input sizes are 3 × 64 × 40, 1 × 64 × 40 and 1 × 64×40, respectively. The kernel size of the first convolutional layer is set to 3 × 3 × 3, and the stride size, to 1 × 1 × 1, generating 64 feature maps. Then, these feature maps are subsampled by the max pooling layer with a kernel size of 3 × 3 × 3 and stride size of 2 × 2 × 2. The kernel size of the second convolutional layer is 3 × 3 × 3, and that of the stride is 2 × 2 × 1, generating 128 feature maps. The kernel size of the third convolutional layer is 3 × 3 × 3, and the stride size is 1 × 1 × 1. Here, in order to fit the function better and avoid the gradient’s disappearance, we use the structure of a residual network for reference and add a shortcut channel. The shortcut channel includes a convolution with a kernel size of 3 × 3 × 3 and stride size of 2 × 2 × 1. Then, the convolution results of each channel are fused through a concatenation layer to form new feature maps. The kernel size of the forth convolution layers is 3 × 3 × 3, and the stride size is 2 × 2 × 2, generating 64 feature maps. The following is an average pooling layer with a kernel size of 3 × 3 × 3 and stride size of 2 × 2 × 2. We use the f l a t t e n function to expand the obtained feature maps and use these as inputs of the fully connected layer. Here, in order to avoid the over-fitting of the network, we set up a dropout layer. In addition, a batch normalization (BN) layer was added after each convolutional layer, and all the activation function is R e L U .
For channel 1, channel 2 and channel 3, the output vectors of the third convolutional layers are defined as F 1 = ( f 1 1 , f 2 1 , , f N 1 1 ) , F 2 = ( f 1 2 , f 2 2 , , f N 2 2 ) and F 3 = ( f 1 3 , f 2 3 , , f N 3 3 ) , respectively. Then the concatenation operation in concatenation layer is defined as:
o = F 1 F 2 F 3 ( f 1 1 , f 2 1 , , f N 1 1 , f 1 2 , f 2 2 , , f N 2 2 , f 1 3 , f 2 3 , , f N 3 3 )
where N 1 = N 2 = N 3 is the number of elements in the vector.
During the backpropagation, we define the cross-entropy loss function as:
C = 1 N i = 1 N j = 1 K y i j log ( y ^ i j )
where N denotes the number of samples, K represents the number of categories, y i j is the expected probability that sample i belongs to category j , and y ^ i j denotes the actual output probability of sample i belongs to category j .

4. Experimental Results

4.1. Dataset Generation

In order to classify and recognize human gait accurately, we used a 3Tx-4Rx MMW array radar to detect targets and obtained gait data in various actual scenes, and the radar model was IWR1443BOOST from Texas Instruments. The experimental scene is shown in Figure 6. The radar was placed at a height of 0.8 m from the ground, and the target moved along an unfixed path within the radar detection range as shown in (a) of Figure 6. The radar detected the human body through scanning and used fast algorithms to realize range, speed and angle measurements of strong scattering points of the human body, thus realizing fast tracking and positioning. The obtained data were point cloud data including the 3-D spatial coordinates, radial velocity and intensity of strong scattering points.
The target moved along an unfixed path in the radar detection area, and the moving modes were normal walking (natural swing of both arms), jogging (raised forearm and natural swing of big arm), lame walking (normal walking for one leg, the dragging of one leg behind, and a slight swing of both arms) and squatting down and standing up. In order to increase the diversity of the sample data and enhance the robustness of the network model, in the experiment, we detected the human gait in various scenes (including corridors, basketball courts and parking lots). Each person collected 5000 frames of data for each type of action, and a total of eight different people (four men and four women) were observed. A total of 40,000 frames of sample data were collected for each type of action. The specific sample number and label division are shown in the Table 2.
The starting frequency of the radar is 77 GHz, and the bandwidth is 4 GHz. Through parameter setting, the radar can achieve a sampling rate of 20 frames per second, and each frame of data contains multi-dimensional data for 0–64 point clouds. The CNN realizes feature extraction by sliding convolution on the data matrix according to a certain step size through a convolution kernel. In this process, the input data are required to be a matrix with row and column rules. However, in this experiment, the data collection was random, that is, the number of point clouds in each frame of data was not always consistent. Therefore, it was necessary to preprocess the collected point cloud data and fill the data dimensions into regular shapes that were suitable for convolution network input. The main steps of preprocessing included data denoising, data filling, data smoothing, dataset expansion, and the division of the training set and test set. The specific operation flow is shown in Figure 7. The specific implementation method was as follows: firstly, a distance threshold was set in the 3-D space, and some noise points or outliers and other miscellaneous data in the data could be eliminated through the distance threshold, thus realizing the denoising of the data. Then, the dimensions of frames with few points were expanded to ensure the consistency of the input data dimensions. Here, we used null values to fill in, so that each frame of data contained 64 point clouds, thus completing the data filling. After dimension expansion, the point cloud data could not be null, because null values cannot provide useful data information for the convolution process. Therefore, we needed to copy the data of adjacent points, so that each frame of the point cloud was a 5-D array containing 64 valid points; thus, we realize data smoothing. In the case of small samples, it is necessary to expand the dataset by means of flipping and cropping, and then, the training set and the test set can be divided. The specific data preprocessing process is shown in the following figure.
In order to evaluate the network model, we adopted a new data partitioning strategy: N-fold crossover, which can avoid the limitations and particularity brought by the fixed partitioning of datasets. In the experiment, we divided the sample data into four folds, that is, as shown for each category, we divided the sample data into four equal parts for four-fold cross-validation (CV), as shown in Table 3. The division results for the training set and test set are shown in Table 4.

4.2. Classification Results and Algorithm Comparison

In the network training, we used a desktop computer with the Windows 10 system for training. The computer was equipped with a TITAN XP graphics card. For each cross-validation group, the training lasted for 60 epochs, and each training took approximately 2 h. The initial learning rate of the network is 1 e 5 , which degrades to 90% after every 20,000 iterations. To demonstrate that the proposed MC-3DCNN achieves higher recognition accuracy by enhancing and the fusion of features for 3-D spatial coordinates, radial velocity and intensity in the three channels, rather than increasing the number of training samples, we designed an experiment for a single-channel three-dimensional convolutional neural network (SC-3DCNN). The structure of the SC-3DCNN is shown in Figure 8, which has similar structural parameters to the MC-3DCNN. The SC-3DCNN integrates the 3-D spatial coordinates, radial velocity and intensity of every 40 frames of data into one sample as the input of the network.
The results of the four-fold cross-training are shown in Figure 9 and Figure 10. It can be seen from the training results that the network structure and parameter setting fit the distribution of the training data well, thus obtaining good training results in the four-fold cross training. The accuracy of the network continuously improved in the training process, and the accuracy of the network had reached more than 95% after 30 epochs of training. The training error converged smoothly, and after 30 epochs of training, the error converged to close to 0.
The trained network models were used for four-fold cross-validation to test the recognition accuracies of the network model for various actions. The recognition results are shown in the data in Table 5.
From the data in the table, it can be seen that the MC-3DCNN can better identify the three types of movements with strong continuity of movements and obvious micro-Doppler characteristics, including jogging, normal walking, squatting, and standing up, with an average accuracy rate of more than 92%. However, for the gait category with weak movement continuity and less obvious micro-Doppler characteristics, such as lame walking, the recognition accuracy is lower than that for the other three movements.
The four-fold cross-validation was carried out on the SC-3DCNN and MC-3DCNN, and the average verification results of each cross-validation are shown in the following Table 6.
The recognition results demonstrate that the proposed MC-3DCNN outperforms SC-3DCNN and achieves higher recognition accuracy by extracting features independently of multiple channels.

5. Conclusions

In this paper, an MMW array radar-based human gait recognition method using MC-3DCNN is proposed. In this work, we used a 77 GHz MMW array radar as the sensor and used a multi-dimensional array antenna to obtain multi-dimensional information on the target. At the same time, in order to fuse multi-dimensional features, we propose a MC-3DCNN based on the 3-D CNN. The 3-D spatial coordinates, radial velocity and intensity of the observed human gait were extracted with hierarchical features, and then, the classification and recognition of human gait were completed by using multi-dimensional feature fusion. The network model proposed by us can achieve more than 92.5% recognition accuracy for daily travel modes such as jogging and normal walking.
Future work can be carried out on the following aspects: First, the recognition at this stage is mainly to classify the gait of the human. The next step could complete the identification of the human under the premise of a large-scale expansion of the sample database. Second, the current target of gait recognition is mainly a single human, and the next step would be mainly gait segmentation and classification recognition in complex scenes with multiple targets. Third, the current design is mainly for simple target gait category classification and identification; the next step could be carried out in the direction of using gait changes to analyze the intention of the perpetrator, which would play a positive role in gait recognition in the field of security inspection.

Author Contributions

Conceptualization, X.J. and Q.Y.; methodology, X.J.; software, Y.Z.; validation, X.J. and Y.Z.; writing—original draft preparation, X.J.; writing—review and editing, B.D. and H.W.; project administration, B.D. and H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by The National Natural Science Foundation of China (No. 61871386 and No. 61701513) and the Preeminence Youth Foundation of Hunan Province (No. 2019JJ20022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, K.J.; Hou, B.B. A review of gait recognition. Chin. J. Image Graph. 2007, 12, 1152–1160. [Google Scholar]
  2. Tahmoush, D.; Clark, J.; Silvious, J. Tracking of dismounts moving in cross-range using GMTI radar. In Proceedings of the 11-th International Radar Symposium, Vilnius, Lithuania, 16–18 June 2010. [Google Scholar]
  3. Dorp, P.V.; Groen, F.C.A. Feature-Based Human Motion Parameter Estimation with Radar. IET Radar Sonar Navig. 2008, 2, 135–145. [Google Scholar] [CrossRef]
  4. Chen, V.C.; Li, F.; Ho, S.S. Analysis of micro-Doppler Signatures. IEE Proc. Radar Sonar Navig. 2003, 150, 271–276. [Google Scholar] [CrossRef]
  5. Chen, V.C. Doppler Signatures of Radar Backscattering from Objects with Micro-Motions. IET Signal Process. 2008, 2, 291–300. [Google Scholar] [CrossRef]
  6. Seifert, A.K.; Amin, M.G.; Zoubir, A.M. Toward Unobtrusive in-home Gait Analysis Based on Radar Micro-Doppler Signatures. IEEE Trans. Biomed. Eng. 2019, 66, 2629–2640. [Google Scholar] [CrossRef] [Green Version]
  7. Seifert, A.K.; Amin, M.G.; Zoubir, A.M. Detection of Gait Asymmetry Using Indoor Doppler Radar. In Proceedings of the IEEE Radar Conference, Boston, MA, USA, 22–26 April 2019. [Google Scholar]
  8. Cui, L.M. Research on Remote Sensing Mechanism and Information Extraction Method of X-band Radar Waves and Current. Ph.D. Thesis, Chinese Academic of Science (Institute of Oceanology), Qingdao, China, June 2010. [Google Scholar]
  9. Zhang, C.Y.; Fang, Z.H.; Zhang, W.H.; Zhao, N.; Wu, M.Q.; Ruan, X.C. Data acquisition and imaging processing of vehicle-mounted X-band synthetic aperture radar. Radar Sci. Technol. 2001, 3, 1–4+16. [Google Scholar]
  10. Sun, C.X. Human Face Recognition with PCA Method. J. Yunnan Univ. Natl. Nat. Sci. Ed. 2010, 19, 439–443. [Google Scholar]
  11. Wang, J.; Chen, Y.; Deng, S. Motion Recognition Method Based on Improved SVM Classifier. J. Chongqing Univ. Nat. Sci. Ed. 2016, 39, 12–17. [Google Scholar]
  12. Tang, Y.; Huang, Z.; Huang, R.; Jiang, J.; Lu, X. Texture image classification based on multi-feature extraction and SVM classifie. Comput. Appl. Softw. 2011, 28, 22–25. [Google Scholar]
  13. Xu, Y.; Gu, Y.; Peng, D.L.; Liu, J. Target Recognition Based on DRGAN and Support Vector Machine in Synthetic Aperture Radar Images. Opt. Precis. Eng. 2020, 28, 727–735. [Google Scholar]
  14. Cui, Z.M.; Deng, B.; Deng, X.J.; Gao, J.K.; Lu, B.; Chen, P.; He, Y.; Qin, Y.L.; Chen, B.B.; Liu, Q.; et al. 340 GHz sparse MIMO array real-time 3-D imaging system. J. Infrared Millim. Wave 2017, 36, 102–106. [Google Scholar]
  15. Gao, J.K.; Deng, B.; Qin, Y.L. Point Cloud and 3-D Surface Reconstruction Using Cylindrical Millimeter-Wave Holography. IEEE Trans. Instrum. Meas. 2019, 68, 4765–4778. [Google Scholar] [CrossRef]
  16. Lv, N.; Chen, C.; Qiu, T. Deep Learning and Superpixel Feature Extraction Based on Sparse Autoencoder for Change Detection in SAR Images. IEEE Trans. Ind. Inform. 2018, 14, 5530–5538. [Google Scholar] [CrossRef]
  17. Lundén, J.; Koivunen, V. Deep Learning for HRRP-based Target Recognition in Multistatic Radar Systems. In Proceedings of the Radar Conference, Philadelphia, PA, USA, 2–6 May 2016. [Google Scholar]
  18. Amin, M.G.; Ahmad, F.; Zhang, Y.D.; Boashash, B. Micro-Doppler characteristics of elderly gait patterns with walking aids. In Proceedings of the SPIE DEFENSE + SECURITY, Baltimore, MD, USA, 20–24 April 2015. [Google Scholar]
  19. Seyfioglu, M.S.; Gurbuz, S.Z. Deep Neural Network Initialization Methods for Micro-Doppler Classification with Low Training Sample Support. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2462–2466. [Google Scholar] [CrossRef]
  20. Haefner, B.; Peng, S.; Verma, A.; Queau, Y.; Cremers, D. Photometric Depth Super-Resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 99, 1. [Google Scholar] [CrossRef] [Green Version]
  21. Shamsolmoali, P.; Zareapoor, M.; Jain, D.K.; Jain, V.K.; Jie, Y. Deep convolution network for surveillance records super-resolution. Multimed. Tools Appl. 2019, 78, 23815–23829. [Google Scholar] [CrossRef]
  22. Sidorov, O.; Hardeberg, J.Y. Deep Hyperspectral Prior: Single-Image Denoising, Inpainting, Super-Resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, (ICCVW), Seoul, Korea, 27–28 October 2019. [Google Scholar]
  23. Xu, J.; Zhang, L.; Zuo, W.; Zhang, D.; Feng, X.C. Patch Group Based Nonlocal Self-Similarity Prior Learning for Image Denoising. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  24. Rueckert, D.; Schnabel, J.A. Model-Based and Data-Driven Strategies in Medical Image Computing. Proc. IEEE 2019, 108, 1–15. [Google Scholar] [CrossRef] [Green Version]
  25. Archirapatkave, V.; Sumilo, H.; See, S.; Achalakul, T. GPGPU Acceleration Algorithm for Medical Image Reconstruction. In Proceedings of the IEEE Ninth International Symposium on Parallel & Distributed Processing with Applications, Busan, Korea, 26–28 May 2011. [Google Scholar]
  26. Wang, L.; Tang, J.; Liao, Q. A Study on Radar Target Detection Based on Deep Neural Networks. IEEE Sens. Lett. 2019, 3, 1–4. [Google Scholar] [CrossRef]
  27. Pei, J.; Huang, Y.; Huo, W.; Zhang, Y.; Yang, J. Target Aspect Identification in SAR Image: A Machine Learning Approach. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018. [Google Scholar]
  28. Gurbuz, S.Z.; Amin, M.G. Radar-Based Human-Motion Recognition with Deep Learning: Promising Applications for Indoor Monitoring. IEEE Signal Process. Mag. 2019, 36, 16–28. [Google Scholar] [CrossRef]
  29. Seyfioglu, M.S.; Erol, B.; Gurbuz, S.Z.; Amin, M.G. DNN Transfer Learning from Diversified Micro-Doppler for Motion Classification. IEEE Trans. Aerosp. Electron. Syst. 2018, 55, 2164–2180. [Google Scholar] [CrossRef] [Green Version]
  30. Seyfioglu, M.S.; Ozbayoglu, A.M.; Gurbuz, S.Z. Deep Convolutional Autoencoder for Radar-Based Classification of Similar Aided and Unaided Human Activities. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 1709–1723. [Google Scholar] [CrossRef]
  31. Bai, X.; Hui, Y.; Wang, L.; Zhou, F. Radar-Based Human Gait Recognition Using Dual-Channel Deep Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9767–9778. [Google Scholar] [CrossRef]
  32. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  33. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal Segment Networks: Towards Good Practices for Deep Action Recognition; Springer: Cham, Germany, 2016. [Google Scholar]
  34. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D Convolutional Neural Networks for Human Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 221–231. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Zhao, M.M.; Li, T.H.; Mohammad, A.A. Through-Wall Human Pose Estimation Using Radio Signals. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  36. Li, T.; Fan, L.; Zhao, M.; Liu, Y.; Katabi, D. Making the Invisible Visible: Action Recognition through Walls and Occlusions. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
Figure 1. Geometric diagram of the relationship between the direction of arrival and array structure.
Figure 1. Geometric diagram of the relationship between the direction of arrival and array structure.
Sensors 20 05466 g001
Figure 2. Radar system: (a) Structure of radar system; (b) Antenna array element and virtual aperture.
Figure 2. Radar system: (a) Structure of radar system; (b) Antenna array element and virtual aperture.
Sensors 20 05466 g002
Figure 3. Point cloud collection.
Figure 3. Point cloud collection.
Sensors 20 05466 g003
Figure 4. Diagram of convolution process [32]: (a) 2-D convolution; (b) 2-D convolution on multiple frames; (c) 3-D convolution.
Figure 4. Diagram of convolution process [32]: (a) 2-D convolution; (b) 2-D convolution on multiple frames; (c) 3-D convolution.
Sensors 20 05466 g004
Figure 5. Structure of multi-channel three-dimensional convolution neural network (MC-3DCNN).
Figure 5. Structure of multi-channel three-dimensional convolution neural network (MC-3DCNN).
Sensors 20 05466 g005
Figure 6. Experimental scenario: (a) Schematic diagram of experimental scene; (b) Real experimental scene.
Figure 6. Experimental scenario: (a) Schematic diagram of experimental scene; (b) Real experimental scene.
Sensors 20 05466 g006
Figure 7. Flow of data processing.
Figure 7. Flow of data processing.
Sensors 20 05466 g007
Figure 8. The structure of single-channel three-dimensional convolutional neural network (SC-3DCNN).
Figure 8. The structure of single-channel three-dimensional convolutional neural network (SC-3DCNN).
Sensors 20 05466 g008
Figure 9. Training loss during cross-training.
Figure 9. Training loss during cross-training.
Sensors 20 05466 g009
Figure 10. Training accuracy during cross-training.
Figure 10. Training accuracy during cross-training.
Sensors 20 05466 g010
Table 1. Direction of arrival (DOA) estimation method.
Table 1. Direction of arrival (DOA) estimation method.
For :   θ   from 9 0 o   to   9 0 o
calculate y α = a H ( α ) X ( n ) ;
end for θ = max α y α
Table 2. Samples and labels.
Table 2. Samples and labels.
Category of ActionLabel of ActionNumber of Samples (Frame)
normal walking040,000
jogging140,000
lame walking240,000
squatting down and standing up340,000
Table 3. Capacity of the dataset.
Table 3. Capacity of the dataset.
DatasetNumber (Frame)
Dataset 110,000 × 4
Dataset 210,000 × 4
Dataset 310,000 × 4
Dataset 410,000 × 4
Table 4. Cross-training and dataset partition.
Table 4. Cross-training and dataset partition.
Cross-ValidationDataset 1Dataset 2Dataset 3Dataset 4
CV_1TrainingTrainingTrainingTesting
CV_2TrainingTrainingTestingTraining
CV_3TrainingTestingTrainingTraining
CV_4TestingTrainingTrainingTraining
Table 5. Recognition results of cross validation.
Table 5. Recognition results of cross validation.
Accuracy (%)CV_1CV_2CV_3CV_4Average Accuracy
Category
Jogging95.2089.8094.6090.4092.50
Normal walking90.4095.2096.8089.6093.00
Lame walking81.6094.2089.6085.6087.75
Squatting down and standing up94.8092.5089.8093.6092.68
Table 6. Recognition results for different networks.
Table 6. Recognition results for different networks.
Accuracy (%)CV_1CV_2CV_3CV_4Average Accuracy
Network
SC-3DCNN84.2089.8081.6089.6086.30
MC-3DCNN90.5092.9392.7089.8093.00

Share and Cite

MDPI and ACS Style

Jiang, X.; Zhang, Y.; Yang, Q.; Deng, B.; Wang, H. Millimeter-Wave Array Radar-Based Human Gait Recognition Using Multi-Channel Three-Dimensional Convolutional Neural Network. Sensors 2020, 20, 5466. https://doi.org/10.3390/s20195466

AMA Style

Jiang X, Zhang Y, Yang Q, Deng B, Wang H. Millimeter-Wave Array Radar-Based Human Gait Recognition Using Multi-Channel Three-Dimensional Convolutional Neural Network. Sensors. 2020; 20(19):5466. https://doi.org/10.3390/s20195466

Chicago/Turabian Style

Jiang, Xinrui, Ye Zhang, Qi Yang, Bin Deng, and Hongqiang Wang. 2020. "Millimeter-Wave Array Radar-Based Human Gait Recognition Using Multi-Channel Three-Dimensional Convolutional Neural Network" Sensors 20, no. 19: 5466. https://doi.org/10.3390/s20195466

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop