You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

1 July 2020

Hyperparameter Optimization Method Based on Harmony Search Algorithm to Improve Performance of 1D CNN Human Respiration Pattern Recognition System

,
and
1
Department of Computer Engineering, Gachon University, Seongnam 13120, Korea
2
Department of Energy IT, Gachon University, Seoongnam 13120, Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Section Intelligent Sensors

Abstract

In this study, we propose a method to find an optimal combination of hyperparameters to improve the accuracy of respiration pattern recognition in a 1D (Dimensional) convolutional neural network (CNN). The proposed method is designed to integrate with a 1D CNN using the harmony search algorithm. In an experiment, we used the depth of the convolutional layer of the 1D CNN, the number and size of kernels in each layer, and the number of neurons in the dense layer as hyperparameters for optimization. The experimental results demonstrate that the proposed method provided a recognition rate for five respiration patterns of approximately 96.7% on average, which is an approximately 2.8% improvement over an existing method. In addition, the number of iterations required to derive the optimal combination of hyperparameters was 2,000,000 in the previous study. In contrast, the proposed method required only 3652 iterations.

1. Introduction

Existing ultra-wideband (UWB) radar sensors have been used for shortrange communication. However, recently, many studies using this sensor have been conducted in various fields that require noncontact signal pattern analysis technology. Among them, in the healthcare field, a technique to analyze the bio-signal data of a person obtained using a UWB sensor is being actively researched. This technology is intended to analyze signal patterns, e.g., the respiration, pulse, and movement of a person, obtained from a UWB sensor and use these signal patterns to diagnose diseases or improve quality of life based on the analyzed data.
Various machine learning techniques have been studied as a method to recognize and analyze human bio-signal patterns. Recently, as deep learning technologies have been researched, attention has been placed on improving performance, e.g., by improving the recognition rate [1,2,3,4,5].
Previously, the pattern recognition method for UWB sensor signals was researched primarily on pulse or apnea pattern detection based on machine learning [6,7,8,9,10]. Since then, as ANN(Artificial Neural Network)-based AI(Artificial Intelligence) technology has emerged, methods using this technique have been investigated; however, in such methods, preprocessing steps are required to construct learning data of a specific pattern. Recently, deep learning models based on CNN have emerged to solve these problems.
In such methods, features can be extracted and learned in the same process without preprocessing the data. Thus, various pattern recognition studies using CNN models have been conducted recently [11,12,13].
Generally, a CNN is used for object detection and recognition from an image; thus, it primarily comprises input data structures in two dimensions; however, one-dimensional signal data can also be used as input. Therefore, when one-dimensional signal data are input, convolution is performed using a one-dimensional vector kernel. The respiratory signal of a human obtained from a UWB sensor is also one-dimensional data; therefore, it is possible to learn and recognize the patterns of respiratory signal by implementing a 1D CNN model. However, generally, CNN performance is greatly affected by the structural design of each layer and hyperparameter values.
Various image processing studies have demonstrated that it is possible to improve performance with little adjustment of the depth and structure of the layer and the number and size of kernels in each layer based on previously studied neural networks and hyperparameters. However, in fields with few related studies, it is necessary to design an optimal neural network via repeated experiments with various hyperparameter combinations using a simple neural network [14,15,16,17].
In previous 1D CNN-based respiration pattern recognition studies, a specific range for hyperparameters is determined in advance for each layer, and experiments are performed repeatedly using a combination of all hyperparameters in this range, and the best combination is selected. In this case, the total number of hyperparameter combinations is 2,000,000; however, this method is limited relative to finding a hyperparameter with better performance because the combination is generated using specific values in the specified range.
To find a combination of hyperparameters that yields improved performance, the search must be performed in a wider range. In this case, experimenting with all combinations is very inefficient because it takes a lot of time. Therefore, it is necessary to study how to find an optimized combination of hyperparameters that provides good performance in less time. There are various algorithms for the optimization method, but among them, the harmony search (HS) algorithm is designed with a structure that is easy to hybridize with other technologies than the traditional optimization methods, namely the genetic algorithm (GA) or particle swarm optimization (PSO), and is an algorithm that efficiently searches for solutions [18,19,20,21,22,23,24].
Therefore, in this paper, we propose a hyperparameter optimization method that yields better performance in less time than the existing method [25] by applying the HS algorithm to the existing 1D CNN model with UWB radar sensor for respiration pattern recognition. A comparative analysis demonstrates that the proposed method outperforms existing methods.

3. Proposed Method

This paper proposes a method to optimize hyperparameters to improve the recognition rate in a 1D CNN neural network configured to recognize respiratory signal patterns from UWB sensor data. In the proposed method, first, the required hyperparameters are set as target variables to be optimized, and HM comprises hyperparameters and objective functions.
Next, the parameters to be used by the HS algorithm are derived, and in the final step, the optimized hyperparameter combination is found by applying the HS algorithm to the 1D CNN respiration pattern recognition system.
The goal of the proposed method is to design an optimal system that integrates the 1D CNN and HS algorithms in order to find an optimal combination of hyperparameters that yields an excellent respiratory pattern recognition rate in the 1D CNN system.
To demonstrate that the proposed method is superior to the existing method, an experiment was conducted under the same conditions as the existing method, and the respiratory signal pattern recognition rate and number of iterations required to find the optimal combination of hyperparameters were compared.

3.1. Design of Harmony Memory and Object Function

To optimize the 1D CNN’s hyperparameters using the HS algorithm, first, the HM must be configured and the objective function must be designed. Here, hyperparameters that significantly influence optimization of the 1D CNN are selected. The hyperparameters used in the proposed method are presented in Table 3, and these hyperparameters affect the recognition rate according to the characteristics of the input data. The hyperparameters selected for the proposed method are KS, KC, and DLNC (Table 4), where KS is the kernel size in the convolutional layer, KC is the number of kernels used in the convolutional layer, and DLNC is the number of neurons in the dense layer.
Table 4. Hyperparameters to be optimized for 1D CNN.
KS and KC exist for each convolutional layer. Thus, if the convolutional layer depth (CLD) increases, KS and KC will also increase.
In the 1D CNN, the dense layer always comprises two layers regardless of the CLD.
The HM is designed to have the convolutional layer as much as CLD (Figure 4) for the existing 1D CNN respiration pattern recognition system. As the value of CLD increases, the number of hyperparameters for KS and KC also increases, and the number of DLNC is fixed to two.
Figure 4. Layout of 1D CNN model used in proposed method.
The structure of the hyperparameters of HM, i.e., the optimization target, is expressed in matrix form as follows.
H M = k s 1 1   k c 1 1     k s 1 c l d   k c 1 c l d   d l n c 1 1   d l n c 1 2   R 1 k s 2 1   k c 2 1     k s 2 c l d   k c 2 c l d   d l n c 2 1   d l n c 2 2   R 2 k s 3 1   k c 3 1     k s 3 c l d   k c 3 c l d   d l n c 3 1   d l n c 3 2   R 3               k s h m s 1   k c h m s 1     k s h m s c l d   k c h m s c l d   d l n c h m s 1   d l n c h m s 2   R h m s
H P i = k s i 1 ,   k c i 1 ,   ,   k s i c l d ,   k c i c l d ,   d l n c i 1 ,   d l n c i 2 ,   i = 1 ,   ,   h m s
The objective function used to evaluate the performance of each hyperparameter combination in HM is set to recognition rate R i .
In other words, in each execution step of the 1D CNN (Figure 5), hyperparameters are substituted, and weights ω and bias b are updated while the feedforward and backpropagation processes are performed with the learning data. Learning for a single selected parameter combination is completed when the loss function’s conditions are satisfied.
Figure 5. Process for hyperparameter optimization between HM and 1D CNN.
Next, the recognition rate using the verification data is calculated and stored at the location of the objective function in the new HM area ( H M n e w ), and the HM is updated by comparing the R n e w of H M n e w to R h m s of HM, which is the lowest priority of HM. When HM is updated, alignment is performed for recognition rate R i , i = 1 , , h m s in HM.
When initializing HM, it is recommended to set HMS to a sufficiently large value (in this study, it is set to 1000). The reason may be described as an example of searching global maxima, which has the highest recognition rate for several CNN hyperparameters.
If a global search is performed through multiple random selections during the HM initialization process, and iteration is performed based on this HM, the HM creation with sufficiently large HMS can be optimized with much fewer Iterations than when HMS is small. Thus, the initialization process requires more time than when HMS is small; however, it is possible to significantly reduce the time required to reach global maxima by iteration. In contrast, if a small HMS value is used, the memory pool will also be small; thus, there is a high probability of falling into local maxima or performing a very high number of iterations.

3.2. Selecting Optimal Hyperparameters for HS Algorithm

The HS algorithm requires hyperparameters to operate HS itself in addition to optimization for a 1D CNN, and, to obtain good results, the hyperparameters should be selected with appropriate values determined through experimentation. Therefore, experiments were performed by constructing various hyperparameter groups, and appropriate hyperparameters were set based on the results. The hyperparameters that have significant influence on the performance of the HS algorithm in the proposed method are HMCR, PAR, and MPAP. Here, as the HMCR value increases, the probability of using a combination of existing hyperparameters stored in HM increases. Pitch is adjusted by PAR, and a higher PAR value results in a higher probability of pitch adjustment. Here, pitch is adjusted by selecting a random value within the MPAP range. In other words, a higher HMCR value results in a higher probability that the PAR and MPAP values will be used frequently. Therefore, HMCR should be set to an appropriate value to achieve efficient optimization without reaching local maxima. Conversely, if the value of the HMCR is low, optimization efficiency is reduced because this is similar to global search using normal random selection [37].
Considering the characteristics of each hyperparameter, we present two cases where optimization becomes inefficient with the local maxima or frequent random selection in the following.
  • Local maxima problem: This problem occurs when using a high HMCR value and low MPAP value regardless of the PAR value. This problem frequently recalls existing memory due to the high HMCR value, and the pitch is adjusted by PAR when creating a new harmony. Here, if the MPAP value is too low, falling into local maxima may occur.
  • Random selection problem: This problem occurs when the HMCR value is low. As a result, the probability of random selection occurring increases, which can yield efficiency that is similar to a typical random search method. This problem may also occur when the HMCR, PAR, and MPAP values are all high. In particular, if MPAP is too large, even if the existing HM is used, the pitch adjustment value becomes too large, which has the same effect as random selection.
Due to these problems, it is inefficient to use HS algorithm variables that are close to 0 or 1. To minimize this problem, HMCR is set to a random value of 0.05 units between 0.5 and 0.7. Thus, when creating a new harmony, it is recommended to recall the existing HM with a probability of 50% or greater.
In addition, PAR was selected with a random value of 0.05 units in the range of 0.6 to 0.8 such that pitch adjustment occurred frequently. Here, the PAR value is set to a slightly higher value than existing method 2 in Table 5. Thus, MPAP must be set to an appropriate value to minimize the occurrence of local maxima and random selection problems.
Table 5. Number of iterations required to reach 95% recognition rate for each hyperparameter combination to find appropriate hyperparameter value for HS algorithm.
The MPAP value was set in the range 10–18 through experiments. The hyperparameter combination and performance of the HS algorithms among studies are compared in Table 5. Note that the performance comparison is expressed as the number of iterations required to reach a 95% recognition rate for each combination of hyperparameters.
As can be seen, the hyperparameter combination of the proposed method reached a recognition rate of 95% with fewer iterations than the hyperparameter combination used in other studies. The hyperparameter values for the existing method 1 and 2 shown in Table 5 are used in different optimization problems [38,39].
As shown in Table 5, when applying the combination of hyperparameters derived from existing methods 1 and 2 to the hyperparameter optimization problem of the proposed ID CNN, approximately 4,357 to 5,912 iterations were required reach the target recognition rate of 95%. When the proposed method was applied to the same environment, it was possible to derive a combination of hyperparameters that reached the target recognition rate by performing approximately 2,011 iterations. Since MPAP is applied to all hyperparameters used in the proposed 1D CNN (i.e., KS, KC, and DLNC), a range of hyperparameters was derived experimentally to have a larger value than the hyperparameters suggested by the existing method. This means that each hyperparameter must be set appropriately for the target optimization problem. The experimentally derived HMCR, PAR, and MPAP values were used in the HS algorithm to optimize the KS, KC, and DLNC hyperparameters of the proposed 1D CNN.

3.3. Optimization of 1D CNN Hyperparameters Using HS Algorithm

Figure 6 shows the hyperparameter optimization process to improve the recognition rate of respiration patterns in the 1D CNN using the HS algorithm with the HM configuration suggested in Section 3.1 and the HS hyperparameter range suggested in Section 3.2.
Figure 6. 1D CNN hyperparameter optimization process using HS algorithm.
The new harmony generation and harmony memory update processes shown in Figure 6 are detailed in Algorithm 1. Here, a combination of hyperparameters as much as HMS is generated and filled in HM. Then, the HM was updated only when the given condition was satisfied. H m a x is a combination of hyperparameters that gives the best performance among the harmonies in HM. If the HM is updated but H m a x has not been updated, variable Δ I (count of non-update for H m a x ) is increased by one. If the Δ I value reaches a given specific value (threshold k), optimization is no longer occurring in HM, and the HM update operation is completed even if the set number of iterations has not been reached.
Finally, the hyperparameter combination H P m a x , which has the highest recognition rate in HM, is derived as the optimal combination of hyperparameters.
Algorithm 1: Algorithm for Optimizing Hyperparameters of 1D CNN using HS
(1)
Set the parameters for Harmony Search (HS) algorithm
H M S = 1000 , H M C R = r a n d 0.5 ,   0.7 ,     P A R = r a n d 0.6 ,   0.8 ,   M P A P = r a n d 10 ,   18 ,
k = threshold   for   optimization   completion   condition
KS-MIN = min v a l u e o f   k e r n e l   s i z e ,   KS-MAX = m a x   v a l u e   o f   k e r n e l   s i z e
KC-MIN = min v a l u e o f   k e r n e l   c o u n t ,   KS-MAX = m a x   v a l u e   o f   k e r n e l   c o u n t
DLNC-MIN = min v a l u e o f   d e n s e   l a y e r   n e u r o n   c o u n t
DLNC-MAX = max v a l u e o f   d e n s e   l a y e r   n e u r o n   c o u n t
 
(2)
Initialization of Harmony Memory (HM).
    Begin Loop (i):
          for(i = 0; i < HMS; i++)
Set the value of each hyperparameter H P i   through random selection
     Begin Loop (d):
        for (d = 1; d <= CLD; d++)
               k s i d = r a n d (KS-MIN, KS-MAX)
             k c i d = r a n d (KC-MIN, KC-MAX)
      End Loop (d)
       d l n c i 1 = r a n d (DLNC-MIN, DLNC-MAX)
       d l n c i 2 = r a n d (DLNC-MIN, DLNC-MAX)
 
Learn with the hyperparameter H P i .
* Use training data set X
* Update w, b by backpropagation
         ω ,   b L   H P i ,   X
 
If the Loss function is not satisfied, go to ②.
 
Calculate the recognition rate R i   by applying H P i   to the verification data Y.
       R i V ω ,   H P i ,   Y
 
Store the values of hyperparameters H P i and recognition rate R i in the i index of HM
 End Loop (i)
 
(3)
Optimization of 1D CNN Hyperparameter ( H P ).
Initialization of variable I p r e v   , I c u r r   ( I p r e v = 0 , I c u r r = 0 )
Begin Loop (j):
   for(j = 0; j<Iteration; j++)
     
Each algorithm is executed according to the following conditions
  if rand(0, 1) < HMCR then
     * Create new harmony H P n e w using the hyperparameter H P r a n d   randomly
       selected from the existing HM.
             H P r a n d = H P i ,   i = rand 0 ,   H M S   a n d   H P i H M
             H P r a n d = k s r a n d 1 ,     k c r a n d 1 ,       ,     k s r a n d c l d ,     k c r a n d c l d ,     d l n c i 1 ,     d l n c i 2
 
        if rand(0, 1) < PAR then
          * Adjust the pitch for the new harmony H P n e w as follows.
           Begin Loop (d):
               for (d = 1; d <= CLD; j++)
                  k s n e w d = k s r a n d d + M P A P
                  k c n e w d = k c r a n d d + M P A P
           End Loop (d)
            d l n c n e w 1 = d l n c r a n d 1 + M P A P
            d l n c n e w 2 = d l n c r a n d 2 + M P A P
        else if
          * Create a new harmony without adjusting the pitch of the existing harmony.
                                              H P n e w = H P r a n d
  else if
    * Set values for each hyperparameter H P n e w   through Random Selection.
        Begin Loop (d):
           for (d = 1; d <= CLD; j++)
               k s n e w d = r a n d (KS-MIN, KS-MAX)
               k c n e w d = r a n d (KC-MIN, KC-MAX)
      End Loop (d)
       d l n c n e w 1 = r a n d (DLNC-MIN, DLNC-MAX)
       d l n c n e w 2 = r a n d (DLNC-MIN, DLNC-MAX)
  
  Learn with the new harmony H P n e w .
    * Use training data set X.
    * Update w, b by backpropagation
                     ω ,   b L   H P n e w ,   X
  
If the Loss function is not satisfied, go to ②.
  
Calculate the recognition rate R n e w   by applying H P n e w to the verification data Y.
                     R n e w V ω ,   H P i ,   Y
  
Compare the recognition rate R n e w with R m i n , the minimum recognition rate in harmony memory. And if R n e w is greater than R m i n , remove H m i n from harmony and update to H n e w .
  
After sorting all harmony H i in descending order based on the recognition rate R of HM, perform the following procedure to find the optimal combination of hyperparameters early.
   if R n e w > R m a x then
        I p r e v = I c u r r = j , Δ I = 0
   else
          I c u r r = j , Δ I = I c u r r I p r e v  
         if Δ I k then
              break Loop (j)
         else
            continue Loop (j)
End Loop (j)
 
(4)
Obtain the hyperparameter combination H P m a x with the highest recognition rate (R) from the harmony memory HM.

4. Experiments

4.1. Learning and Test Dataset

The UWB respiration signal dataset used for learning in this study is the same dataset used in a previous study to compare the performance to the exiting method [35].
The dataset includes five signal patterns, i.e., general respiration, slow respiration, tachypnea, apnea, and movement. Here, 10 people participated in the experiment. Table 6 shows samples of the respiratory signal data extracted for each pattern type.
Table 6. Samples of five respiration patterns collected from participants.
The experimental data comprised a total of 2500 pieces (i.e., 10 peoples × 5 types × 50 pieces) by extracting 50 pieces for each type from each participant’s respiratory signal and 1 piece is a respiratory signal for 10 seconds. Here, 1500 pieces were used for learning, and the remaining 1000 were used for testing. In addition, 1500 pieces were divided into training and verification datasets at a ratio of 8:2. The two datasets comprised randomly selected data from 1500 data for each learning process.

4.2. Comparison of Recognition Rate between Proposed and Existing Methods

The range of hyperparameters used in the HS algorithm and 1D CNN for the proposed method were set as shown in Table 7. Here, KS, KC, and DLNC were approximately two times wider than in the previous studies.
Table 7. Hyperparameters used in proposed method and range of each parameter value.
For the experiment with the proposed method, the hyperparameters for the HS algorithm and 1D CNN were set as shown in Table 7, and number of iterations was set to 10,000 to perform repeated experiments. Here, if we calculate the number of all cases of the combination in the range of KS1, KS2, KS3, KC1, KC2, KC3, DLNC1, and DLNC2 used as hyperparameters of 1D CNN, the number of KSi cases is 40 (odd numbers from 3 to 81), KCi is 1009 (16 to 1024), and DLNCi is 3841 (256 to 4096), so the total number of cases is 9.69934 × 1020 (= 403 × 10093 × 38412).
Figure 7 shows the process of optimizing the hyperparameter combination of the 1D CNN. As can be seen, the recognition rate gradually is increased as iteration progressed. For H P m a x , the recognition rate reached 96.7% after 3652 iterations, and optimization was terminated after 3852 iterations because it had not been updated continuously for over Δ I (200 iterations) repetitions.
Figure 7. Trend of recognition rate in 1D CNN hyperparameter optimization process using proposed method.
The combination of hyperparameters and recognition rate R i for each iteration step is shown in Table 8, and it is possible to derive the optimal hyperparameter combination H P m a x at the location (i.e., recognition rate of H m a x has maximum values) where optimization was completed.
Table 8. Hyperparameter combination and recognition rate when H m a x was updated.
Comparing the above result with the existing method (Table 9), the final recognition rate and total number of iterations performed until searching for the optimal hyperparameter can be represented. In the existing method, a set of hyperparameters was constructed using a grid search method, and the recognition rate was tested for 2,000,000 combinations to optimize hyperparameters.
Table 9. Number of iterations required for hyperparameter optimization and recognition rate using optimized hyperparameters.
However, with the proposed method, it was possible to obtain a better recognition rate result in only 3652 iterations. In addition, with the existing method, the values of each optimized hyperparameter used in the 1D CNN were derived in a specific range. However, with the proposed method, each optimized hyperparameter was derived as a single value.
The recognition rate of 1000 test datasets was tested with optimized hyperparameters and compared to the existing method. The results are shown in Figure 8. By comparing all patterns, we found that the proposed method (Figure 8b) increased the recognition rate by approximately 2.8% compared to the existing method (Figure 8a).
Figure 8. Comparison of recognition rate by respiration pattern for existing and proposed methods. (a) Recognition rate of existing method [28]; (b) Recognition rate of proposed method.

5. Conclusions

In this paper, have proposed a method to optimize the hyperparameters required to improve the recognition rate in a 1D CNN-based UWB respiration pattern recognition study, and we have proven the validity of the proposed method by comparing it with an existing method. Conventionally, to optimize the hyperparameters of a 1D CNN, a specific range is determined for each hyperparameter, and a combination of the range values is constructed. Then, repeated experiments are performed using the combination in all cases. However, this method requires many experiments without special rules to find a better hyperparameter. Thus, it takes a long time to find the optimal hyperparameter combination, which is tedious and inefficient.
In contrast, the proposed hyperparameter optimization method is implemented to have scalability according to the CLD by integrating the HS algorithm into the 1D CNN and focusing on deriving an optimal combination of hyperparameters with fewer iterations.
The proposed method was tested with CLD value of 3 to maintain the same environment as the existing method. Experiments were conducted using the same dataset used in a previous study using the optimal hyperparameter combination derived by the proposed method. As the result, the recognition rate of the 1D CNN’s respiration pattern was approximately 96.7%, which is an improvement of approximately 2.8% over the previous study.
In addition, to derive hyperparameters, the number of experimental repetitions was 2,000,000 with the existing method. In contrast, the proposed method could derive an optimal combination of hyperparameters in only 3652 iterations.
This demonstrates the excellence of the proposed method and suggests that the time required to derive an optimal combination of hyperparameters can be reduced by up to approximately 548 times compared to the existing method.

Author Contributions

S.-H.K. performed calculation and wrote the paper; G.-T.H. and Z.W.G. supervised the research direction; G.-T.H. obtained the funds. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the GRRC program of Gyeonggi province (GRRC-Gachon2017(B02), Bio-Data Construction and Prediction based on Artificial Intelligence).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lazaro, A.; Girbau, D.; Villarino, R. Analysis of vital signs monitoring using an IR-UWB radar. Prog. Electromagn. Res. 2010, 100, 265–284. [Google Scholar] [CrossRef]
  2. Cho, H.; Yoon, S.M. Divide and Conquer-Based 1D CNN Human Activity Recognition Using Test Data Sharpening. Sensors 2018, 18, 1055. [Google Scholar] [CrossRef]
  3. Kravchik, M.; Shabtai, A. Detecting Cyber Attacks in Industrial Control Systems Using Convolutional Neural Networks. In Proceedings of the 2018 Workshop on Cyber-Physical Systems Security and PrivaCy—CPS-SPC ’18, Toronto, ON, Canada, 15–19 October 2018; pp. 72–83. [Google Scholar]
  4. Kim, T.; Lee, J.; Nam, J. Sample-Level CNN Architectures for Music Auto-Tagging Using Raw Waveforms. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 366–370. [Google Scholar]
  5. Lee, S.-M.; Yoon, S.M.; Cho, H. Human activity recognition from accelerometer data using Convolutional Neural Network. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju Island, Korea, 13–16 February 2017; pp. 131–134. [Google Scholar]
  6. Kim, M.G.; Pan, S.B. Deep Learning Based on 1-D Ensemble Networks Using ECG for Real-Time User Recognition. IEEE Trans. Ind. Inform. 2019, 15, 5656–5663. [Google Scholar] [CrossRef]
  7. Tran, V.P.; Al-Jumaily, A.; Islam, S. Doppler Radar-Based Non-Contact Health Monitoring for Obstructive Sleep Apnea Diagnosis: A Comprehensive Review. Big Data Cogn. Comput. 2019, 3, 3. [Google Scholar] [CrossRef]
  8. Javaid, A.Q.; Noble, C.M.; Rosenberg, R.; Weitnauer, M.A. Towards Sleep Apnea Screening with an Under-the-Mattress IR-UWB Radar Using Machine Learning. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015; pp. 837–842. [Google Scholar]
  9. Huang, X.; Sun, L.; Tian, T.; Huang, Z.; Clancy, E. Real-time non-contact infant respiratory monitoring using UWB radar. In Proceedings of the 2015 IEEE 16th International Conference on Communication Technology (ICCT), Hangzhou, China, 18–21 October 2015; pp. 493–496. [Google Scholar]
  10. Fedele, G.; Pittella, E.; Pisa, S.; Cavagnaro, M.; Canali, R.; Biagi, M. Sleep-Apnea Detection with UWB Active Sensors. In Proceedings of the IEEE International Conference on Ubiquitous Wireless Broadband (ICUWB), Montreal, QC, Canada, 4–7 October 2015. [Google Scholar]
  11. Zhang, Q.; Zhou, D.; Zeng, X. HeartID: A Multiresolution Convolutional Neural Network for ECG-Based Biometric Human Identification in Smart Health Applications. IEEE Access 2017, 5, 11805–11816. [Google Scholar] [CrossRef]
  12. Yildirim, O.; Baloglu, U.B.; Acharya, U.R. A deep convolutional neural network model for automated identification of abnormal EEG signals. Neural Comput. Appl. 2018, 1–12. [Google Scholar] [CrossRef]
  13. Yildirim, O.; Pławiak, P.; Tan, R.-S.; Acharya, U.R. Arrhythmia detection using deep convolutional neural network with long duration ECG signals. Comput. Biol. Med. 2018, 102, 411–420. [Google Scholar] [CrossRef]
  14. Andonie, R.; Florea, A.-C. Weighted Random Search for CNN Hyperparameter Optimization. Int. J. Comput. Commun. Control. 2020, 15, 432–445. [Google Scholar] [CrossRef]
  15. Passricha, V.; Aggarwal, R.K. PSO-based optimized CNN for Hindi ASR. Int. J. Speech Technol. 2019, 22, 1123–1133. [Google Scholar] [CrossRef]
  16. Wu, J.; Chen, X.-Y.; Zhang, H.; Xiong, L.-D.; Lei, H.; Deng, S.-H. Hyperparameter Optimization for Machine Learning Models Based on Bayesian Optimization. J. Electron. Sci. Technol. 2019, 17, 26–40. [Google Scholar]
  17. Bochinski, E.; Senst, T.; Sikora, T. Hyper-parameter optimization for convolutional neural network committees based on evolutionary algorithms. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3924–3928. [Google Scholar]
  18. Yadav, N.; Yadav, A.; Bansal, J.C.; Deep, K.; Kim, J.H. Harmony Search and Nature Inspired Optimization Algorithms; Advances in Intelligent Systems and Computing; Springer Nature Singapore pte Ltd.: Singapore, 2018. [Google Scholar]
  19. Peraza, C.; Valdez, F.; Castillo, O. A Harmony Search Algorithm Comparison with Genetic Algorithms. Adv. Intell. Inf. Database Syst. 2014, 574, 105–123. [Google Scholar] [CrossRef]
  20. Khan, L.; Mumtaz, S.; Khattak, A. Comparison of Genetic Algorithm and Harmony Search for Generator Maintenance Scheduling. Mehran Univ. Res. J. Eng. Technol. 2012, 31, 587–598. [Google Scholar]
  21. Zhang, T.; Geem, Z.W. Review of harmony search with respect to algorithm structure. Swarm Evol. Comput. 2019, 48, 31–43. [Google Scholar] [CrossRef]
  22. Saka, M.; Hasançebi, O.; Geem, Z.W. Metaheuristics in structural optimization and discussions on harmony search algorithm. Swarm Evol. Comput. 2016, 28, 88–97. [Google Scholar] [CrossRef]
  23. Geem, Z.W. Music-Inspired Harmony Search Algorithm and Its Experience-Based Derivative. New Phys. Sae Mulli 2017, 67, 608–614. [Google Scholar] [CrossRef]
  24. Geem, Z.W. Particle-swarm harmony search for water network design. Eng. Optim. 2009, 41, 297–311. [Google Scholar] [CrossRef]
  25. Kim, S.-H.; Geem, Z.W.; Han, G.-T. A Novel Human Respiration Pattern Recognition Using Signals of Ultra-Wideband Radar Sensor. Sensors 2019, 19, 3340. [Google Scholar] [CrossRef] [PubMed]
  26. Li, C.; Mak, P.-I.; Gomez-Garcia, R.; Chen, Y. Guest Editorial Wireless Sensing Circuits and Systems for Healthcare and Biomedical Applications. IEEE J. Emerg. Sel. Top. Circuits Syst. 2018, 8, 161–164. [Google Scholar] [CrossRef]
  27. Milici, S.; Lazaro, A.; Villarino, R.; Girbau, D.; Magnarosa, M. Wireless Wearable Magnetometer-Based Sensor for Sleep Quality Monitoring. IEEE Sens. J. 2018, 18, 2145–2152. [Google Scholar] [CrossRef]
  28. Li, C.; Lubecke, V.M.; Boric-Lubecke, O.; Lin, J. A Review on Recent Advances in Doppler Radar Sensors for Noncontact Healthcare Monitoring. IEEE Trans. Microw. Theory Tech. 2013, 61, 2046–2060. [Google Scholar] [CrossRef]
  29. Bernardi, P.; Cicchetti, R.; Pisa, S.; Pittella, E.; Piuzzi, E.; Testa, O. Design, Realization, and Test of a UWB Radar Sensor for Breath Activity Monitoring. IEEE Sens. J. 2013, 14, 584–596. [Google Scholar] [CrossRef]
  30. Fan, D.; Ren, A.; Zhao, N.; Yang, X.; Zhang, Z.; Shah, S.A.; Hu, F.; Abbasi, Q.H. Breathing Rhythm Analysis in Body Centric Networks. IEEE Access 2018, 6, 32507–32513. [Google Scholar] [CrossRef]
  31. Loughlin, P.C.; Sebat, F.; Kellett, J.G. Respiratory Rate: The Forgotten Vital Sign—Make It Count! Jt. Comm. J. Qual. Patient Saf. 2018, 44, 494–499. [Google Scholar] [PubMed]
  32. Brownlee, J. Deep Learning for Time Series Forecasting; Machine Learning Mastery Pty. Ltd.: Vermont Victoria, Australia, 2018; pp. 1–575. [Google Scholar]
  33. Dubey, S.R.; Chakraborty, S.; Roy, S.K.; Mukherjee, S.; Singh, S.K.; Chaudhuri, B.B. diffGrad: An Optimization Method for Convolutional Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 1–12. [Google Scholar] [CrossRef]
  34. Nalçakan, Y.; Ensari, T. Decision of Neural Networks Hyperparameters with a Population-Based Algorithm. In Proceedings of the Intelligent Tutoring Systems, Montreal, QC, Canada, 11–15 June 2018; pp. 276–281. [Google Scholar]
  35. Geem, Z.W. Music-Inspired Harmony Search Algorithm: Theory and Applications; Springer: Berlin, Germany, 2009. [Google Scholar]
  36. Abdel-Raouf, O.; Metwally, M.A.-B. A Survey of Harmony Search Algorithm. Int. J. Comput. Appl. 2013, 70, 17–26. [Google Scholar] [CrossRef]
  37. Maheri, M.R.; Narimani, M. An enhanced harmony search algorithm for optimum design of side sway steel frames. Comput. Struct. 2014, 136, 78–89. [Google Scholar] [CrossRef]
  38. Askarzadeh, A.; Rashedi, E.; Giannoccaro, I.; Patnaik, S. Harmony Search Algorithm. In Advances in Computational Intelligence and Robotics; IGI Global: Hershey, PA, USA, 2017; pp. 1–36. [Google Scholar]
  39. Rosa, G.; Papa, J.P.; Marana, A.N.; Scheirer, W.J.; Cox, D. Fine-Tuning Convolutional Neural Networks Using Harmony Search. In Natural Language Processing and Information Systems; Springer Science and Business Media LLC: Berlin, Germany, 2015; Volume 9423, pp. 683–690. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.