Next Article in Journal
Behavioral Analysis and Individual Tracking Based on Kalman Filter: Application in an Urban Environment
Next Article in Special Issue
The Biomechanical Mechanism of Upper Airway Collapse in OSAHS Patients Using Clinical Monitoring Data during Natural Sleep
Previous Article in Journal
Fully Automated and Robust Cable Tension Estimation of Wireless Sensor Networks System
Previous Article in Special Issue
Enhanced Monitoring of Sleep Position in Sleep Apnea Patients: Smartphone Triaxial Accelerometry Compared with Video-Validated Position from Polysomnography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Detection of K-Complexes Using the Cohen Class Recursiveness and Reallocation Method and Deep Neural Networks with EEG Signals

by
Catalin Dumitrescu
1,*,
Ilona-Madalina Costea
1,
Angel-Ciprian Cormos
1 and
Augustin Semenescu
2
1
Department Telematics and Electronics for Transports, University “Politehnica” of Bucharest, 060042 Bucharest, Romania
2
Department Engineering and Management for Transports, University “Politehnica” of Bucharest, 060042 Bucharest, Romania
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(21), 7230; https://doi.org/10.3390/s21217230
Submission received: 13 September 2021 / Revised: 17 October 2021 / Accepted: 27 October 2021 / Published: 30 October 2021

Abstract

:
Evoked and spontaneous K-complexes are thought to be involved in sleep protection, but their role as biomarkers is still under debate. K-complexes have two major functions: first, they suppress cortical arousal in response to stimuli that the sleeping brain evaluates to avoid signaling danger; and second, they help strengthen memory. K-complexes also play an important role in the analysis of sleep quality, in the detection of diseases associated with sleep disorders, and as biomarkers for the detection of Alzheimer’s and Parkinson’s diseases. Detecting K-complexes is relatively difficult, as reliable methods of identifying this complex cannot be found in the literature. In this paper, we propose a new method for the automatic detection of K-complexes combining the method of recursion and reallocation of the Cohen class and the deep neural networks, obtaining a recursive strategy aimed at increasing the percentage of classification and reducing the computation time required to detect K-complexes by applying the proposed methods.

1. Introduction

It is hard to explain the mechanism that triggers sleep. Science can only put forth some hypotheses, but there is still no certainty. Sleep can only be studied in its external manifestations, while the causes that generate it remain unknown. The mechanisms that maintain the state of wakefulness and the transition from wakefulness to sleep sparked the interest of scholars many centuries ago.
However, since the first theories were laid down by Alcmaeon of Croton and Aristotle, there has not been much progress in this field. It is only in the past 80 years that a review of old ideas about wakefulness and sleep has begun, largely due to the enhancement of research tools that have allowed for the precise exploration of deep brain structure and the recording of electrical brain activity down to the level of a single cell. Existing theories on sleep and wakefulness feature great diversity but can be grouped into two categories: humors-related, including particularly old and very old theories, and neural theories, specific to the modern age and based on experimental investigations.
The study of electroencephalograms (EEG) began in the 1930s as the first attempt at classifying human cerebral activities during sleep. Only after a long period of analytical interpretation were efforts eventually directed towards the analysis of the transient phenomena occurring during sleep. At present, the main objective lies in understanding the mechanisms that generate sleep and the psychological role of electrical activity in the brain.
The applicability of the method proposed in this article focuses on the analysis, detection, and classification of the K-complex, which is a transient brain wave in the microstructure of the electroencephalogram (EEG) of sleep. Detection of the K-complex brain wave is difficult because during sleep other nonstationary phenomena occur, including an increase in delta waves, which looks very similar to the K-complex. The K-complex investigation method proposed in this article allows for an automated system to detect and classify—with very good probability—the aperiodic K-complex that occurs in the second stage of EEG sleep. According to the literature, the K-complex has an important role in the study of sleep [1,2,3,4], the detection of diseases associated with sleep disorders [5,6,7,8,9,10,11,12,13,14,15,16,17], and recently has been suggested as a biomarker for predicting the occurrence of Parkinson’s [18,19,20,21,22,23,24,25,26,27] and Alzheimer’s diseases [28,29,30,31,32,33,34,35,36,37].
The results reported in a series of papers on the study of sleep electroencephalograms justify the interest in using time–frequency-pooled analyses that can characterize the transient phenomena that occur during the stages of sleep [38,39,40,41,42].
The two-dimensional representations in the time–frequency domain provide a powerful signal analysis tool that has the advantage of making it possible to highlight certain “hidden” properties of the signals. As far as the system of analysis is concerned, the main interest lies in analyzing signals at the lowest level, comparable to the noise made by the EEG apparatus. This is why time–frequency analyses should be performed on signals affected by noise, the signal/noise ratio being of particular importance for evaluating the parameters of the analyzed signal. It is this method of representation and analysis of transient signals that we have resorted to in order to develop a K-complex detector. The proposed methodology for the synthesis of imposed structure detectors has wider applicability, for example, in the processing of vocal signals.
Form recognition has become a growing field in both theory and application, being, on the one hand, an advanced form of information processing, and on the other, a component of artificial intelligence. Many mathematical methods have been proposed to solve the problem of form recognition, and have been grouped into two categories—statistical–decisional and syntactic–structural ones [43,44,45,46,47,48,49,50].
The theory of detection aims at making an optimal decision, chosen from a finite number of possible alternatives, with the purpose of obtaining a random sample. In particular, when detecting a disturbed noise signal, the decisions taken into consideration envisage the validation of one of the following hypotheses: “observation x(t) consists of noise” or “signal s(t) is present in observation x(t)”. The criterion that sets the objectives of the detection structure in optimal terms may then, for example, be to minimize the average cost of a decision or even increase the probability of detecting signal s(t) under the condition of an imposed false alarm probability. For these two criteria, one has to demonstrate that the optimal value is achieved by choosing the plausibility ratio for the Λ detection statistics. Comparing it to a λ0 threshold then generates the decision: signal s(t) is assumed to be present in observation x(t) if Λ is superior to λ0 and absent otherwise. One refers to it in this case as frame detection, especially since no constraint acts upon the structure of detection statistics, the latter relying only on the choice of a rule and the knowledge/implementation of plausibility laws [51]. At the same time, in a large number of situations, the information provided by the statistical properties of the sample for each hypothesis is not available, which makes the previous study impossible. Under these circumstances, a possible but suboptimal case consists of defining in advance the nature of the detection statistics, then optimizing the characteristic parameters according to a given criterion. For this reason, this study is known as imposed structure detection [52].
Under these circumstances, the problems encountered during the development of a detection rule are generally related to the difficulty of answering the following questions:
  • How to choose the pattern/model of detection?
  • What are the optimal criteria for determining the characteristic parameters of this model? What optimization procedure is being adopted?
In this article, we propose a new approach to the analysis, recognition, and classification of K-complexes based on the application of recursion properties to Cohen class distributions modified by the reallocation method, developing a recursive strategy to capitalize the reallocation algorithm for time–frequency representations. We also demonstrate the advantage of the size of the time support of the parameterization function for Cohen class distributions, but especially for the Wigner–Ville distribution, proposing an algorithm for using recursion in the reallocation process of the Wigner–Ville distribution spectrogram. Next, we established the necessary and sufficient conditions allowed by the analysis windows and the parameterization function, so that the Cohen class distributions could be used after the introduction of recursion and we extended these conditions for the distributions modified by the reallocation method. We used the results obtained from the Wigner–Ville distribution to train a convolutional neural network (deep learning), which we trained with the Calibrated Stochastic Gradient Descend (CSGD) optimization algorithm because it updates the parameters for each training stage. SGD also eliminates the redundant calculations that occur for large datasets and EEG records, by performing frequent updates. Another advantage of SGD is the speed of calculations. The test results showed a detection percentage of 98.30%.
The rest of the paper is structured as follows: it presents the model of analysis and detection of K-complexes in Section 2.1, the EEG microstructure of sleep in Section 2.2, a comparative analysis of Cohen class energy distributions in Section 2.3, the new algorithm for recursiveness and reallocation in the Section 2.4 and Section 2.5, and the results obtained by the new method proposed together with the deep neural network used in Section 3.

2. Materials and Methods

2.1. Proposed Methods

Once the signals have been acquired, according to the traditional method of acquiring EEG signals during sleep, they are processed according to the sequence of steps illustrated in Figure 1.
First, the EEG signals must be filtered. This step is imperative because the acquired data overlap a parasitic signal of 50 Hz coming from the power supply network that can affect the useful signal, making the classification almost impossible. To eliminate the area of network noise, an Infinite Impulse Response (IIR) notch filter, implemented in MATLAB, was used. The alternative to the IIR filter is the Finite Impulse Response (FIR) filter. The differences include:
  • IIR has a faster response time than FIR [53];
  • IIR phases out different components frequently; it is different from FIR, which generates a constant phase shift (IIR has a nonlinear phase and FIR has a linear phase) [53];
  • FIR is always stable, while IIR is sometimes unstable [53].
Since the information of interest (through which the class will be made) is not in the phase component but in the frequency component, we chose for this first stage of processing the IIR filter. After segmentation, the Wigner–Ville algorithm from the Cohen class is applied, for which we introduce a new method of recursiveness and reallocation of the spectrum so as to extract the fundamental parameters that characterize the K-complexes. After extracting the characteristics, we use a deep neural network for the automatic detection and classification of K-complexes.

2.2. Microstructure of Sleep EEG

According to this classification, the four stages of sleep, represented in Figure 2, are as follows [49]:
  • Stage 1: In this stage, we have smooth sleep: we easily pass from the waking state to the sleep state and can be awakened by even the slightest noise. Our eyes move very slowly and muscle activity slows down. We also experience sudden, involuntary muscle contractions called myoclonus, often preceded by a feeling of falling into emptiness. People awakened from this stage often remember fragmented visual images.
  • Stage 2: When we enter the second stage of sleep, eye movements stop, the heart begins to beat slower, the muscles relax, and the body temperature drops. Also, brain waves (fluctuations in electrical activity, which can be measured by electrodes) become slower and a series of occasional fast waves called sleep spindles appear. Basically, the body prepares for deep sleep.
  • Stage 3: Stage 3 of sleep is characterized by extremely slow brain waves and the lack of any eye movement or muscle spasms, which promotes deep sleep. Once we reach this stage, it is very difficult to wake up. People awakened during deep sleep do not immediately adapt to reality and, for a few minutes, are dizzy and disoriented.
  • Stage 4: In this stage, the heart rate, breathing, and eye movements become faster and faster. The brain becomes more active, processing the things we learned during the day to help us form memories. Usually, during REM (Rapid Eye Movement) sleep, people dream; that is why those who are awakened at this stage often tell bizarre and illogical stories of what they experienced.
Sleep is of two types, each with its own characteristics:
(1)
REM (rapid eye movement);
(2)
non-REM, with four depth stages, 1, 2, 3, and 4.
When we fall asleep, we enter a state of non-REM sleep, in different stages; then we go into a state of REM sleep (rapid eye movement) characterized by rapid and simultaneous movements of both eyes, hence the name. Research studies show that the two types of sleep alternate with each other and take place in the form of cycles. The duration of these cycles is about 70–100 min (the first non-REM–REM cycle), the next cycles being longer, 90–120 min. Also, the passage through all stages of non-REM sleep is not observed. In the last period of sleep, morning sleep, the duration of REM sleep increases. It is not known exactly why these phases of sleep alternate. What is known is that there is a higher proportion of non-REM sleep (about 80%) compared to REM sleep (about 20–25%) [54].
The K-complex, together with the sleep cycles, is one of the main “markers” of sleep onset as it appears in stage 2.
The K-complex can have different morphologies, which can appear every fourth second, as follows: isolated K-complex, without spindle; K-complex and spindle insulated; K-complex pair; and K-complex with negatively flattened component. A K-complex is an EEG waveform that occurs in stage 2 of REM sleep. It is “the biggest event in a healthy person,” being more common in the first sleep cycles. The K-complexes have two proposed functions: first, to suppress cortical arousal in response to stimuli that the sleeping brain does not evaluate to signal a danger, and second, to support sleep-based memory consolidation. The K-complex consists of a short peak of high negative voltage, usually greater than 100 μV, followed by a slower positive complex around 350–550 ms and 900 ms at the final negative peak. The K-complexes occur approximately every 1.0 to 1.7 min and are often followed by bursts of sleep spindles. These occur spontaneously, but also in response to external stimuli, such as sounds, touching the skin, and internal stimuli, such as inspiratory interruptions. They are generated in diffuse cortical locations, although they tend to predominate over the frontal parts of the brain. Both the K-complex and the activity of delta waves in stage 2 sleep create a slow wave of 0.8 Hz and delta oscillations (1.6–4.0 Hz). However, their topographic distribution is different and the delta power of the K-complexes is higher. They are created by the diffuse appearance in the cortical areas of the external dendritic currents in the middle (III) and (I) starts of the cerebral cortex. This is accompanied by a decrease in the broadband power of the EEG, including the activity of gamma waves. This produces “down-states” of neural silence in which the activity of the neural network is reduced.
The K-complex, sleep spindle and delta wave are shown in Figure 3 and Figure 4.

2.3. Comparative Analysis of Cohen Class Energy Distributions

As this is a comparative study that resorts to time–frequency representations, a linear frequency modulated pulse was used, of unit amplitude, and the frequency, rated/normed at sampling frequency, was increased linearly from 0.1 to 0.3. The signal length corresponds to 512 samples [55]. Table 1 presents the Cohen-class frequency–time representations and associated core functions.
Figure 5 (left) shows the smoothed-out Wigner–Ville pseudodistribution applied for the test signal. The dimensions of the smoothing windows for the simultaneous analysis in the time and frequency axes are 64 points and the Fourier transform is calculated at 512 points. It is noted that this “knife blade” representation preserves the temporal and frequency support and provides a precise picture of how the signal energy is distributed along the time–frequency plane. Figure 5 (right) shows the time–frequency plane viewed from above at an angle of 90°. It may be seen how thin the “energy blade” of the test signal is and how sharply it comes into prominence against the remaining background. It has been found that this has reduced the representation’s resolution in both time and frequency. Another effect of dimming the windows and the FFT dimension has been an increase in the amplitude of the secondary lobes. The advantages of using small-scale analytical windows include better preservation of the time and frequency supports, as well as a reduction in the calculation time. Even under these conditions, one may determine on whose signal type basis such representation was made in case the latter is not known a priori. Good enough estimates may also be made of the parameters of the signal that originated the transformation. The analysis window was in both cases of the Hamming type.
The test signal will be further analyzed using other Cohen class time–frequency representations, such as the Born–Jordan distribution represented in Figure 6, for which Hamming-type windows were used. Compared to the Wigner–Ville distribution, it is similar in that it looks like a somewhat thicker “knife blade,” with a lower resolution, but the secondary lobes are almost unnoticed when the window size H is equal to 64; when the window size decreases to 32, the resolution decreases and the secondary lobes become visible.
Figure 7 contains the Choi–Williams distributions of the linear variable frequency signal for a Hamming window sized 64. A first observation is related to the positive sign of the distribution and its large amplitude. The resolution of the representation is almost as good as the smooth Wigner–Ville pseudo distribution. Interestingly, the secondary lobes have not become visible for any of the dimensions of the analysis window. Instead, some disturbances perpendicular to the time axis and wide frequency support appeared at the signal edge, which are more pronounced when the size of the smoothing window is small.
The Zao–Mark–Atlas distribution of the test signal is shown in Figure 8. As a first conclusion, the amplitude of the representation is lower that that of the other representations studied so far.
The resolution is very good, compared to the other distributions.
The decrease in the size of the analytical window (to 32 vs. 64) led to the emergence of the secondary lobes, as well as to disturbances in the signal edges. We left to last the signal analysis made with the help of the spectrogram (Figure 9), which is also a Cohen class distribution; it had the poorest performance but was the simplest of all. The window used was rectangular, size 64.
From the analysis of a test signal based on time–frequency representations in the Cohen class, one may draw some conclusions:
  • of all the time–frequency distributions investigated, the spectrogram provides the temporal resolution at the lowest frequency and the Zao–Mark–Atlas distribution features the smallest amplitude;
  • the Choi–Williams distribution is positive and has the largest amplitude;
  • the Born–Jordan distribution features almost no secondary lobes;
  • the choice of analytical windows (Blackman, Hanning, Hamming, Bartlett, or rectangular) influences the temporal and frequency resolutions as well as secondary lobe levels.
In the wake of our investigations, some important aspects may be highlighted with respect to the use of signal analysis based on time–frequency distributions in the Cohen class, namely:
  • the energy structure of the analyzed signals may be fairly accurately identified and located in the time–frequency plane;
  • when the type, duration, frequency, and timing of signals are not known in advance, they may be estimated by using time–frequency distributions;
  • one can thus foresee the possibility of implementing these analytical algorithms in the systems for identifying the EEG transient signals;
  • even in the case of signals covering the same spectral range, generated by different sources, the time–frequency distributions allow each of them to be highlighted;
  • one may set up databases that are useful for identifying EEG transient signals as their “signature” can be individualized by using time–frequency representations.
For the analysis of EEG signals in order to detect and classify K-complexes, we will use pseudo-Wigner–Ville reallocated distributions.

2.4. Recursive Implementation of Cohen Representations

Section 2.3 deals with practical problems that depend on the calculus of representations, a lesser subject if we consider that the development of rapid and elegant economic algorithms has widened and continues to broaden the applicability domains of the time–frequency analysis.
Two methods may be identified in this area: the first method envisages the limitation of the large volume of calculations. For this purpose, some authors look for an optimal ordering of operators [56,57,58]. Other authors, on the other hand, choose the symmetry properties of the instant autocorrelation function and some parameterization functions [59,60]. Finally, the distribution of the Cohen class may become the object of a division into a weighted sum of spectrograms, from which only the most significant terms have been retained.
Except for this latter situation, which, at most, can only provide an approximation of the representation, the above methods lead to a substantial reduction of the calculation time. Because of this, some of them are systematically associated with recursive implementation algorithms that are covered by the second method.
Recursiveness and reallocation method
The improvement of the contrast function for time–frequency representations is an important step for the analysis of nonstationary signals. That is why numerous solutions have been proposed, including the reallocation method. In this section, we propose a recursive strategy that aims at reducing the computation time required. Thus, we are interested in introducing recursiveness into the spectrogram reallocation process, then to the pseudo-Wigner–Wille distributions.
(a)
Pseudo-Wigner–Ville reallocated distributions (Recursiveness and reallocation operators)
In [55], the authors formulated the relationships of reallocation operators associated with Cx Cohen class distributions as follows:
n ^ [ n , v ; x ] = n C X [ n , v ; T φ ) C X [ n , v ; φ ) ,
v ^ [ n , v ; x ) = v + j C X [ n , v ; D ϕ ) C X [ n , v ; ϕ ) ,
where T ϕ T R [ m , l ] = m ϕ T R [ m , l ] and D ϕ T R [ m , l ] are the partial derivative of ϕ T R ( t , τ ) / τ evaluated in (m, l). By adopting the same method as for the reallocated spectrogram, it may be shown that Cx[.;) and Cx[.;) may be evaluated recursively when Cx[.;Φ) verifies the direct recursiveness relation (Equation (3)):
C x [ n , v ; ϕ ] = α c x [ n l , v ; ϕ ] + c α M C x [ n + m , v ; φ ] c α M + 1 C x [ n M l , v ; φ ]
To reduce the complexity of the various recursive relationships presented, we have confined the present study only to the cases of pseudo-Wille distribution.
Theorem 1.
Let Φ belong to the class of parameterization functions ξ ( W ˜ X ) with separable variables that allow a direct recursive implementation of the pseudo-Wille–Ville distributions. The following relationships determine the direct and indirect recursiveness of Wx[.;DΦ] and W ˜ x [.;TΦ]:
W ˜ X [ n , v ; D ϕ ) = α W ˜ X [ n 1 , v ; D ϕ ) + l = 1 L l 1 n [ n , l ) e j 4 π v l with η [ n , l ] = D ϕ T R [ M , l ] R X [ n + M , l ] α D ϕ T R [ M , l ] R X [ n M 1 , l ] .
W ˜ X [ n , v ; T ϕ ) = α ( W ˜ X [ n 1 , v ; T ϕ ) W ˜ X [ n 1 , v ; ϕ ) + l = 1 L l 1 ψ [ m , l ) e j 4 π v l with ψ [ n , l ] = M ϕ T R [ M , l ] R X [ n + M , l ] α ( M + 1 ) ϕ T R [ M , l ] R X [ n M 1 , l ]
Equation (4) results ifΦbelongs toξ( W ˜ x ); then DΦ also belongs to ξ( W ˜ x ). Consequently, the direct recursive relation
PWL X [ n , v ; ϕ ] = α PWL X [ N 1 , V ; ϕ ] + L = + k [ n , l ] e j 4 π v l
remains valid for the distribution W ˜ x [.;).
A simple calculation allows us to verify Equation (5). Note that the association of several recursive processes of the same type as Equations (4) and (5) also allows for the extension of the recursive strategy that we have previously presented to the class of parameterization functions ξext( W ˜ x ). In this case, use is made of the linearity property of the Fourier transform.
(b)
Performances
The kernels of the separable variables g[l] h [m], written as gh, which we use to demonstrate interest in the recursive strategy, verify the Hermitian symmetry property:
g [ l ] h [ m ] = g [ l ] h [ m ]
This results in:
( D g h [ l , m ) ) * = D g h [ l , m ]   si   ( T g h [ l , m ] ) * = T g h [ l , m ] ,
which allows us to rewrite the pseudo-Wigner–Ville distributions associated with nuclei Tgh şi Dgh as follows:
W ˜ X [ n , v ; T g h ) = 2 Re { Ξ x [ n , v ; T g h ) } m = M M T g h ( 0 , m ] R x [ n + m , 0 ]
W ˜ X [ n , v ; D g h ) = 2 Im { Ξ x [ n , v ; D g h ) } m = M M D g h [ 0 , m ] R x ( n + m , 0 ] ,  
where
Ξ x [ n , v ; ϕ ) = l = 0 l 1 m = M M ϕ [ l , m ] R x [ n + m , l ] e j 4 π v l .
The relationships retain the general characteristics found in Equations (4) and (5). If the adopted strategy is classical or recursive, the use of Equations (9)–(11) allows for a reduction in the time needed to calculate distributions [., Tgh] and [., Dgh] and thus the reallocation operators. Finally, in order to compare the two methods under the most favorable conditions, the calculations of algorithmic complexity mentioned were made on the basis of these relationships. These results emphasize the fact that, for the recursive algorithm, complexity is independent of the temporal support [−M, M] of the temporal window h. This represents an advantage of the proposed method, as shown in Figure 10a. In contrast, Figure 10b shows that the calculation time gain decreases when the window width g increases for a given time window h.
Finally, it is noted that the calculation time gain is not affected by the number of samples W ˜ x [n,v;gh) reallocated at each moment n, and the calculation of the reallocation algorithm function shown in Figure 11 requires a maximum of 3N additions and 2N multiplications. This specification is valid for the reallocated spectrogram.

2.5. Deep Neuronal Network

To see the separability of the classes, various architectures of deep neural networks were used. As previously mentioned, the features provided as input data consist of the correlations extracted with the Cohen class algorithm. Thus, the networks were tested with a number of hidden layers ranging from 1 to 7. Also, the number of neurons on each layer varied between 8 and 256 (more precisely, 8, 16, 32, 128, 256 neurons per layer were used). The activations tested were sigmoid, tanh, and arctan. They have been used due to the fact that they are known to have good results, especially for class problems [61,62], but also estimation problems [63].
Results with activation functions that deny components’ negatives (e.g., ReLU, Leaky ReLU) were not used in this paper because they provided random responses (an accuracy of around 50%). One possible reason for this is the phenomenon called dying ReLU. It refers to the fact that if too many weights take values below 0 the activation result will be 0, so data discrimination becomes impossible.
The optimization technique used was Stochastic Gradient Descent (SGD). This algorithm is used in the field of deep neural networks and generally provides satisfactory results [64,65,66,67]. The proposed method is a simple one from GD (Gradient Descent) in the sense that, instead of calculating the gradient for the whole lot at each iteration, the gradient is calculated only for a randomly chosen value from the lot [68]. Therefore, the learning process is stochastic and depends on these chosen values. In this way, it is desired that Equation (12) (3.8) for GD behave relatively similarly to Equation (13) (3.9) for SGD, having the advantage of a much smaller calculation volume:
w t + 1 = w t η 1 n i = 1 n w ( z i , w t )
w t + 1 = w t η w ( z t , w t ) .  
In these formulas, z refers to the pair (x; y), meaning the input value (associated output value), and z t is the pair chosen at time t.

3. Results

Due to the relatively small size of the database, we used a batch size of value 1 (only one example was evaluated for updating the parameters). In addition, to ensure that all architectures were tested under the same conditions, the number of epochs was always 100. The tests were performed in two stages.
In the first stage, the database was divided into two: the training part with a size of 350 and the test part with a size of 39, equivalent to a division of 90% and 10%. During training, the purpose of the network is to determine the weights that allocate the ideal labels to the input values. The data that the network learns are called the training batch. In order to test the possibility of generalizing the found function, some values are kept separately (called the test batch) and passed through the network. If the results are good, we can say that the network has learned and is able to classify the new data.
At this stage, different deep neural network architectures were tested. During the tests, the hyperparameters of the system were modified. More precisely, we experimented with various values for the number of neurons, the layer, and various activation functions—tanh, arctan, and sigmoid. The three activation functions were chosen because they have an S shape, which increases the separability of the input data.
The neural architecture tested for the tanh function had five layers and 16 neurons per layer, and the best cost recorded on the test was 0.32 with an accuracy of 90.18%.
The results obtained for the tanh activation function are presented in Table 2.
The next activation function tested was sigmoid, and the best cost on the test batch was 0.35 at an accuracy of 94.65%. It has also been observed that, with the increase in the number of hidden layers and even the number of neurons, the network begins to stop learning and gives random results for training data. The results obtained for the sigmoid activation function are presented in Table 3.
Next, we tested the activation function arctan. This time the number of hidden layers was 1–7, and the number of neurons was varied according to Table 4 In this case, the minimum cost found on the test batch was 0.31 at an accuracy of 95.67% with seven hidden layers and 16 neurons on each layer. The results obtained for the arctan activation function are presented in Table 4.
The best result was obtained using an architecture with six hidden states, eight neurons on each hidden layer, and each layer (excluding the output layer, which had softmax) having the arctan activation function. The accuracy of classes (in this case, on the test group) was 98.30%.
In the second stage, optimization, validation, and increase of generalization capacity, the database used in stage one was divided into 80% for training and 20% for testing (another test was done for ordering the datasets into 80% for training, 10% for validation, and 10% for testing) using the SGD optimization algorithm.
The algorithms presented in this article were developed in the LabVIEW 2020 SP1 programming environment with a 17-inch high-performance portable computing system with an Intel processor, four cores @ 3.2 GHz, 16 GB DDR3 RAM @ 1600 MHz, 256 GB SSD, 1 TB HDD, and high-performance dedicated video card, optimized for computer graphics, with K-complex detection being offline for optimized for computer graphics to train and test the MLP-CNN neural network with Wigner–Ville. In the future, we hope to develop a portable system for online analysis based on an FPGA architecture with an Intel Core i7 processor (3.1 GHz) or Coral Dev Board that can be used for a machine learning embedded system; again, the software will be developed in the Pytorch framework.
The final part of a deep network is represented by one or more fully connected multilayer perceptron (MLP) layers that perform the classification part, using as inputs the outputs of the initial convolutional part (characteristics automatically extracted by it) obtained for each applied signal network input.
The optimizer is the algorithm that decides in which direction and how strongly to change the weights in the network. In order to adjust the parameters of a network (weights and displacements), an optimizer must be used that decides the modification strategy according to the gradients obtained with the help of the loss function. The algorithm used is Calibrated Stochastic Gradient Descent (CSGD). The optimizer requires specifying the size of the optimization step, also called the learning rate, which in our tests will be kept constant at 10−5.
It is, practically speaking, an iterative optimization process that aims to reduce the error (respectively the difference between the expected output and the one obtained in that iteration), estimated using a cost (or loss) function. In this way, those inputs and characteristics that are the most relevant for obtaining the desired output are enhanced.
This experiment resulted in a classification accuracy of 98.3%, recall of 0.96, and microaverage F1 score of 0.97. The confusion matrix is shown in Table 5 and the classification report is given in Table 6.
This was to be expected as the cost does not provide information on accuracy. The latter is calculated according to the formula:
a c c u r a c y = n u m b e r   o f   c a s e s   c l a s s i f i e d   c o r r e c t l y n u m b e r   o f   t o t a l   c a s e s
and the cost function used is binary cross entropy and behaves according to the following law:
= 1 n i = 1 N y i l o g ( y ^ ) + ( 1 y i ) l o g ( 1 y ^ ) .
In the field of neural networks, there are two phenomena that can occur during training: underfitting and overfitting. The first refers to when learning is stopped too quickly compared to the number of epochs. As can be seen from Figure 12, underfitting occurs when the cost is still decreasing, both for the training group and for the test group, but a minimum cost has not been reached. Therefore, the subtraining area is characterized by low accuracy in the learning and testing groups. Overfitting is the extreme opposite of underfitting. This time the network is left to learn for a large number of eras. In this case, parameters were found that fit the drive data almost perfectly. The problem is that the network did not learn a real connection between the signals, but only managed to map the input data to the corresponding output values. In this case, the cost per test batch increases because the network is not robust and does not manage new data well.
Analyzing the results presented in this article, we can say that in no variant of the tested architecture was there overfitting. However, the rule of saving the model was: keep the one that has the lowest cost on the test lot. Thus, all results presented are not affected by either underfitting or overfitting.
The computational time required to extract the characteristics of spectrograms using MLP-CNN was 1.26 s. The total training time required for the EEG spectrograms was 13 min. The trained model classified K-complexes’ shape in 5 s for offline analysis and classifications.
To compare the results obtained with the proposed method, we referred to the following articles [69,70,71,72,73], which have as their theme the automatic detection of the K-complex. The comparative results are presented in Table 7.

4. Discussion and Conclusions

The present study was spurred by the need to develop a definition of a methodology for the synthesis of imposed structure detectors with applications in the detection of K-complexes using time–frequency representation. Furthermore, before drawing any conclusions, we must recall the main outcomes, from which we can highlight some future research directions. Initially, we presented the notion of time–frequency representation, insisting on the distribution of energy in the Cohen class. We have presented some practical problems related to their use, such as the presence of interference terms that are harmful to the interpretation of results.
Among other difficulties that may arise, we have been particularly interested in the large amount of computing that occurs in the analysis of long signals. Thus, after emphasizing the potential of recursive algorithms for the rapid calculation of time–frequency representations, we presented in a homogeneous and unified manner the direct and indirect recursive properties of the Cohen class.
Furthermore, we expanded their distribution by the reallocation method. Then, we made the first strategic choice in defining a methodology for the design of imposed structure detectors using a deep neural network. The aim of the present study was to select the best detection test for a given pattern, as far as the definition of an optimal criterion is concerned. To determine the decision rule, our choice was a deep neural network with arctan-type activation function, due to its consistency, which is a guarantee of performance. The method allowed us to present an answer to this problem, leading to the analytical determination of the best criterion, for which the detector for the obtained K-complexes has a minimum probability of error. Finally, the time–frequency plan, which is an optimal space for analyzing the configuration of some detectors, made it possible to verify the method. An original presentation of the potential of the Cohen class in solving certain detection problems with an imposed structure is provided, which gives the representation a decisive role in the decision-making process. In the final stage, different deep neural network architectures were tested. During the tests, the hyperparameters of the system were modified. More precisely, we experimented with various values for the number of neurons and layers, and various activation functions—tanh, arctan, and sigmoid. These three activation functions were chosen because they have an S shape, which increases the separability of the input data. The best result was obtained using an architecture with six hidden states, eight neurons on each hidden layer, and with each layer (excluding the output layer, which had softmax) having the arctan activation function. The accuracy of classes (in this case, in the test group) was 98.30%.
The limitations of this study were: (a limitation of all sleep EEG studies) recorded EEG signals represent the sum of the electrical activity in large areas of the brain, which can lead to an inaccurate location; to improve the performance of the proposed method it is useful to identify K-complexes from multi-electrode data; EEG recordings for somnology are often not of a quality requiring the use of filters, which can introduce the smoothing of the EEG signal, which is interpreted as uniformity of brain activity during sleep; in most cases, public databases for brain activity during sleep are not annotated, which makes it difficult to train the neural network; future studies with larger annotated sleep EEG databases will be needed to assess and compare the robustness of our method at each stage of sleep, as access to specialized somnology clinic databases is currently very limited due to GDPR legislation; one direction of increasing the performance of our method is the fine adjustment of the analysis and recognition parameters from one subject to another, taking into account the individual differences of the subjects and the properties of the K-complex.
Given both the theoretical and practical knowledge gained from the research presented in this article, and the software developments made to validate the theoretical results, we believe that they can be successfully addressed in the future: analysis and implementation of theoretical research that can lead to a link between time–frequency analysis and the SRM (Structural Risk Minimization) principle; optimizing the reconstruction property of Wigner–Ville representations, starting from a partial, but not random, knowledge of it, to remove the signals from the noise and the exact representation of the K-complex and the sleep spindle, synthesizing a linear detector that operates only on independent linear components and can fully process Wigner–Ville representations and the realization of a portable hardware (e.g., a Coral Dev Board can be used for a machine learning embedded system) and software system for the analysis, recognition, and classification of a K-complex and sleep spindle in real time so that this system can be used in patients’ homes (home care).
Finally, the methodology presented in this paper has been validated for an application with a certain degree of difficulty, namely, the detection of the K-complex in the sleep electroencephalogram. The proposed solution is among the performance solutions described in the literature. In addition, these results may be improved by correlating information from multiple electroencephalogram records.

Author Contributions

Conceptualization, C.D.; Methodology, I.-M.C.; Laboratory testing, C.D. and A.-C.C.; Analysis, C.D.; Software, C.D.; Validation, C.D., I.-M.C. and A.-C.C.; Formal analysis, C.D. and A.S.; Investigation, C.D., I.-M.C. and A.-C.C.; Resources, C.D., I.-M.C. and A.S.; Data curation, C.D.; Writing—original draft preparation, C.D. and I.-M.C.; Writing—review and editing, C.D.; Visualization, I.-M.C., A.-C.C. and A.S.; Supervision, C.D., I.-M.C., A.S. and A.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Anonymized records obtained from a somnology clinic were used, and the resulting software application is to be tested in the same clinic.

Acknowledgments

We thank the management of the Faculty of Transportation, the Department of Electronics and Telecommunications, University Politehnica of Bucharest, Romania, for the use of research from the Artificial Intelligence Laboratory.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alger, S.E.; Chambers, A.M.; Cunningham, T.; Payne, J.D. The role of sleep in human declarative memory consolidation. Sleep Neuronal Plast. Brain Funct. 2014, 25, 269–306. [Google Scholar] [CrossRef]
  2. What Is Sleep? American Sleep Association. 2016. Available online: https://www.sleepassociation.org/patients-general-public/what-is-sleep/ (accessed on 23 September 2021).
  3. Rechtschaffen, A. A manual of standardized terminology, techniques and scoring system for sleep stages of human subjects. Public Health Serv. 1969, 20, 246–247. [Google Scholar]
  4. Moser, D.; Anderer, P.; Gruber, G.; Parapatics, S.; Loretz, E.; Boeck, M.; Kloesch, G.; Heller, E.; Schmidt, A.; Danker-Hopfe, H.; et al. Sleep Classification According to AASM and Rechtschaffen & Kales: Effects on Sleep Scoring Parameters. Sleep 2009, 32, 139–149. [Google Scholar] [PubMed]
  5. Fletcher, J. Why Sleep is Essential for Health? Available online: https://www.medicalnewstoday.com/articles/325353/ (accessed on 25 August 2021).
  6. American Academy of Sleep Medicine. International Classification of Sleep Disorders. 2005, p. 51. Available online: https://ci.nii.ac.jp/naid/20001061569/ (accessed on 13 August 2021).
  7. Thorpy, M.J. Classification of sleep disorders. Neurotherapeutics 2012, 9, 687–701. [Google Scholar] [CrossRef]
  8. Sharma, M.; Dhiman, H.S.; Acharya, U.R. Automatic identification of insomnia using optimal antisymmetric biorthogonal wavelet filter bank with ECG signals. Comput. Biol. Med. 2021, 131, 104246. [Google Scholar]
  9. Bhattacharya, D.; Sen, M.; Suri, J. Epidemiology of insomnia: A review of the global and Indian scenario. Ind. J. Sleep Med. 2013, 8, 100–110. [Google Scholar] [CrossRef]
  10. Murali, R.; Rangarajan, P.; Mounissamy, A. Bruxism: Conceptual discussion and review. J. Pharm. Bioallied Sci. 2015, 7, S265. [Google Scholar] [CrossRef]
  11. Anderson, K.N. An update in sleep neurology: The latest bed time stories. J. Neurol. 2015, 262, 487–491. [Google Scholar] [CrossRef] [PubMed]
  12. Blagrove, M.; Ruby, P.; Eichenlaub, J.-B. Dreams are made of memories, but may be not formemory. Behav. Brain Sci. 2013, 36, 609–610. [Google Scholar] [CrossRef]
  13. Al-Salman, W.; Li, Y.; Wen, P.; Diykh, M. An efficient approach for EEG sleep spindles detection based on fractal dimension coupled with time frequency image. Biomed. Signal Process. Control 2018, 41, 210–221. [Google Scholar] [CrossRef]
  14. Nobili, L.; Proserpio, P.; Combi, R.; Provini, F.; Plazzi, G.; Bisulli, F.; Tassi, L.; Tinuper, P. Nocturnal frontal lobe epilepsy. Curr. Neurol. Neurosci. Rep. 2014, 14, 424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Bernhardt, B.C.; Bonilha, L.; Gross, D.W. Network analysis for a network disorder: The emerging role of graph theory in the study of epilepsy. Epilepsy Behav. 2015, 50, 162–170. [Google Scholar] [CrossRef] [PubMed]
  16. Natarajan, R. Review of periodic limb movement and restless leg syndrome. J. Postgrad. Med. 2010, 56, 157. [Google Scholar] [CrossRef] [PubMed]
  17. Carskadon, M.A.; Dement, W.C. Monitoring and Staging Human Sleep. In Principles and Practice of Sleep Medicine, 5th ed.; Kryger, M.H., Roth, T., Dement, W.C., Eds.; Elsevier Saunders: St. Louis, MO, USA, 2011; pp. 16–26. [Google Scholar] [CrossRef]
  18. Bhat, S.; Acharya, U.R.; Hagiwara, Y.; Dadmehr, N.; Adeli, H. Parkinson’s disease: Cause factors, measurable indicators, and early diagnosis. Comput. Biol. Med. 2018, 102, 234–241. [Google Scholar] [CrossRef]
  19. Politis, M.; Wu, K.; Molloy, S.; Bain, P.G.; Chaudhuri, K.R.; Piccini, P. Parkinson’s disease symptoms: The patient’s perspective. Mov. Disord. 2010, 25, 1646–1651. [Google Scholar] [CrossRef] [PubMed]
  20. Balestrino, R.; Schapira, A.H.V. Parkinson disease. Eur. J. Neurol. 2020, 27, 27–42. [Google Scholar] [CrossRef]
  21. Báez-Mendoza, R.; Schultz, W. The role of the striatum in social behavior. Front. Neurosci. 2013, 7, 233. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Savitt, J.M.; Dawson, V.L.; Dawson, T.M. Diagnosis and treatment of Parkinson disease: Molecules to medicine. J. Clin. Investig. 2006, 116, 1744–1754. [Google Scholar] [CrossRef] [Green Version]
  23. Yuvaraj, R.; Murugappan, M.; Acharya, U.R.; Adeli, H.; Ibrahim, N.M.; Mesquita, E. Brain functional connectivity patterns for emotional state classification in Parkinson’s disease patients without dementia. Behav. Brain Res. 2016, 298, 248–260. [Google Scholar] [CrossRef]
  24. Imbriani, P.; Schirinzi, T.; Meringolo, M.; Mercuri, N.B.; Pisani, A. Centrality of Early Synaptopathy in Parkinson’s Disease. Front. Neurol. 2018, 9, 103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Pagano, G.; Niccolini, F.; Politis, M. Imaging in Parkinson’s disease. Clin. Med. 2016, 16, 371–375. [Google Scholar] [CrossRef]
  26. Orozco, J.L.; Valderrama-Chaparro, J.A.; Pinilla-Monsalve, G.D.; Molina-Echeverry, M.I.; Pérez Castaño, A.M.; Ariza-Araújo, Y.; Prada, S.I.; Takeuchi, Y. Parkinson’s disease prevalence, age distribution and staging in Colombia. Neurol. Int. 2020, 12, 8401. [Google Scholar] [CrossRef]
  27. Miller, I.N.; Cronin-Golomb, A. Gender differences in Parkinson’s disease: Clinical characteristics and cognition. Mov. Disord. 2010, 25, 2695–2703. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Silva, M.V.F.; Loures, C.D.M.G.; Alves, L.C.V.; De Souza, L.C.; Borges, K.B.G.; Carvalho, M.D.G. Alzheimer’s disease: Risk factors and potentially protective measures. J. Biomed. Sci. 2019, 26, 33. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Kang, D.W.; Lee, C.U.; Lim, H.K. Role of Sleep Disturbance in the Trajectory of Alzheimer’s Disease. Clin. Psychopharmacol. Neurosci. 2017, 15, 89–99. [Google Scholar] [CrossRef] [Green Version]
  30. Cooke, J.R.; Ancoli-Israel, S. Normal and abnormal sleep in the elderly. Handb. Clin. Neurol. 2011, 98, 653–665. [Google Scholar]
  31. Tatineny, P.; Shafi, F.; Gohar, A.; Bhat, A. Sleep in the elderly. Sci. Med. 2020, 17, 490–495. [Google Scholar]
  32. Peter-Derex, L.; Yammine, P.; Bastuji, H.; Croisile, B. Sleep and Alzheimer’s disease. Sleep Med. Rev. 2015, 19, 29–38. [Google Scholar] [CrossRef] [PubMed]
  33. Chwiszczuk, L.; Breitve, M.; Hynninen, M.; Gjerstad, M.D.; Aarsland, D.; Rongve, A. Higher Frequency and Complexity of Sleep Disturbances in Dementia with Lewy Bodies as Compared to Alzheimer’s Disease. Neurodegener. Dis. 2015, 16, 152–160. [Google Scholar] [CrossRef] [PubMed]
  34. Bagyinszky, E.; Youn, Y.C.; An, S.S.; Kim, S. The genetics of Alzheimer’s disease. Clin. Interv. 2014, 9, 535–551. [Google Scholar] [CrossRef] [Green Version]
  35. Minakawa, E.N.; Wada, K.; Nagai, Y. Sleep Disturbance as a Potential Modifiable Risk Factor for Alzheimer’s Disease. Int. J. Mol. Sci. 2019, 20, 803. [Google Scholar] [CrossRef] [Green Version]
  36. Busche, M.A.; Kekuš, M.; Adelsberger, H.; Noda, T.; Förstl, H.; Nelken, I.; Konnerth, A. Rescue of long-range circuit dysfunction in Alzheimer’s disease models. Nat. Neurosci. 2015, 18, 1623–1630. [Google Scholar] [CrossRef] [PubMed]
  37. Bedrosian, T.A.; Nelson, R.J. Sundowning syndrome in aging and dementia: Research in mouse models. Exp. Neurol. 2013, 243, 67–73. [Google Scholar] [CrossRef] [PubMed]
  38. Khare, S.K.; Bajaj, V.; Acharya, U.R. Detection of Parkinson’s disease using automated tunable Q wavelet transform technique with EEG signals. Biocybern. Biomed. Eng. 2021, 41, 679–689. [Google Scholar] [CrossRef]
  39. Yuvaraj, R.; Acharya, U.R.; Hagiwara, Y. A novel Parkinson’s Disease Diagnosis Index using higher-order spectra features in EEG signals. Neural Comput. Appl. 2018, 30, 1225–1235. [Google Scholar] [CrossRef]
  40. Sharma, M.; Goyal, D.; Achuth, P.; Acharya, U.R. An accurate sleep stages classification system using a new class of optimally time-frequency localized three-band wavelet filter bank. Comput. Biol. Med. 2018, 98, 58–75. [Google Scholar] [CrossRef]
  41. Dhok, S.; Pimpalkhute, V.; Chandurkar, A.; Bhurane, A.A.; Sharma, M.; Acharya, U.R. Automated phase classification in cyclic alternating patterns in sleep stages using Wigner-Ville Distribution based features. Comput. Biol. Med. 2020, 119, 103691. [Google Scholar] [CrossRef]
  42. Hartmann, S.; Baumert, M. Automatic a-phase detection of cyclic alternating patterns in sleep using dynamic temporal information. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 1695–1703. [Google Scholar] [CrossRef]
  43. Biswal, S.; Sun, H.; Goparaju, B.; Westover, M.B.; Sun, J.; Bianchi, M.T. Expert-level sleep scoring with deep neural networks. J. Am. Med. Inform. Assoc. 2018, 25, 1643–1650. [Google Scholar] [CrossRef] [Green Version]
  44. Lajnef, T.; Chaibi, S.; Ruby, P.; Aguera, P.-E.; Eichenlaub, J.-B.; Samet, M.; Kachouri, A.; Jerbi, K. Learning machines and sleeping brains: Automatic sleep stage classification using decision-tree multi-class support vector machines. J. Neurosci. Methods 2015, 250, 94–105. [Google Scholar] [CrossRef] [PubMed]
  45. Hartmann, S.; Baumert, M. Improved A-phase Detection of Cyclic Alternating Pattern Using Deep Learning. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 1842–1845. [Google Scholar]
  46. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  47. Yildirim, O.; Baloglu, U.B.; Acharya, U.R. A Deep Learning Model for Automated Sleep Stages Classification Using PSG Signals. Int. J. Environ. Res. Public Health 2019, 16, 599. [Google Scholar] [CrossRef] [Green Version]
  48. Hafezi, M.; Montazeri, N.; Saha, S.; Zhu, K.; Gavrilovic, B.; Yadollahi, A.; Taati, B. Sleep Apnea Severity Estimation from Tracheal Movements Using a Deep Learning Model. IEEE Access 2020, 8, 22641–22649. [Google Scholar] [CrossRef]
  49. Dumitrescu, C.; COSTEA, I.M.; Nemtanu, F.C.; Stan, V.A.; Gheorghiu, A.R. Application of the Wavelet Transform in Machine-Learning. Univ. Politeh. Buchar. Sci. Bull. Ser. Appl. Math. Phys. 2014, 76, 167–178. [Google Scholar]
  50. Ozsen, S. Classification of sleep stages using class-dependent sequential feature selection and artificial neural network. Neural Comput. Appl. 2013, 23, 1239–1250. [Google Scholar] [CrossRef]
  51. Dodge, Y. The Concise Encyclopedia of Statistics; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  52. McHugh, M.L. Lessons in biostatistics. Biochem. Med. 2009, 19, 120–126. [Google Scholar] [CrossRef]
  53. Litwin, L. FIR and IIR digital filters. IEEE Potentials 2000, 19, 28–31. [Google Scholar] [CrossRef]
  54. Lee, Y.-H.; Chen, Y.-S.; Chen, L.-F. Automated sleep staging using single EEG channel for REM sleep deprivation. In Proceedings of the Ninth IEEE International Conference on Bioinformatics and Bioengineering, Taichung, Taiwan, 22 June 2009; pp. 439–442. [Google Scholar]
  55. Cohen, L. Time-Frequency Analysis; Prentice Hall: Englewood Cliffs, NJ, USA, 1995; Volume 778. [Google Scholar]
  56. Bhati, D.; Sharma, M.; Pachori, R.B.; Nair, S.S.; Gadre, V.M. Design of Time–Frequency Optimal Three-Band Wavelet Filter Banks with Unit Sobolev Regularity Using Frequency Domain Sampling. Circuits Syst. Signal Process. 2016, 35, 4501–4531. [Google Scholar] [CrossRef]
  57. Lilly, J.M.; Olhede, S.C. Generalized morse wavelets as a superfamily of analytic wavelets. IEEE Trans. Signal. Process. 2012, 60, 6036–6041. [Google Scholar] [CrossRef] [Green Version]
  58. Erdamar, A.; Duman, F.; Yetkin, S. A wavelet and teager energy operator based method for automatic detection of K-Complex in sleep EEG. Expert Syst. Appl. 2012, 39, 1284–1290. [Google Scholar] [CrossRef]
  59. Sharma, M.; Acharya, U.R. Analysis of knee-joint vibroarthographic signals using bandwidth-duration localized three-channel filter bank. Comput. Electr. Eng. 2018, 72, 191–202. [Google Scholar] [CrossRef]
  60. Sharma, M.; Acharya, U.R. A new method to identify coronary artery disease with ECG signals and time-Frequency concentrated antisymmetric biorthogonal wavelet filter bank. Pattern Recognit. Lett. 2019, 125, 235–240. [Google Scholar] [CrossRef]
  61. Han, S.-J.; Song, M.; Kim, J.; Lee, W.-S.; Lee, H.-K. Classification of temporal bone pneumatization based on sigmoid sinus using computed tomography. Clin. Radiol. 2007, 62, 1110–1118. [Google Scholar] [CrossRef]
  62. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 36393655. [Google Scholar] [CrossRef] [Green Version]
  63. Liao, B.; Ma, C.; Xiao, L.; Lu, R.; Ding, L. An arctan-activated wasd neural network approach to the prediction of dow jones industrial average. In International Symposium on Neural Networks; Springer: Cham, Switzerland, 2017. [Google Scholar]
  64. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of COMPSTAT’2010; Physica-Verlag HD: Heidelberg, Germany, 2010. [Google Scholar]
  65. Johnson, R.; Zhang, T. Accelerating stochastic gradient descent using predictive variance reduction. Adv. Neural Inf. Process. Syst. 2013, 26, 315–323. [Google Scholar]
  66. Zinkevich, M.; Weimer, M.; Smola, A.J.; Li, L. Parallelized stochastic gradient descent. Adv. Neural Inf. Process. Syst. 2010, 4, 2595–2603. [Google Scholar]
  67. Zhang, T. Solving large scale linear prediction problems using stochastic gradient descent algorithms. In Proceedings of the Twenty First International Conference on Machine Learning, New York, NY, USA, 4–8 July 2004; p. 116. [Google Scholar]
  68. Bottou, L. Stochastic gradient descent tricks. In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  69. Elajnef, T.; Echaibi, S.; Eeichenlaub, J.-B.; Ruby, P.M.; Eaguera, P.-E.; Esamet, M.; Ekachouri, A.; Ejerbi, K. Sleepspindle and K-complex detection using tunable Q-factor wavelet transform and morphological component analysis. Front. Hum. Neurosci. 2015, 9, 414. [Google Scholar] [CrossRef]
  70. Al-Salman, W.; Li, Y.; Wen, P. Detection of EEG K-Complexes Using Fractal Dimension of Time Frequency Images Technique Coupled with Undirected Graph Features. Front. Hum. Neurosci. 2019, 13, 45. [Google Scholar] [CrossRef] [Green Version]
  71. Al-Salman, W.; Li, Y.; Wen, P. Detection of k-complexes in EEG signals using a multi-domain feature extraction coupled with a least square support vector machine classifier. Neurosci. Res. 2012, 172, 26–40. [Google Scholar] [CrossRef] [PubMed]
  72. Chambon, S.; Thorey, V.; Arnal, P.J.; Mignot, E.; Gramfort, A. A deep learning architecture to detect events in EEG signals during sleep. In Proceedings of the 2018 IEEE 28th International Workshop on Machine Learning for Signal Processing (MLSP), Aalborg, Denmark, 17–20 September 2018. [Google Scholar]
  73. Oliveira, G.H.; Coutinho, L.R.; da Silva, J.C.; Pinto, I.J.; Ferreira, J.M.; Silva, F.J.; Santos, D.V.; Teles, A.S. Multitaper-based method for automatic k-complex detection in human sleep EEG. Expert Syst. Appl. 2020, 151, 113331. [Google Scholar] [CrossRef]
Figure 1. The architecture of the proposed method for the detection of K-complex brain waves.
Figure 1. The architecture of the proposed method for the detection of K-complex brain waves.
Sensors 21 07230 g001
Figure 2. Sleep stage [49].
Figure 2. Sleep stage [49].
Sensors 21 07230 g002
Figure 3. Brain waves: K-complex and sleep spindle [49].
Figure 3. Brain waves: K-complex and sleep spindle [49].
Sensors 21 07230 g003
Figure 4. Waveforms characteristic of the K-complex and delta waves that occur during sleep [49].
Figure 4. Waveforms characteristic of the K-complex and delta waves that occur during sleep [49].
Sensors 21 07230 g004
Figure 5. The pseudo-Wigner–Ville representation of the test signal.
Figure 5. The pseudo-Wigner–Ville representation of the test signal.
Sensors 21 07230 g005
Figure 6. Born–Jordan representation of the test signal.
Figure 6. Born–Jordan representation of the test signal.
Sensors 21 07230 g006
Figure 7. Choi–Williams representation of the test signal.
Figure 7. Choi–Williams representation of the test signal.
Sensors 21 07230 g007
Figure 8. Zao–Mark–Atlas representation of test signal.
Figure 8. Zao–Mark–Atlas representation of test signal.
Sensors 21 07230 g008
Figure 9. STFT spectrogram of the test signal.
Figure 9. STFT spectrogram of the test signal.
Sensors 21 07230 g009
Figure 10. Estimation of time gain by using recursive calculation applied to the classical method for evaluating the re-allocated pseudo-Wigner–Ville distribution. Different time windows h are analyzed: rectangular (●), semi-sinusoidal (■), Hamming, and Hanning (♦). (a) Temporal window h of semilength l that is set at 64. (b) Frequency window g with semilength M set at 32.
Figure 10. Estimation of time gain by using recursive calculation applied to the classical method for evaluating the re-allocated pseudo-Wigner–Ville distribution. Different time windows h are analyzed: rectangular (●), semi-sinusoidal (■), Hamming, and Hanning (♦). (a) Temporal window h of semilength l that is set at 64. (b) Frequency window g with semilength M set at 32.
Sensors 21 07230 g010
Figure 11. Algorithm with recursive strategy for the evaluation of the reallocated pseudo-Wigner–Ville distribution.
Figure 11. Algorithm with recursive strategy for the evaluation of the reallocated pseudo-Wigner–Ville distribution.
Sensors 21 07230 g011
Figure 12. Illustration of underfitting and overfitting.
Figure 12. Illustration of underfitting and overfitting.
Sensors 21 07230 g012
Table 1. Cohen class frequency–time representations and associated core functions.
Table 1. Cohen class frequency–time representations and associated core functions.
Φ ( ξ , τ ) C x ( t , ν , Φ ) Distribution
1 f ( t + τ 2 ) f ( t τ 2 ) e i 2 π ν τ d τ Wigner-Ville
sin ( π ξ τ ) π ξ τ [ 1 | r | t τ 2 t + τ 2 f ( s + τ 2 ) f ( s τ 2 ) d s ] e i 2 π ν τ d τ Born-Jordan
e ( π ξ τ σ ) 2 2 σ | r | e 2 σ 2 ( s t ) 2 τ 2 f ( s + τ 2 ) f ( s τ 2 ) e i 2 π ν τ d s d τ Choi-Williams
A h ( ξ , τ ) | f ( s ) h ( s t ) e i 2 π ν s d s | 2 Spectrograma
Table 2. The results obtained for the tanh activation function.
Table 2. The results obtained for the tanh activation function.
Number of Neurons TrainingTesting
CostAccuracyCostAccuracy
80.5775.570.3587.05
160.5872.000.3290.18
320.5873.290.4688.62
640.5674.560.3487.18
1280.5474.230.3885.65
2560.5575.830.3889.97
Table 3. The results obtained for the sigmoid activation function.
Table 3. The results obtained for the sigmoid activation function.
Number of Neurons TrainingTesting
CostAccuracyCostAccuracy
80.5272.870.35 94.65
160.5274.000.3884.65
320.5872.890.4184.62
640.5374.000.3893.23
1280.5864.250.3882.35
2561.5552.630.75 90.17
Table 4. The results obtained for the arctan activation function.
Table 4. The results obtained for the arctan activation function.
Number of NeuronsTrainingTesting
CostAccuracyCostAccuracy
80.6767.870.3598.30
160.6764.530.3195.67
320.5572.790.3192.62
640.5874.310.3388.53
1280.5278.250.3487.38
2560.5571.130.35 94.17
Table 5. The confusion matrix from MLP-CNN with Wigner–Ville spectrogram classes.
Table 5. The confusion matrix from MLP-CNN with Wigner–Ville spectrogram classes.
Predicted Label
Background NoiseSingle EEG SignalTwo EEG Signals
True LabelBackground Noise1.000.000.00
Single EEG Signal0.000.960.06
Two EEG Signals0.000.930.57
Table 6. Classification report for MLP-CNN with Wigner–Ville spectrograms.
Table 6. Classification report for MLP-CNN with Wigner–Ville spectrograms.
ClassesPrecisionRecallF1 Score
Background Noise111
Single EEG Signal0.990.980.98
Two EEG Signals0.980.970.97
Average/total0.980.950.96
Table 7. Comparison of methods used in the automatic detection of K-complexes.
Table 7. Comparison of methods used in the automatic detection of K-complexes.
Reference Year Method Used Detection Result
[69]2015Wavelet81.57%
[70]2019Fractal Time–frequency97.00%
[71]2021SVM97.70%
[72]2018Deep Learning94.00%
[73]2020Multitaper Method93.70%
Proposed Method2021Time–Frequency and Deep Neural98.30%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dumitrescu, C.; Costea, I.-M.; Cormos, A.-C.; Semenescu, A. Automatic Detection of K-Complexes Using the Cohen Class Recursiveness and Reallocation Method and Deep Neural Networks with EEG Signals. Sensors 2021, 21, 7230. https://doi.org/10.3390/s21217230

AMA Style

Dumitrescu C, Costea I-M, Cormos A-C, Semenescu A. Automatic Detection of K-Complexes Using the Cohen Class Recursiveness and Reallocation Method and Deep Neural Networks with EEG Signals. Sensors. 2021; 21(21):7230. https://doi.org/10.3390/s21217230

Chicago/Turabian Style

Dumitrescu, Catalin, Ilona-Madalina Costea, Angel-Ciprian Cormos, and Augustin Semenescu. 2021. "Automatic Detection of K-Complexes Using the Cohen Class Recursiveness and Reallocation Method and Deep Neural Networks with EEG Signals" Sensors 21, no. 21: 7230. https://doi.org/10.3390/s21217230

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop