Next Article in Journal
Rapid Detection of Deployment Errors for Segmented Space Telescopes Based on Long-Range, High-Precision Edge Sensors
Previous Article in Journal
Pilot Study: Step Width Estimation with Body-Worn Magnetoelectric Sensors
Previous Article in Special Issue
NeuroSafeDrive: An Intelligent System Using fNIRS for Driver Distraction Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gate Control Mechanisms of Autoencoders for EEG Signal Reconstruction

1
School of Systems and Computing, University of New South Wales, Canberra, ACT 2612, Australia
2
Faculty of Science and Technology, University of Canberra, Canberra, ACT 2617, Australia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3389; https://doi.org/10.3390/s25113389
Submission received: 27 February 2025 / Revised: 17 May 2025 / Accepted: 24 May 2025 / Published: 28 May 2025

Abstract

:
Electroencephalography (EEG) is a non-invasive and portable way to capture neurophysiological activity, which provides the basis for brain–computer interface systems and more innovative applications, from entertainment to security. However, the acquisition of EEG signals often suffers from noise contamination and even signal interruption problems due to poor contact of the electrodes, body movement, or heavy noise. Such heavily contaminated and lost signal segments are usually removed manually, which can hinder practical system deployment and application performance, especially in scenarios where continuous signals are required. In our previous work, we proposed the weighted gate layer autoencoder (WGLAE) and demonstrated its effectiveness in learning dependencies in EEG time series and encoding relationships among EEG channels. The WGLAE adopts a gate layer to encourage the AE to approximate multiple relationships simultaneously by controlling the data flow of each input variable. However, it only applies a sequential control scheme without taking into account the physical meaning of EEG channel locations. In this study, we investigate the gating mechanism for WGLAE and validate the importance of having a proper gating scheme for learning relationships between EEG channels. To this end, several gate control mechanisms are designed that embed EEG channel locations and their corresponding underlying physical meanings. The influences introduced by the proposed gate control mechanisms are examined on an open dataset with different scales and associated with various stimuli. The experimental results suggest that the gating mechanisms have varying influences on reconstructing EEG signals.

1. Introduction

Understanding brain activity patterns is fundamental to unravelling the mysteries of cognitive processes, neurological disorders, and brain-related phenomena. Electroencephalography (EEG) has emerged as a vital tool in the field of neuroscience and brain–computer interfaces, offering a non-invasive means to study and exploit the electrical activity of the brain [1,2]. However, the acquisition of EEG signals is often impeded by various challenges, including noise, artefacts, and missing channels, which cause signal discontinuity and contamination [1,3]. The reconstruction of EEG signals has thus become a critical task, as it aims to enhance the quality and reliability of recorded brain activity data. Enhanced EEG data quality and quantity facilitate more precise analysis, leading to a deeper comprehension of neural dynamics and cognitive functions, and contribute to more reliable and accurate recognition performance in brain–computer interface applications [4].
Traditional methods for EEG signal reconstruction include interpolation, filtering, and time-series analysis. Interpolation techniques, such as linear or cubic spline interpolation [5], aim to estimate missing values by considering the neighbouring data points. While effective in some cases, these methods often oversimplify the complex nature of EEG signals and may introduce inaccuracies, especially in the presence of irregular patterns or abrupt changes. Filtering approaches [6] aim to eliminate noise and artefacts, but they may inadvertently alter the underlying signal or fail to address the non-linear characteristics of EEG data. Time-series analysis methods, such as autoregressive modelling [7], have also been utilised, but they may struggle to capture the intricate dynamics of brain activity.
Recent advancements in machine learning, particularly neural networks, have revolutionised EEG signal reconstruction. Deep learning models, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), have demonstrated remarkable capabilities in capturing intricate temporal and spatial dependencies within EEG signals [8]. Unsupervised learning paradigms further contribute to EEG signal reconstruction when labelled data are scarce and demonstrate flexibility in uncovering diverse patterns, rather than relying on predefined classes. Among the unsupervised learning models, autoencoders (AEs) have demonstrated a strong capacity for inferring latent information and learning complex representations of EEG data [9,10], enabling them to effectively recover EEG signals with higher fidelity.
In our previous study, we proposed the weighted gate layer autoencoder (WGLAE) [11], which is an AE architecture with gate control regularisation for relationship learning. The WGLAE learns to approximate relationships among multiple variables simultaneously through a single model. Specifically, it introduces a gate layer to control the input stream being fed into the AE learner and a weight layer to adjust the update of the reconstruction error according to the importance of the variables. The experimental results demonstrated its effectiveness in reconstructing EEG signals and missing channels. However, our previous work on the WGLAE had a limitation: it only applied a sequential control scheme to turn on and off each input variable/channel sequentially, without considering the physical meaning of the locations of the EEG sensors. It remains an open research question whether the adoption of different gate control mechanisms impacts the learning outcome. Our hypothesis is that having a control strategy that embeds the physical locations of EEG sensors can further enhance signal reconstruction performance.
The integration of EEG electrode positions into the WGLAE model’s learning process is grounded in well-established neurophysiological principles. EEG signals are generated by synchronised neural activity, which propagates through the scalp via volume conduction, causing spatially proximate electrodes to capture overlapping activity from underlying brain regions. Functional brain networks also exhibit spatial organisation, where neighbouring regions often participate in related processes (e.g., occipital electrodes in visual tasks or central electrodes in sensorimotor rhythms). This spatial dependency is leveraged in neurophysiological methodologies, such as source localisation, spatial filtering, and functional connectivity studies. These principles justify our hypothesis: electrode proximity correlates with functional similarity, enabling models to exploit spatial relationships for reconstruction. In this study, we focus on evaluating the impact of the gate control strategies of the WGLAE on EEG signal reconstruction.
The main contributions of this paper are summarised as follows:
  • We propose gate control mechanisms that take into account the underlying physical locations of EEG sensors for regularising the learning process of a deep learning model in learning the inter-channel dependencies of EEG signals.
  • We investigate the gate control mechanisms, including four mechanisms, under different experimental settings. The analysis confirms the impact of gating regularisation on the learning outcomes. Specifically, with the gating sequence embedding the sensors’ physical locations, the WGLAE is able to reconstruct EEG signals with higher accuracy, fidelity, and efficiency.

2. Related Works

Autoencoders, which are a subfamily of neural networks aimed at learning compressed representations of input data via encoding–decoding layers, are considered very useful for different machine learning tasks, including feature extraction, dimensionality reduction, and learning generative models. The architecture of autoencoders can be thought of as a simple feedforward neural network, where the size of both the input layer and output layer is the same. An AE consists of an encoder Y = f ( X ) that generates a latent representation and a decoder Z = g ( Y ) that decodes this latent representation back to the original representation space. The objective is to minimise the difference between the final representation Z and the original input X using a loss function that captures this distance as an error l o s s ( X , g ( f ( X ) ) . The loss functions are the same as those used by the neural network, i.e., MSE, RMSE, MAE, etc. The size of the middle layer determines whether it is an under-complete autoencoder or an over-complete autoencoder. An under-complete autoencoder is used to overcome the potential of an AE learning to copy the input to the output g ( f ( X ) ) = c o p y ( x ) , where it does not learn any significant information. Regularised autoencoders constrain what the autoencoder can learn, where additional constraints are applied to guide the learning process.
Each regularised AE uses specific terms crafted to control the properties of the generated representation. Sparse AEs [12,13] add the sparsity term ( h ) to impose a sparsity penalty in the loss function l o s s ( x , g ( f ( x ) ) ) + λ Ω ( h ) . The sparsity penalty Ω ( h ) = i | h i | is computed using a Laplace prior. Contractive AEs [14] penalise the representation’s sensitivity to the input data using the Frobenius norm of the Jacobian matrix, which generates a localised space contraction with l o s s ( x , g ( f ( x ) ) ) + λ Ω ( h ) , where Ω ( h ) = λ f ( X ) X F 2 . Variational AEs [15] use the Kullback–Leibler divergence principle Ω ( h ) = K L [ Q ( ( z x ) ) | | P ( z ) ] with l o s s ( x , g ( f ( x ) ) ) + λ Ω ( h ) to learn to minimise the difference between the learnt representation distribution and the original sample space distribution. Denoising AEs [16] aim to minimise the distance between the noised input x ˜ and its original form x using a loss function l o s s ( x , g ( f ( x ˜ ) ) .
The GLAE [10] and WGLAE [11] use an input gate layer to guide the autoencoder to learn simultaneous multi-function approximation. This facilitates learning inter-relationships for every variable in the dataset, which can be a complex and time-consuming task for high-dimensional datasets. Both the GLAE and WGLAE have shown superior incomplete data recovery for EEG data compared to CAEs, DAEs, and VAEs when tested using different datasets. Both the GLAE and WGLAE share the same underlying architecture, which gates out the target variable to enforce learning its estimation using the remaining variables. This process is repeated systematically to enable learning for all variables. Therefore, the same sample can be seen multiple times, but with a different target variable each time.
The GLAE and WGLAE architectures enable a systematic teaching process that considers the full training dataset as a curriculum and structures the learning into lessons and sessions, as will be explained further in the next section. A single session is a single pass of a sample with a specific dependent variable gated. A lesson consists of sessions, and the lesson is completed when the number of sessions allows the network to see the same sample the required number of times to learn, approximating every single variable in this sample. However, the original architectures only apply a random sequence scheme without taking into account the physical locations of EEG electrodes. The order of sequential signals is critical due to the temporal structure, while the order of non-sequential signals can be significant in the spatial aspect. As EEG signals reflect both temporal and spatial aspects, we hypothesise that they impact learning performance. Many machine learning methods assume a certain level of order [17] without exploring the actual influence of the data order on performance. Therefore, we will focus on the gate control mechanisms in this study.
In addition to the AE family, recent advances in EEG reconstruction include diffusion models [18] for super-resolution, although their computational demands hinder real-time use, and GAN-based approaches [19,20], which achieve high fidelity but suffer from training instability and artefact sensitivity. Meanwhile, spatiotemporal correlation methods [21] reconstruct missing channels via fixed spatial heuristics, limiting their adaptability to dynamic neural interactions. In contrast, our WGLAE framework can uniquely integrate neurophysiological priors (e.g., electrode proximity and hemispheric symmetry) into lightweight, spatially informed gating mechanisms, thereby balancing efficiency and interpretability. By learning adaptive dependencies through task-specific gates, the WGLAE avoids the computational overhead of diffusion models, the instability of GANs, and the rigidity of correlation-based interpolation, positioning it as a pragmatic solution for real-time EEG applications requiring biologically plausible reconstruction.

3. Methodology

This section provides a brief description of the WGLAE model and proposes layout-inspired gate control mechanisms for it.

3.1. Weighted Gate Layer Autoencoder (WGLAE)

The gate layer autoencoder (GLAE) [10] was introduced to address the limited ability of conventional autoencoders to explicitly model inter-variable dependencies within multivariate data. Standard autoencoders, whether classical or regularised, treat all input dimensions equally and reconstruct the input vector holistically, thereby lacking an inductive bias towards learning functional relationships among variables. The GLAE augments the autoencoder architecture with an input gate layer controlled by a binary mask, allowing the selective suppression of input variables during training. The variables masked in each training iteration are treated as the dependent variables (DVs), while the unmasked variables serve as the independent variables (IDVs) from which the model must infer the DVs. This targeted masking strategy enables the autoencoder to approximate variable-level conditional mappings, promoting the discovery of inter-variable structure. The GLAE organises the training process into hierarchically structured iterations consisting of sessions and lessons. A session corresponds to a single training pass with a specific subset of input variables gated off. A lesson comprises multiple sessions and is completed once every variable in the input vector has been gated and reconstructed at least once. This mechanism ensures comprehensive exposure of the model to the full set of conditional estimation tasks embedded in the dataset. Unlike prior models referred to as “gated autoencoders”, which rely on multiplicative gating and were derived from extensions of Boltzmann machines, the GLAE introduces gating at the architectural level to structure learning. Empirical studies, particularly in EEG signal recovery, have demonstrated the GLAE’s efficacy in reconstructing missing data and capturing latent variable dependencies, with consistent improvements over classical autoencoder variants, including vanilla AEs, denoising AEs, and the variational GLAE [10].
Building on the foundational structure of the GLAE, the weighted gate layer autoencoder (WGLAE) [11] introduced a generalised framework that incorporates variable-specific weighting into both the gating and learning processes. While the GLAE applies uniform binary gating to mask variables, the WGLAE integrates a dynamic weighting mechanism that modulates the contribution of each variable to the reconstruction loss. This is achieved through the addition of two key modules: a weight controller and an error-weight generator. These modules assign distinct weights to DVs and IDVs during training, enabling the model to prioritise learning relationships that are considered more informative or relevant based on task-specific criteria. The WGLAE training pipeline is governed by three synchronised control sequences: (i) data control, which schedules training samples; (ii) gate control, which defines the binary masking of input variables; and (iii) weight control, which determines the loss contribution of each variable. This structure enables session-level modulation of both the input space and the learning signal. Furthermore, the WGLAE introduces a temporal decay mechanism for the weights of dependent variables (DVs), allowing the network to initially emphasise challenging estimation targets before gradually normalising their influence. Through this flexible and adaptive training scheme, the WGLAE achieves improved performance in variable reconstruction, particularly under conditions of structured or large-scale missingness. Comparative evaluations on synthetic and real EEG datasets have confirmed the superiority of the WGLAE over the GLAE and other baseline autoencoder architectures in both reconstruction fidelity and generalisation capacity [11]. Furthermore, the WGLAE was employed as a benchmark in a 2024 study on spatial super-resolution using spatiotemporal diffusion models [18], where its ability to reconstruct occluded channels was explicitly validated.
The WGLAE is an advanced autoencoder architecture designed to enhance dependency learning among input variables. Specifically, the GLAE introduces gate and network controllers, where the gate controller manages input data flow by switching variables ON or OFF using a binary gating sequence, treating switched-off variables as dependent variables (DVs) to be predicted from switched-on independent variables (IDVs). The network controller synchronises the gate and data control sequences, enabling the GLAE to learn variable dependencies beyond traditional data representations. The WGLAE extends the GLAE by incorporating a dynamic error-weighting mechanism, adding flexibility and accuracy in learning. It produces a weight sequence to regulate the error-weight generator, enabling more precise backpropagation by considering the importance of variables, making it suitable for adaptable applications. Figure 1 illustrates the complete architecture of the WGLAE, which includes the following four modules:
  • Input Gate Control Module: This module generates a binary gate control sequence θ j , which is applied to the input layer via element-wise multiplication, forming the gated input:
    xg i j = x i θ j .
    Each element of θ j controls whether the corresponding input variable is activated or deactivated. The deactivated variables are set to zero and excluded from the forward path during training.
  • Network Control Module: This module orchestrates the learning process by generating and synchronising three control sequences: the gate control θ j , the data control γ j , and the weight control Ω j . These sequences determine which variables serve as inputs or targets (IDVs/DVs) and assign corresponding weights. The network dynamically updates these roles across training iterations to enhance the model’s ability to learn inter-variable dependencies.
  • Error Weight Generation Module: The weight control sequence Ω j includes the error weights ω j , which are applied to modulate the contribution of each variable in the loss function. Specifically, ω j decays the influence of the DVs during training, adjusting their learning impact adaptively.
  • Error Calculation Module: The reconstruction x ^ i g is computed via feedforward propagation using the gated input. Using a sigmoid activation function and a single-layer encoding, the reconstructed variable can be computed using:
    xg ^ i j = σ W h o · σ W i h · ( x i θ j ) .
    The weighted reconstruction error is then calculated using:
    wLoss i j = k = 1 N v ω k j · x ^ i , k j x i , k 2 N v .
    This error is used in backpropagation to update the network parameters.
Refer to the original paper [11] for detailed descriptions of the model and algorithm. In this study, we focus on the input gate control module and investigate different gate control mechanisms.

3.2. Gate Control Mechanisms

In the original WGLAE model [11], a random sequence gate control scheme was adopted in the input gate layer. That is, the DVs were selected randomly without considering any dependencies either within the DVs or between the DVs and IDVs. This random scheme demonstrates the resilience of the WGLAE model. However, it may not be able to fulfil the potential of the WGLAE model, thus making it less efficient in applications. Inspired by developments in neuroscience, three gate control mechanisms are proposed—order-based, partition-based, and hemispheric symmetry rotation gate control mechanisms—in addition to the random scheme adopted in the original WGLAE model [11]. While the random scheme applies a random permutation, the three proposed gate control mechanisms incorporate the order, partition, and hemispheric symmetry features of brain signals corresponding to EEG electrode placement.
From a machine learning perspective, integrating EEG electrode positions into the gating mechanism of our model is motivated by structured inductive bias [22], a fundamental concept in machine learning, where prior knowledge about data structure guides model design to improve generalisation. EEG signals exhibit spatial dependencies due to volume conduction (where signals from a neural source propagate to multiple nearby electrodes) and functional correlations between adjacent brain regions. By encoding electrode locations into the gating mechanism, we explicitly introduce a spatial prior that reduces the hypothesis space and enables efficient feature interaction. While neurophysiological studies support the spatial nature of EEG, our work demonstrates that explicitly encoding the spatial structure as an inductive bias enhances model robustness and efficiency in a purely data-driven setting.

3.2.1. The Order-Based Gate Control Mechanism

Brain lateralisation, also known as hemispheric specialisation, refers to the functional dominance of one hemisphere of the brain [23]. It is involved in many advanced neural functions, including cognitive performance, memory, and handedness [23]. As pointed out in [24], brain asymmetries are significant to the functioning of the sensory, cognitive, and motor systems in humans and other animals. For instance, when conducting analytical or logical tasks, active beta waves can be observed in the left hemisphere. Therefore, it is fair to assume that EEG electrodes located at the left scalp region exhibit stronger interdependence than those at the right scalp region. For this reason, it is essential to learn the dependencies of EEG signals in the left brain hemisphere on those in the right hemisphere. Additionally, EEG electrodes generally have a stronger influence on nearby electrodes than on those farther away. Inspired by this idea, the order-based gate control mechanism first selects the DVs corresponding to the EEG electrode(s) located at the front-left scalp region, followed by those in the middle-left and back-left regions, and then applies the same sequence on the right scalp region.
For a clearer explanation, the order-based gate control mechanism is explained using the international EEG scalp electrode system. Figure 2 shows a head map of EEG electrode placements derived from the international 10-10 system and international 10-20 system (blue circles). In the international 10-10 system, there are 64 EEG electrodes in total, whereas in the international 10-20 system, there are 19 electrodes. Each electrode is labelled based on the lobe and specific area of the brain it records. The lobe location is indicated by letters and numbers. For letters, Fp stands for fronto-polar, F for frontal, C for central, T for temporal, P for parietal, and O for occipital. The numbers denote lateralisation. In particular, electrodes located on the right brain hemisphere are indicated by even numbers, while those on the left brain hemisphere are indicated by odd numbers.
As shown in the head map, the order-based gate control mechanism selects DVs from the top-left section to the bottom-left section, followed by the top-right and bottom-right sections. The mechanism is exemplified in Figure 3a. A full list of the order-based sequence, based on the international 10-20 system, is presented in Table 1.

3.2.2. The Partition-Based Gate Control Mechanism

A human’s cerebral cortex is divided into different lobes, as illustrated in the legend of Figure 3. Each cortex lobe is responsible for processing certain information and is associated with specific functions [26]. Importantly, the pattern of brain waves can be closely related to neural functions. For instance, gamma-band oscillations (30–80 Hz) play a crucial role in attention and consciousness [27]. Assuming that brain signals located in the same lobe are more dependent on each other, it is expected to be more efficient to learn the relationships among signals generated by different cortex lobes. Inspired by this assumption, the proposed partition-based gate control mechanism sets the sequence of DVs based on the cortex lobes.
In the partition-based gate control mechanism, the channels at the fronto-polar lobe are set as the DVs, and their dependency on the channels associated with the other cortical lobes is learned first, followed by the channels located at the frontal and central lobes. The channels located at the parietal and occipital lobes are learned last. A complete list of the sequence based on the 19-channel EEG electrode system is presented in Table 1. With the partition-based gate control mechanism, channel dependency within and across lobes is learned when very few channels are missing, and the dependency of channels in one lobe on those in other lobes is learned when more channels are missing.

3.2.3. The Hemispheric Symmetry Rotation Gate Control Mechanism

Despite its functional asymmetry, the human brain is a structurally mirrored organ [28] and inherently displays bilateral symmetry [23]. In certain circumstances, such as a restful state with closed eyes, brain waves can show symmetrical patterns of activity across the hemispheres [29]. This symmetrical brain feature promotes the efficiency of data representation, which further improves the capability for information processing and responding [30].
Considering the anatomical symmetry of the left and right hemispheres of the brain [23], it can be assumed that the brain waves captured by EEG electrodes located in symmetrical regions have a stronger correlation. Accordingly, the hemispheric symmetry rotation gate control mechanism explores the feasibility and benefit of learning channels based on symmetry. As depicted in Figure 3c, the gate control sequence starts from the EEG electrode located at the fronto-polar region on the left side, followed by the symmetrical EEG electrode on the right side.
Similar to the partition-based gate control mechanism, the gate control sequence in the hemispheric symmetry rotation mechanism is based on the cerebral cortex areas. However, within each cortical area, the hemispheric symmetry rotation mechanism prioritises symmetry over proximity. The full channel sequence of the hemispheric symmetry rotation mechanism, based on the international 10-20 system, is exemplified in Table 1. It can be seen from the EEG electrode labels that even numbers and odd numbers appear alternately. This is because a left EEG electrode is always followed by the electrode placed either on the right side or in the central area in this mechanism.

3.2.4. Summary

The gating mechanism in our WGLAE framework functions as a structured form of regularisation, explicitly encoding domain-specific priors to guide the model towards learning neurophysiologically meaningful dependencies. By constraining interactions based on spatial proximity (e.g., partition-based gates), functional symmetry (e.g., hemispheric gates), or task relevance, the gate introduces inductive biases that reflect neurophysiological principles such as volume conduction and modular brain networks. This structured sparsity suppresses noisy or implausible correlations, reduces overfitting, and prioritises neurophysiologically meaningful relationships. Empirically, gates with such priors outperform unstructured variants [31], validating their role in enhancing robustness and generalisability. Thus, the gate operates not merely as a control mechanism but as a neurophysiologically grounded regulariser, bridging data-driven learning with domain knowledge to ensure biologically plausible signal reconstruction.

4. Experimental Design

4.1. Database and Pre-Processing

The proposed method was evaluated using the publicly available EEG Motor Movement/Imagery Database (MMIDB) [32]. The MMIDB contains EEG recordings from 109 healthy subjects under various signal elicitation protocols [33], including resting states with eyes open and closed, motor-movement tasks involving opening and closing fists or feet, and motor imagery tasks requiring subjects to imagine these movements. This database is widely used in EEG classification studies due to its diversity in subject count and recording conditions [34,35]. The EEG signals were recorded using the BCI2000 system [36], which features 64 scalp-mounted sensors following the international 10-10 system. The system operated at a 160 Hz sampling rate, using the earlobes as signal references.
For EEG signal pre-processing, we implemented an automated pipeline consisting of filtering, artefact removal, and segmentation. First, the raw EEG data were band-pass filtered between 0.5 Hz and 42 Hz to preserve the canonical EEG frequency range. Next, independent component analysis (ICA) and the Multiple Artefact Rejection Algorithm (MARA) [37] were applied to eliminate artefacts caused by eye movements, muscle activity, and sensor contact issues. Finally, a non-overlapping sliding window was used to segment the continuous signals into two-second segments.
While the MMIDB provides a controlled benchmark for evaluating gating mechanisms, future studies will test the WGLAE’s generalisability on clinical EEG datasets with diverse pathologies and noise profiles. The framework’s integration of neurophysiological priors positions it as a promising candidate for cross-dataset adaptation, particularly when combined with proper domain adaptation strategies.

4.2. Experiments

In this paper, the WGLAE with the different gating control mechanisms was used to learn the relationships between EEG channels. In order to investigate how the physical locations of EEG channel(s) affect the learning capability of the WGLAE, the EEG channels were blocked in different sequences and in different quantities when training the WGLAE. Then, the trained WGLAE model was used to reconstruct the EEG data with incomplete channels. The proposed approach was evaluated through the following: an evaluation of the impact of gate control mechanisms on reconstruction performance, a scalability analysis, and an evaluation of reconstructed EEG in classification tasks.
In the reconstruction performance experiment and the scalability analysis, the parameter settings were the same for all datasets to ensure fair comparisons. The WGLAE model consisted of a single hidden layer, with the number of neurons set to half the input size to build an under-complete structure of the AE. Based on the preliminary work in [11], the WGLAE model was configured as follows: the activation function was sigmoid, the optimiser was ADAM [38], the learning rate was 0.00001, the total number of epochs was 10,000, and the batch size was 1024.
The data setup of the WGLAE model followed the widely used 80% to 20% ratio. That is, for each dataset, the training data accounted for 80% of the total data, while the testing data accounted for 20%. In addition, to avoid overfitting, 20% of the training data were randomly selected to serve as validation data. The same training, validation, and testing datasets were employed across all four gating strategies.
The training and learning process conformed to the curriculum, session, and lesson framework proposed in [11]. In a curriculum, a specific gate control mechanism is applied, and the number of missing channels is fixed. Within one curriculum, each session is assigned one data sample, under which different lessons are in charge of different locations of the missing channels (DVs). Therefore, the number of sessions in one curriculum equals the number of samples, and the WGLAE learns multiple sessions simultaneously. The number of lessons N l e s s o n in one session is determined by the number of channels N D V + I D V and the number of missing channels N D V :
N l e s s o n = N D V + I D V / N D V
where is the floor function. Reconstruction performance was evaluated based on the differences between the reconstructed data and the original complete data, and the average reconstruction errors of 10 runs were employed.

4.2.1. Impact of Gate Control Mechanisms on Reconstruction Performance

The first set of experiments evaluated the reconstruction errors of WGLAE models trained with different gate control mechanisms. The reconstruction errors from the three proposed mechanisms, i.e., partition-based, hemispheric symmetry rotation, and order-based, were compared with those from the random sequence gate control mechanism. This experiment assessed how different gate control strategies influence the accuracy and quality of EEG signal reconstruction, providing insights into their effectiveness in modelling inter-channel dependencies.
To evaluate the reconstruction performance of the three mechanisms under different scenarios, the experiments were conducted based on two EEG electrode configurations: the international 10-20 system, where 19 channels of signals were considered, and the extended system, where 29 channels were examined. The EEG electrode placements for the two configurations are presented in Figure 2, where the blue circles indicate the configuration based on the international 10-20 system, and the yellow circles denote the additional 10 channels of the extended system. The comparison results are expected to enhance our understanding of the correlations among EEG channels.

4.2.2. Scalability Analysis

The scalability analysis aimed to examine the influence of the number of blocked channels (DVs) on the reconstruction error. When more channels are blocked, the WGLAE receives less ‘supervision’ during training, and thus learns relatively less about the EEG data. However, a resilient model should still maintain a certain level of reconstruction accuracy. The scalability analysis evaluated the model across varying numbers and configurations of missing channels, demonstrating its adaptability to different EEG electrode configurations and real-world applications.

4.2.3. Evaluation of Reconstructed EEG in Classification Tasks

The reconstruction error reflects the discrepancy between the original signal and its reconstructed version in the time domain, providing a measure of the accuracy and fidelity of the reconstruction process. While the reconstruction error is a valuable metric for evaluating signal reconstruction performance, it offers limited insight into data patterns and biological relevance [39]. For example, the reconstruction error does not inherently capture whether the reconstructed signal retains meaningful information about neural processes or cognitive states. Hence, in this experiment, we evaluated the usability of EEG data reconstructed under different gating mechanisms using a classification task: biometric identification [34,40]. The Motor Movement/Imagery Database was used for this evaluation.
Classification performance can reflect the extent to which the reconstructed data capture the underlying patterns identified by the learning model in the original data. In this task, we classified EEG signals into 109 classes corresponding to the identities of the 109 subjects. A convolutional neural network (CNN) model was trained using 80% of the original data. For the remaining 20% of testing data, we randomly discarded one, five, and ten channels and replaced them with reconstructed data using the proposed method to simulate scenarios where certain channel data are lost in real applications. Performance was evaluated using the Correct Recognition Rate (CRR), which indicates the classification accuracy. The details of the classification model configurations are summarised in Table 2. The training process used the ADAM optimiser with a learning rate of 0.0005, a batch size of 32, 200 training epochs with early stopping, batch normalisation, and a dropout rate of 0.1 after each pooling layer.

5. Experimental Results Analysis

5.1. Reconstruction Accuracy

The reconstruction accuracy evaluates the credibility of the proposed WGLAE model by calculating the mean squared difference between the original and reconstructed signals. The performance of the three gate control mechanisms, as well as the random sequence gate control mechanism, was examined based on the international EEG placement system with 19 channels. The MMIDB with three elicitation protocols, including resting states, motor imagery, and motor movement, was employed.
The reconstruction errors across the entire sample are presented in Table 3, and the best-performing gate control mechanism in each case is highlighted. For all the examined gate control mechanisms and datasets, the reconstruction accuracy decreased as more channels were blocked. This aligns with real-world experiences; that is, as missing information increased, it became more challenging to reconstruct EEG signals. Another common phenomenon observed across all three datasets was that the three proposed gating mechanisms performed better than the random sequence gating mechanism in most cases. This indicates the necessity of considering EEG electrode placement when designing algorithms for signal reconstruction.
On the other hand, the results demonstrated the various influences that the three gate control mechanisms had on the reconstruction performance of the WGLAE. On the resting-state and motor-imagery datasets, the hemispheric symmetry rotation gate control mechanism outperformed the other three mechanisms in 9 out of 10 different quantities of missing channels. This observation can be explained by the brain’s alpha waves, which are active in states of relaxed and passive attention [29]. Particularly in an eyes-closed state, alpha waves were typically balanced between the left and right hemispheres, suggesting similarities and dependencies between the EEG signals captured by the left electrodes and those captured by the right electrodes. As the hemispheric symmetry mechanism aims to learn EEG signals obtained from symmetrical areas, it should facilitate better learning capabilities for the WGLAE. A different trend was observed in the motor-movement dataset, where the partition-based gate control mechanism exhibited the best performance in 8 out of 10 cases. This could be due to the active beta and gamma waves in the brain, which are closely related to movement execution and coordination. As beta waves and gamma waves show dominance in various cerebral areas when undertaking different tasks, learning the dependencies between different cerebral cortex lobes can be significant for building an effective EEG signal reconstruction model. The advantages observed in the hemispheric symmetry rotation and partition-based gate control mechanisms validate the benefits of employing gate control mechanisms. The disadvantage of the hemispheric symmetry rotation gate control mechanism can be justified by the lateralisation of motor functions such as handedness [23].
The results suggest that the application context should be taken into account when selecting gate control mechanisms. Specifically, in scenarios involving resting states, the hemispheric symmetry rotation gate control mechanism should be prioritised, while the partition-based gate control mechanism is more suitable when motor movement is involved. The observed differences in the efficacy of the gating mechanisms across tasks can be attributed to distinct functional brain connectivity patterns associated with resting states versus motor movements. Resting-state EEG was characterised by synchronised activity within default mode networks, which spanned the medial prefrontal and posterior regions. These networks exhibited strong long-range and inter-hemispheric coherence, reflecting idling cortical activity. The hemispheric symmetry rotation gate control mechanism, which explicitly models interactions between homologous brain regions (e.g., left vs. right motor cortex), aligns with this bilateral synchronisation. By emphasising cross-hemispheric dependencies, this gate effectively captured the large-scale functional connectivity inherent to resting states. On the other hand, motor tasks (e.g., hand movement) engaged localised, task-specific networks primarily within the motor cortex. These regions exhibited short-range and intra-hemispheric coherence. The partition-based gate control mechanism, which groups electrodes into localised clusters (e.g., frontal, central, or occipital), prioritised short-range spatial relationships. This design mirrored the organisation of the motor cortex, where adjacent electrodes captured activity from functionally related representations, enhancing the reconstruction of task-specific signals. The proposed gating strategies aligned with the neural mechanisms. Specifically, the hemispheric symmetry rotation gate control mechanism acted as a coarse-grained prior for global connectivity, while the partition-based gate control mechanism imposed a fine-grained prior for local connectivity. Their task-dependent efficacy suggests that the model successfully leveraged the spatial structure of functional networks in a data-driven setting.
The standard deviations in some cases were relatively high, for example, when reconstructing 10/19 channels for motor-movement data. An analysis of the extreme cases revealed the context-dependent efficacy of the gate mechanisms. For example, the partition-based gate mechanism exhibited larger deviations when critical task-relevant channels (e.g., C3 for right-hand motor imagery) were blocked, underscoring the need for adaptive strategies in future work.
To understand how the different gate control mechanisms influenced the reconstruction accuracy in cerebral cortex regions, the reconstruction error distributions over the scalp are presented in Figure 4. It can be observed that the central and posterior areas show higher errors compared to the frontal area, and this trend is consistent across the different gate control mechanisms. This can be explained by the brain functions associated with the regions of the cerebral cortex. Specifically, the frontal lobe is responsible for executive functions and motor control [41], which are active from stimuli incorporated into the MMIDB. On the other hand, the occipital lobe is mainly in charge of visual processing [41], which might not have been intensively activated by the three elicitation protocols.
To assess spectral fidelity, we conducted post hoc frequency-domain analyses to validate spectral preservation. These analyses focused on quantifying the fidelity of oscillatory dynamics across the five canonical EEG frequency bands: delta (1–4 Hz), theta (4–8 Hz), alpha (8–13 Hz), beta (13–30 Hz), and gamma (30–50 Hz). Specifically, we computed the power spectral density (PSD) for both the original and reconstructed signals using Welch’s method and estimated band-specific power. Then, Pearson’s correlation coefficients were calculated between the PSD of the original and reconstructed signals across all channels and subjects to assess the global spectral alignment. The normalised RMSE (nRMSE) was computed for each frequency band to quantify the deviations in the reconstructed band power. Paired t-tests were performed to evaluate the significance of the band-power differences between the gating strategies. The results of the spectral fidelity analysis are summarised in Table 4.
For the resting-state data, the WGLAE framework demonstrated better reconstruction performance using the hemispheric symmetry rotation gating scheme, achieving higher spectral correlation and lower spectral nRMSE compared to the random sequence gating scheme, with statistically significant differences (p < 0.05). Similarly, the partition-based gating scheme achieved better spectral fidelity than the random sequence gating scheme for motor-movement data reconstruction. Although marginal in absolute value, these improvements reflect meaningful noise suppression, particularly in middle- and high-frequency bands (alpha, beta, and gamma). The results show strong preservation of oscillatory dynamics and demonstrate the importance of neurophysiologically informed gating in capturing task-relevant spatial dependencies while mitigating noise. As many EEG features are in the frequency domain, spectral fidelity is important in downstream tasks such as sleep-stage classification, cognitive load assessment, and other BCI control applications. While our model applies time-domain metrics in the learning process, minimising amplitude errors inherently enforces implicit spectral fidelity. This is supported by studies showing strong correlations between time-domain reconstruction accuracy and spectral preservation in autoencoder-based EEG models [42].
While neurophysiological studies validate the spatial basis of EEG signals (e.g., interpolation of missing channels using splines), our work computationally confirms this by demonstrating improved reconstruction when spatial gates encode electrode locations. Our experimental results align with prior neurophysiological findings, showing that models incorporating spatial information outperform random approaches. Thus, our hypothesis is supported by both established literature and empirical validation in this study.

5.2. Scalability Analysis

The capabilities of the proposed gate control mechanisms were further tested on the MMIDB with 29 channels. The placement of the 29 channels is illustrated in Figure 2. In this section, the reconstruction accuracy obtained by different gate control mechanisms is discussed, followed by a comparison with the accuracy obtained on the dataset with 19 channels.
As presented in Table 5, the hemispheric symmetry rotation, partition-based, and order-based gate control mechanisms outperformed the random sequence gate control mechanism on 26, 25, and 26 out of 30 instances, respectively. Similar to the observations on the 19-channel dataset, the hemispheric symmetry rotation gate control mechanism performed the best on the resting-state dataset, while the partition-based gate control mechanism showed dominant advantages on the motor-movement dataset. This observation is similar to that on the 19-channel dataset, thus validating the benefits of aligning EEG sensor locations with the features of active brainwaves. The difference between the performance on the 19-channel and 29-channel datasets lies in the motor-imagery protocol. In this elicitation protocol, the partition-based gate control mechanism exhibited the best performance on the 29-channel dataset, while the hemispheric symmetry rotation gate control mechanism exhibited the best performance on the 19-channel dataset. Therefore, the features of datasets, especially the data size, should also be considered when deciding on the most fitting gate control mechanism.
For the scalability analysis, we evaluated the performance of the WGLAE model with different gate control mechanisms in reconstructing different numbers of channels from 1 to 10. The mean reconstruction errors on the 19-channel and 29-channel datasets are presented in Figure 5. With an increasing number of missing channels (DVs), the reconstruction errors from all four gate control mechanisms showed an upward trend. This upward trend is consistent with the observations on WGLAE models, where more input information generally yields a better model. At the same time, it is worth noting that the reconstruction accuracy was consistently higher than 91%, even when more than half of the EEG channels were blocked during WGLAE training. This shows the resilience and reliability of the WGLAE model in recovering real-world EEG signals. The advantages of the proposed gate control mechanisms over the random sequence gate control mechanism tended to decline with an increasing number of missing channels. This is because when many channels are blocked simultaneously, it is futile to consider the functions of the cortex lobes, as the channels in the lobe are completely missing. This is also the reason why this performance deterioration was less pronounced in the 29-channel instances.
To examine the statistical difference between the performance of the random sequence gate control mechanism and that of the hemispheric symmetry rotation, partition-based, and order-based gate control mechanisms, a Wilcoxon signed-rank test was conducted [43]. A holistic comparison was made based on the reconstruction errors obtained in 10 scenarios (with 1 to 10 missing channels) of 10 runs. The significance values in Table 6 show that both the hemispheric symmetry rotation and partition-based gate control mechanisms were significantly different from the random sequence gate control mechanism. However, the order-based gate control mechanism was not significantly different from the random sequence gate control mechanism in the 19-channel configuration, which can be explained by the limited information available compared to that in the 29-channel configuration. On the other hand, it aligns with the aforementioned observations and indicates that the hemispheric symmetry rotation and partition-based gate control mechanisms were more effective in capturing signal patterns.
With our hardware setup, a computer with an i7-8550U CPU and 8 GB of RAM, the computational time for reconstructing the 19-channel dataset was between 0.08 and 0.16 s and between 0.15 and 0.2 s for the 29-channel dataset during the inference stage. This inference speed aligns well with the requirements for efficient and low-latency deployment. The proposed model adopts a structured gating mechanism on a lightweight autoencoder backbone, which ensures fast forward passes, with inference time comparable to that of standard simple neural networks or linear models.
The experiments adopted 19- and 29-channel configurations, which are common in clinical and BCI applications. The WGLAE’s structured gating principles are theoretically scalable to high-density systems. In future work, we will evaluate the model on high-density setups and refine the gating mechanisms for high-channel configurations, including hierarchical partitioning and dynamic sparsity adaptation.

5.3. Classification Performance

The utility of the reconstructed EEG signals was examined by applying them to classification problems, validating the model’s practical effectiveness in downstream tasks. Table 7 summarises the classification results, which complement the reconstruction error results and provide a more comprehensive understanding of the quality and fidelity of the reconstructed data. An increased number of missing channels led to a decline in classification accuracy. This was expected, as missing channels reduce the available information and impair the model’s ability to extract meaningful patterns. The decline appears more gradual when fewer channels are missing (1 or 5) but becomes more pronounced as more data are lost. The results show the effect of the different gate control mechanisms on classification accuracy. Across all conditions, the partition-based and hemispheric symmetry rotation gate control mechanisms consistently performed better than the order-based and random sequence gate control mechanisms, which suggests that structured control over missing data can be beneficial. The performance gap between the different mechanisms increased as more channels were missing, which demonstrates the importance of a structured approach to handling missing data. The results suggest that the proposed gate control mechanisms can help mitigate the effects of missing EEG channels, making classification systems more robust.
While subject identification validates the retention of discriminative features, future work will test the WGLAE’s clinical efficacy by integrating reconstructed signals into applications such as seizure detection by leveraging open datasets such as CHB-MIT.

6. Conclusions and Future Work

To facilitate reliable EEG signal reconstruction, this study proposed various gate control mechanisms based on the effective WGLAE model [11]. By considering the physical locations of EEG sensors and the corresponding brain functions, the three proposed gate control mechanisms exploit different features of brain activities. In particular, the order-based gate control mechanism is inspired by the brain lateralisation phenomenon, the partition-based gate control mechanism considers various functions related to different cerebral lobes, and the hemispheric symmetry rotation gate control mechanism emphasises the largely inherent symmetry of the brain structure and brain wave patterns in certain circumstances. The incorporation of such preliminary knowledge about brain waves is expected to enhance the applicability and effectiveness of the WGLAE, hence contributing to the understanding of cognitive processes.
The three proposed gate control mechanisms were examined on the commonly used MMIDB with three signal elicitation protocols, i.e., resting states, motor movement, and motor imagery. The experimental results on the dataset with 19 EEG channels showed that the three proposed gate control mechanisms dominated the random sequence gate control in most instances, although the best mechanism varied for different protocols. The scalability analysis, based on the extended dataset with 29 EEG channels, further validated the effectiveness of gating regularisation in deep learning models. Across the 19-channel and 29-channel datasets, the hemispheric symmetry rotation gate control mechanism performed best in the resting-state dataset, while the partition-based gate control mechanism showed advantages in the motor-movement dataset. This can be explained by the different brain activity patterns for different stimuli. In addition, the reconstructed EEG signals were applied to classification problems. The consistent superiority shown by the partition-based and hemispheric symmetry rotation gate control mechanisms in classification problems further validates the effectiveness of the proposed gate control mechanisms.
In this study, a classical AE backbone was chosen to isolate the efficacy of the gating mechanisms. However, future work will explore advanced architectures (e.g., VAEs) and hybrid models that integrate gates with adversarial or variational components for enhanced performance and artefact suppression. Moreover, future research could explore adaptive mechanisms that dynamically update the gating strategies proposed in this study, based on signal characteristics or task demands. Such approaches, potentially leveraging attention or meta-learning frameworks, may further optimise reconstruction performance across diverse EEG paradigms. In addition, future iterations could integrate domain-specific channel weights into the gating mechanism to align the model with neurophysiological principles and enhance task-specific reconstruction fidelity.
In terms of evaluation, future work will evaluate patient-independent generalisation using leave-one-subject-out validation to assess robustness against anatomical and noise variability across individuals. The WGLAE’s structured gates, which encode neurophysiological priors (e.g., spatial proximity and symmetry), are theoretically positioned to enhance cross-subject generalisation by prioritising conserved neurophysiological principles rather than subject-specific artefacts. Empirical validation of this hypothesis remains a critical next step. Meanwhile, the model’s robustness against dynamic artefacts (e.g., muscle noise or electrode drift) in practical settings is worth further examination. It would also be interesting to extend the work to other physiological data, such as ECG and EMG. While this study focuses on EEG, the gating framework’s reliance on structured priors suggests applicability to these signals by adapting spatial and temporal constraints to their unique properties [44,45].
Our model’s ability to learn spatiotemporal dependencies positions it as a promising alternative to purely spatial (e.g., spline) or temporal (e.g., AR) interpolation, particularly for scenarios requiring joint reconstruction of multi-channel, time-continuous signals. While this study has focused on evaluating gate control mechanisms within the WGLAE framework, future work will include direct comparisons with classical EEG reconstruction methods (e.g., spline interpolation) to further validate the practical advantages of our approach in both spatial and temporal recovery tasks. The proposed model also offers promising potential for applications such as EEG channel density enhancement [46] and data augmentation [47], both of which have proven critical for improving the performance of machine learning models, particularly deep neural networks, in EEG classification tasks.

Author Contributions

Conceptualization, M.W. and H.E.-F.; methodology, M.W. and H.E.-F.; validation, K.L., M.W. and H.E.-F.; investigation, K.L. and M.W.; resources, H.E.-F.; data curation, M.W. and K.L.; writing—original draft preparation, K.L., M.W. and H.E.-F.; writing—review and editing, M.W., K.L. and H.E.-F.; visualization, M.W., K.L. and H.E.-F.; supervision, M.W. and H.E.-F.; project administration, H.E.-F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the University of New South Wales, Canberra, through an Internal Research Funding Program.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in PysioNet EEG Motor Movement Imagery Dataset, DOI: 10.13026/C28G6P. The code is openly available in Weighted Gate Layer Autoencoders, Code Ocean Capsule: 4751126.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bablani, A.; Edla, D.R.; Tripathi, D.; Cheruku, R. Survey on brain-computer interface: An emerging computational intelligence paradigm. ACM Comput. Surv. (CSUR) 2019, 52, 20. [Google Scholar] [CrossRef]
  2. Mridha, M.F.; Das, S.C.; Kabir, M.M.; Lima, A.A.; Islam, M.R.; Watanobe, Y. Brain-computer interface: Advancement and challenges. Sensors 2021, 21, 5746. [Google Scholar] [CrossRef]
  3. Chaddad, A.; Wu, Y.; Kateb, R.; Bouridane, A. Electroencephalography signal processing: A comprehensive review and analysis of methods and techniques. Sensors 2023, 23, 6434. [Google Scholar] [CrossRef]
  4. Gu, X.; Cao, Z.; Jolfaei, A.; Xu, P.; Wu, D.; Jung, T.P.; Lin, C.T. EEG-based brain-computer interfaces (BCIs): A survey of recent studies on signal sensing technologies and computational intelligence approaches and their applications. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 1645–1666. [Google Scholar] [CrossRef] [PubMed]
  5. Abidi, A.; Nouira, I.; Assali, I.; Saafi, M.A.; Bedoui, M.H. Hybrid multi-channel EEG filtering method for ocular and muscular artifact removal based on the 3D spline interpolation technique. Comput. J. 2022, 65, 1257–1271. [Google Scholar] [CrossRef]
  6. Ranjan, R.; Sahana, B.C.; Bhandari, A.K. Motion artifacts suppression from EEG signals using an adaptive signal denoising method. IEEE Trans. Instrum. Meas. 2022, 71, 4000410. [Google Scholar] [CrossRef]
  7. Haderlein, J.F.; Peterson, A.D.; Burkitt, A.N.; Mareels, I.M.; Grayden, D.B. Autoregressive models for biomedical signal processing. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023; pp. 1–6. [Google Scholar]
  8. Wang, M.; Yin, X.; Zhu, Y.; Hu, J. Representation learning and pattern recognition in cognitive biometrics: A survey. Sensors 2022, 22, 5111. [Google Scholar] [CrossRef]
  9. Wang, M.; Abdelfattah, S.; Moustafa, N.; Hu, J. Deep Gaussian mixture-hidden Markov model for classification of EEG signals. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 278–287. [Google Scholar] [CrossRef]
  10. El-Fiqi, H.; Kasmarik, K.; Bezerianos, A.; Tan, K.C.; Abbass, H.A. Gate-layer autoencoders with application to incomplete EEG signal recovery. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  11. El-Fiqi, H.; Wang, M.; Kasmarik, K.; Bezerianos, A.; Tan, K.C.; Abbass, H.A. Weighted gate layer autoencoders. IEEE Trans. Cybern. 2021, 52, 7242–7253. [Google Scholar] [CrossRef]
  12. Ranzato, M.A.; Boureau, Y.L.; LeCun, Y. Sparse Feature Learning for Deep Belief Networks. In Proceedings of the 20th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; Curran Associates Inc.: Red Hook, NY, USA, 2007; pp. 1185–1192. [Google Scholar]
  13. Goodfellow, I.J.; Le, Q.V.; Saxe, A.M.; Lee, H.; Ng, A.Y. Measuring Invariances in Deep Networks. In Proceedings of the 22Nd International Conference on Neural Information Processing Systems, Istanbul, Turkey, 9–12 November 2015; Curran Associates Inc.: Red Hook, NY, USA, 2009; pp. 646–654. [Google Scholar]
  14. Rifai, S.; Vincent, P.; Muller, X.; Glorot, X.; Bengio, Y. Contractive Auto-encoders: Explicit Invariance During Feature Extraction. In Proceedings of the 28th International Conference on International Conference on Machine Learning, Washington, DC, USA, 28 June–2 July 2011; Omnipress: Madison, WI, USA, 2011; pp. 833–840. [Google Scholar]
  15. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  16. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.A. Extracting and Composing Robust Features with Denoising Autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; ACM: New York, NY, USA, 2008; pp. 1096–1103. [Google Scholar] [CrossRef]
  17. Winter, R.; Noé, F.; Clevert, D.A. Permutation-invariant variational autoencoder for graph-level representation learning. Adv. Neural Inf. Process. Syst. 2021, 34, 9559–9573. [Google Scholar]
  18. Wang, S.; Zhou, T.; Shen, Y.; Li, Y.; Huang, G.; Hu, Y. Generative AI Enables EEG Super-Resolution via Spatio-Temporal Adaptive Diffusion Learning. IEEE Trans. Consum. Electron. 2025. [Google Scholar] [CrossRef]
  19. Vo, K.; Vishwanath, M.; Srinivasan, R.; Dutt, N.; Cao, H. Composing graphical models with generative adversarial networks for EEG signal modeling. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 1231–1235. [Google Scholar]
  20. Lee, W.; Lee, J.; Kim, Y. Contextual imputation with missing sequence of EEG signals using generative adversarial networks. IEEE Access 2021, 9, 151753–151765. [Google Scholar] [CrossRef]
  21. Bahador, N.; Jokelainen, J.; Mustola, S.; Kortelainen, J. Reconstruction of missing channel in electroencephalogram using spatiotemporal correlation-based averaging. J. Neural Eng. 2021, 18, 056045. [Google Scholar] [CrossRef]
  22. Goyal, A.; Bengio, Y. Inductive biases for deep learning of higher-level cognition. Proc. R. Soc. A 2022, 478, 20210068. [Google Scholar] [CrossRef]
  23. Wang, J.; Ma, S.; Yu, P.; He, X. Evolution of Human Brain Left–Right Asymmetry: Old Genes with New Functions. Mol. Biol. Evol. 2023, 40, msad181. [Google Scholar] [CrossRef]
  24. Güntürkün, O.; Ströckens, F.; Ocklenburg, S. Brain lateralization: A comparative perspective. Physiol. Rev. 2020, 100, 1019–1063. [Google Scholar] [CrossRef]
  25. Perera, P.; Harshani, H.; Shiratuddin, M.F.; Wong, K.W.; Fullarton, K. EEG signal analysis of writing and typing between adults with dyslexia and normal controls. Int. J. Interact. Multimed. Artif. Intell. 2018, 5, 62. [Google Scholar] [CrossRef]
  26. Mendoza-Halliday, D.; Major, A.J.; Lee, N.; Lichtenfeld, M.J.; Carlson, B.; Mitchell, B.; Meng, P.D.; Xiong, Y.; Westerberg, J.A.; Jia, X.; et al. A ubiquitous spectrolaminar motif of local field potential power across the primate cortex. Nat. Neurosci. 2024, 27, 547–560. [Google Scholar] [CrossRef]
  27. Kim, T.; Thankachan, S.; McKenna, J.T.; McNally, J.M.; Yang, C.; Choi, J.H.; Chen, L.; Kocsis, B.; Deisseroth, K.; Strecker, R.E.; et al. Cortically projecting basal forebrain parvalbumin neurons regulate cortical gamma band oscillations. Proc. Natl. Acad. Sci. USA 2015, 112, 3535–3540. [Google Scholar] [CrossRef]
  28. Saltoun, K.; Adolphs, R.; Paul, L.K.; Sharma, V.; Diedrichsen, J.; Yeo, B.T.; Bzdok, D. Dissociable brain structural asymmetry patterns reveal unique phenome-wide profiles. Nat. Hum. Behav. 2023, 7, 251–268. [Google Scholar] [CrossRef]
  29. Kurgansky, A. Functional organization of the human brain in the resting state. Neurosci. Behav. Physiol. 2019, 49, 1135–1144. [Google Scholar] [CrossRef]
  30. Anselmi, F.; Patel, A.B. Symmetry as a guiding principle in artificial and brain neural networks, volume II. Front. Comput. Neurosci. 2024, 18, 1527725. [Google Scholar] [CrossRef]
  31. Battaglia, P.W.; Hamrick, J.B.; Bapst, V.; Sanchez-Gonzalez, A.; Zambaldi, V.; Malinowski, M.; Tacchetti, A.; Raposo, D.; Santoro, A.; Faulkner, R.; et al. Relational inductive biases, deep learning, and graph networks. arXiv 2018, arXiv:1806.01261. [Google Scholar]
  32. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. Physiobank, physiotoolkit, and physionet. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed]
  33. EEG Motor Movement/Imagery Dataset. Available online: https://physionet.org/content/eegmmidb/1.0.0/ (accessed on 9 September 2009).
  34. Wang, M.; El-Fiqi, H.; Hu, J.; Abbass, H.A. Convolutional neural networks using dynamic functional connectivity for EEG-based person identification in diverse human states. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3259–3272. [Google Scholar] [CrossRef]
  35. Wang, M.; Hu, J.; Abbass, H.A. BrainPrint: EEG biometric identification based on analyzing brain connectivity graphs. Pattern Recognit. 2020, 105, 107381. [Google Scholar] [CrossRef]
  36. Schalk, G.; McFarland, D.J.; Hinterberger, T.; Birbaumer, N.; Wolpaw, J.R. BCI2000: A general-purpose brain-computer interface (BCI) system. IEEE Trans. Biomed. Eng. 2004, 51, 1034–1043. [Google Scholar] [CrossRef]
  37. Winkler, I.; Brandl, S.; Horn, F.; Waldburger, E.; Allefeld, C.; Tangermann, M. Robust artifactual independent component classification for BCI practitioners. J. Neural Eng. 2014, 11, 035013. [Google Scholar] [CrossRef]
  38. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Xiao, J.; Hu, F.; Shao, Q.; Li, S. A low-complexity compressed sensing reconstruction method for heart signal biometric recognition. Sensors 2019, 19, 5330. [Google Scholar] [CrossRef] [PubMed]
  40. El-Fiqi, H.; Wang, M.; Salimi, N.; Kasmarik, K.; Barlow, M.; Abbass, H. Convolution neural networks for person identification and verification using steady state visual evoked potential. In Proceedings of the 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 1062–1069. [Google Scholar]
  41. Moini, J.; Piran, P. Functional and Clinical Neuroanatomy: A Guide for Health Care Professionals; Academic Press: Warsaw, Poland, 2020. [Google Scholar]
  42. Lashgari, E.; Liang, D.; Maoz, U. Data augmentation for deep-learning-based electroencephalography. J. Neurosci. Methods 2020, 346, 108885. [Google Scholar] [CrossRef]
  43. Wilcoxon, F. Individual comparisons by ranking methods. Biometrics 1945, 1, 80–83. [Google Scholar] [CrossRef]
  44. Jain, U.; Butchy, A.A.; Leasure, M.T.; Covalesky, V.A.; McCormick, D.; Mintz, G.S. Redundancy and Novelty Between ECG Leads Based on Linear Correlation. In Proceedings of the 16th International Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC 2023), Lisbon, Portugal, 16–18 February 2023; pp. 359–365. [Google Scholar]
  45. Tankisi, H.; Burke, D.; Cui, L.; De Carvalho, M.; Kuwabara, S.; Nandedkar, S.D.; Rutkove, S.; Stålberg, E.; van Putten, M.J.; Fuglsang-Frederiksen, A. Standards of instrumentation of EMG. Clin. Neurophysiol. 2020, 131, 243–258. [Google Scholar] [CrossRef]
  46. Wang, M.; Kasmarik, K.; Bezerianos, A.; Tan, K.C.; Abbass, H. On the channel density of EEG signals for reliable biometric recognition. Pattern Recognit. Lett. 2021, 147, 134–141. [Google Scholar] [CrossRef]
  47. Abdelfattah, S.M.; Abdelrahman, G.M.; Wang, M. Augmenting the size of EEG datasets using generative adversarial networks. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–6. [Google Scholar]
Figure 1. WGLAE model architecture, with the gate control module highlighted.
Figure 1. WGLAE model architecture, with the gate control module highlighted.
Sensors 25 03389 g001
Figure 2. Head map of EEG scalp electrode placements [25].
Figure 2. Head map of EEG scalp electrode placements [25].
Sensors 25 03389 g002
Figure 3. Illustration of gate control sequences generated by the four gate control mechanisms on the 19-channel dataset.
Figure 3. Illustration of gate control sequences generated by the four gate control mechanisms on the 19-channel dataset.
Sensors 25 03389 g003
Figure 4. Reconstruction error distributions over the scalp (resting state). The corresponding gate control mechanisms, from left to right, are hemispheric symmetry rotation, partition-based, order-based, and random.
Figure 4. Reconstruction error distributions over the scalp (resting state). The corresponding gate control mechanisms, from left to right, are hemispheric symmetry rotation, partition-based, order-based, and random.
Sensors 25 03389 g004
Figure 5. The mean reconstruction errors (RMSE) across all samples when reconstructing different numbers of missing channels. The left and right columns show the results from the 19- and 29-channel datasets, respectively. (a) 19 channels—Resting State; (b) 29 channels—Resting State; (c) 19 channels—Motor Movement; (d) 29 channels—Motor Movement; (e) 19 channels—Motor Imagery; (f) 29 channels—Motor Imagery.
Figure 5. The mean reconstruction errors (RMSE) across all samples when reconstructing different numbers of missing channels. The left and right columns show the results from the 19- and 29-channel datasets, respectively. (a) 19 channels—Resting State; (b) 29 channels—Resting State; (c) 19 channels—Motor Movement; (d) 29 channels—Motor Movement; (e) 19 channels—Motor Imagery; (f) 29 channels—Motor Imagery.
Sensors 25 03389 g005
Table 1. Electrode sequences under the three proposed gate control mechanisms.
Table 1. Electrode sequences under the three proposed gate control mechanisms.
Order-basedFP1F7F3T7C3P7P3O1FzCz
PzFp2F4F8C4T8P4P8O2
Partition-basedFp1Fp2F7F3FzF4F8T7C3Cz
C4T8P7P3PzP4P8O1O2
Hemispheric symmetry rotationFp1Fp2F7F8F3F4FzT7T8C3
C4CzP7P8P3P4PzO1O2
Table 2. Classification model configurations.
Table 2. Classification model configurations.
Person Identification Task:
Layers Kernel Size#Kernels/Neurons Ch.MaxPoolStridePaddingBN/AF/DP
Conv1 3 × 3 64[2, 2][1, 1][0, 0]Yes/ReLU/Yes
Conv2 5 × 5 64[2, 2][1, 1][0, 0]Yes/ReLU/Yes
Conv3 5 × 5 64[2, 2][1, 1][0, 0]Yes/ReLU/Yes
FLnul128nulnulnulNo/ReLU/No
FLnul109nulnulnulNo/Softmax/No
BN: Batch normalisation; AF: Activation function; DP: Dropout; FL: Fully connected layer.
Table 3. The average and standard deviation of the reconstruction errors (RMSE) of entire samples when reconstructing 1 to 10 channels out of 19 channels.
Table 3. The average and standard deviation of the reconstruction errors (RMSE) of entire samples when reconstructing 1 to 10 channels out of 19 channels.
Resting State:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.059 ± 0.0030.060 ± 0.0030.060 ± 0.0030.060 ± 0.003
20.064 ± 0.0030.064 ± 0.0040.064 ± 0.0040.064 ± 0.004
30.067 ± 0.0040.068 ± 0.0040.068 ± 0.0050.069 ± 0.005
40.069 ± 0.0040.070 ± 0.0050.070 ± 0.0050.072 ± 0.007
50.074 ± 0.0070.074 ± 0.0050.074 ± 0.0070.075 ± 0.006
60.076 ± 0.0090.077 ± 0.0080.077 ± 0.0090.077 ± 0.007
70.078 ± 0.0090.079 ± 0.0090.079 ± 0.0100.080 ± 0.013
80.078 ± 0.0080.079 ± 0.0100.078 ± 0.0070.084 ± 0.012
90.077 ± 0.0040.079 ± 0.0050.079 ± 0.0050.083 ± 0.010
100.087 ± 0.0180.087 ± 0.0160.088 ± 0.0180.083 ± 0.007
Motor Movement:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.049 ± 0.0020.049 ± 0.0020.050 ± 0.0030.049 ± 0.002
20.053 ± 0.0030.053 ± 0.0030.054 ± 0.0040.053 ± 0.003
30.056 ± 0.0050.056 ± 0.0050.058 ± 0.0060.057 ± 0.004
40.058 ± 0.0050.057 ± 0.0050.059 ± 0.0050.061 ± 0.005
50.062 ± 0.0060.061 ± 0.0060.064 ± 0.0080.061 ± 0.006
60.063 ± 0.0070.061 ± 0.0060.064 ± 0.0090.063 ± 0.005
70.061 ± 0.0040.062 ± 0.0050.063 ± 0.0040.064 ± 0.005
80.067 ± 0.0060.067 ± 0.0070.069 ± 0.0080.072 ± 0.012
90.071 ± 0.0110.069 ± 0.0080.073 ± 0.0120.070 ± 0.010
100.071 ± 0.0170.068 ± 0.0110.071 ± 0.0150.069 ± 0.008
Motor Imagery:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.052 ± 0.0020.052 ± 0.0020.053 ± 0.0020.053 ± 0.002
20.056 ± 0.0040.057 ± 0.0040.058 ± 0.0040.057 ± 0.003
30.060 ± 0.0050.060 ± 0.0050.061 ± 0.0060.062 ± 0.004
40.061 ± 0.0060.061 ± 0.0060.063 ± 0.0070.067 ± 0.010
50.065 ± 0.0050.066 ± 0.0060.068 ± 0.0080.066 ± 0.007
60.064 ± 0.0030.065 ± 0.0030.066 ± 0.0050.066 ± 0.005
70.066 ± 0.0040.066 ± 0.0040.067 ± 0.0040.077 ± 0.013
80.070 ± 0.0080.071 ± 0.0090.072 ± 0.0100.072 ± 0.009
90.075 ± 0.0120.075 ± 0.0120.077 ± 0.0130.075 ± 0.015
100.073 ± 0.0220.074 ± 0.0220.077 ± 0.0280.068 ± 0.007
The best performance among the four gating mechanisms is highlighted in bold.
Table 4. Spectral fidelity analysis results (5 out of 19 missing channels).
Table 4. Spectral fidelity analysis results (5 out of 19 missing channels).
Reconstructed Resting State Data
Frequency bandSpectral correlation (Hemispheric vs. Random)nRMSE (Hemispheric vs. Random)
Delta (1–4 Hz)0.83 ± 0.04 (H) vs. 0.79 ± 0.05 (R) *0.11 ± 0.02 (H) vs. 0.13 ± 0.03 (R) *
Theta (4–8 Hz)0.86 ± 0.03 (H) vs. 0.81 ± 0.04 (R) *0.10 ± 0.02 (H) vs. 0.12 ± 0.03 (R) *
Alpha (8–13 Hz)0.90 ± 0.04 (H) vs. 0.85 ± 0.05 (R) *0.06 ± 0.01 (H) vs. 0.09 ± 0.02 (R) *
Beta (13–30 Hz)0.82 ± 0.05 (H) vs. 0.78 ± 0.06 (R) *0.12 ± 0.02 (H) vs. 0.14 ± 0.03 (R) *
Gamma (30–50 Hz)0.75 ± 0.06 (H) vs. 0.72 ± 0.07 (R) *0.15 ± 0.03 (H) vs. 0.17 ± 0.04 (R) *
Reconstructed Motor Movement Data
Frequency bandSpectral correlation (Partition vs. Random)nRMSE (Partition vs. Random)
Delta (1–4 Hz)0.83 ± 0.05 (P) vs. 0.80 (R) ± 0.05 *0.12 ± 0.02 (P) vs. 0.14 ± 0.03 (R) *
Theta (4–8 Hz)0.84 ± 0.04 (P) vs. 0.81 (R) ± 0.06 *0.09 ± 0.02 (P) vs. 0.11 ± 0.03 (R) *
Alpha (8–13 Hz)0.93 ± 0.03 (P) vs. 0.86 (R) ± 0.05 *0.07 ± 0.01 (P) vs. 0.10 ± 0.02 (R) *
Beta (13–30 Hz)0.87 ± 0.04 (P) vs. 0.82 (R) ± 0.06 *0.08 ± 0.01 (P) vs. 0.11 ± 0.03 (R) *
Gamma (30–50 Hz)0.77 ± 0.06 (P) vs. 0.71 (R) ± 0.05 *0.13 ± 0.02 (P) vs. 0.15 ± 0.03 (R) *
* p < 0.05.
Table 5. The average and standard deviation of the reconstruction errors (RMSE) across all samples when reconstructing 1 to 10 channels out of 29.
Table 5. The average and standard deviation of the reconstruction errors (RMSE) across all samples when reconstructing 1 to 10 channels out of 29.
Resting State:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.058 ± 0.0030.058 ± 0.0030.057 ± 0.0030.058 ± 0.003
20.061 ± 0.0030.061 ± 0.0030.061 ± 0.0030.063 ± 0.004
30.064 ± 0.0030.064 ± 0.0030.064 ± 0.0040.067 ± 0.004
40.068 ± 0.0040.069 ± 0.0040.068 ± 0.0040.071 ± 0.005
50.072 ± 0.0040.073 ± 0.0040.072 ± 0.0050.073 ± 0.007
60.076 ± 0.0050.077 ± 0.0060.076 ± 0.0070.074 ± 0.005
70.075 ± 0.0040.075 ± 0.0040.075 ± 0.0050.078 ± 0.007
80.078 ± 0.0040.079 ± 0.0050.078 ± 0.0050.082 ± 0.008
90.082 ± 0.0080.082 ± 0.0090.083 ± 0.0110.084 ± 0.009
100.084 ± 0.0090.085 ± 0.0090.083 ± 0.0080.083 ± 0.009
Motor Movement:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.048 ± 0.0020.048 ± 0.0020.047 ± 0.0020.048 ± 0.002
20.05 ± 0.0020.05 ± 0.0020.05 ± 0.0020.051 ± 0.002
30.053 ± 0.0020.052 ± 0.0020.053 ± 0.0010.054 ± 0.003
40.055 ± 0.0030.055 ± 0.0030.056 ± 0.0030.056 ± 0.004
50.057 ± 0.0020.056 ± 0.0020.057 ± 0.0020.06 ± 0.005
60.059 ± 0.0030.058 ± 0.0030.059 ± 0.0030.064 ± 0.005
70.061 ± 0.0030.061 ± 0.0030.063 ± 0.0040.069 ± 0.005
80.064 ± 0.0040.063 ± 0.0030.065 ± 0.0040.067 ± 0.007
90.064 ± 0.0030.064 ± 0.0030.067 ± 0.0040.067 ± 0.005
100.068 ± 0.0060.068 ± 0.0070.07 ± 0.0080.07 ± 0.005
Motor Imagery:
Missing channelsHemispheric symmetry rotationPartition-basedOrder-basedRandom
10.05 ± 0.0020.051 ± 0.0020.05 ± 0.0020.051 ± 0.002
20.054 ± 0.0020.054 ± 0.0020.054 ± 0.0020.056 ± 0.003
30.057 ± 0.0020.056 ± 0.0020.057 ± 0.0020.059 ± 0.003
40.06 ± 0.0030.059 ± 0.0020.061 ± 0.0030.062 ± 0.004
50.062 ± 0.0020.062 ± 0.0020.064 ± 0.0020.067 ± 0.009
60.064 ± 0.0030.063 ± 0.0030.065 ± 0.0030.07 ± 0.005
70.07 ± 0.0090.068 ± 0.0070.072 ± 0.0090.068 ± 0.005
80.071 ± 0.0040.069 ± 0.0040.074 ± 0.0060.072 ± 0.005
90.069 ± 0.0040.068 ± 0.0030.071 ± 0.0040.078 ± 0.008
100.074 ± 0.0110.072 ± 0.0090.075 ± 0.0110.079 ± 0.018
The best performance among the four gating mechanisms is highlighted in bold.
Table 6. Wilcoxon signed-rank test between the random sequence and hemispheric symmetry rotation gate control mechanisms, and between the partition-based and order-based gate control mechanisms.
Table 6. Wilcoxon signed-rank test between the random sequence and hemispheric symmetry rotation gate control mechanisms, and between the partition-based and order-based gate control mechanisms.
19-Channel29-Channel
EOMIMMEOMIMM
Hemispheric symmetry rotation vs. Random0.0010.0020.0080.0050.0000.000
Partition-based vs. Random0.0260.0070.0060.0470.0000.000
Order-based vs. Random0.0260.2280.2780.0030.0030.000
Table 7. Subject identification results with 0, 1, 5, and 10 reconstructed channels out of 19 and 29 total channels.
Table 7. Subject identification results with 0, 1, 5, and 10 reconstructed channels out of 19 and 29 total channels.
Motor Movement:
#Missing channels/#Total channelsGate control mechanisms
Hemispheric symmetry rotationPartition-basedOrder-basedRandom
0/1987.77
1/1986.0686.3585.4685.04
5/1985.5385.9285.1784.19
10/1981.9682.3481.0380.72
0/2992.97
1/2992.0492.1791.3590.85
5/2989.7390.2188.9289.81
10/2988.2188.3587.7786.48
Motor Imagery:
#Missing channels/#Total channelsGate control mechanisms
Hemispheric symmetry rotationPartition-basedOrder-basedRandom
0/1987.16
1/1986.3386.4885.6885.43
5/1984.6585.0784.884.07
10/1980.5781.1879.8880.09
0/2995.72
1/2994.8695.0892.0892.41
5/2990.9192.6489.1589.83
10/2991.5991.2990.2285.78
The best performance among the four gating mechanisms is highlighted in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, K.; El-Fiqi, H.; Wang, M. Gate Control Mechanisms of Autoencoders for EEG Signal Reconstruction. Sensors 2025, 25, 3389. https://doi.org/10.3390/s25113389

AMA Style

Li K, El-Fiqi H, Wang M. Gate Control Mechanisms of Autoencoders for EEG Signal Reconstruction. Sensors. 2025; 25(11):3389. https://doi.org/10.3390/s25113389

Chicago/Turabian Style

Li, Kangjing, Heba El-Fiqi, and Min Wang. 2025. "Gate Control Mechanisms of Autoencoders for EEG Signal Reconstruction" Sensors 25, no. 11: 3389. https://doi.org/10.3390/s25113389

APA Style

Li, K., El-Fiqi, H., & Wang, M. (2025). Gate Control Mechanisms of Autoencoders for EEG Signal Reconstruction. Sensors, 25(11), 3389. https://doi.org/10.3390/s25113389

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop