Next Article in Journal
A Novel Prototype for Safe Driving Using Embedded Smart Box System
Next Article in Special Issue
Light Field Reconstruction Using Residual Networks on Raw Images
Previous Article in Journal
Motion Blur Kernel Rendering Using an Inertial Sensor: Interpreting the Mechanism of a Thermal Detector
Previous Article in Special Issue
Multi-Excitation Infrared Fusion for Impact Evaluation of Aluminium-BFRP/GFRP Hybrid Composites
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Post Hoc Explainability of Optimized Self-Organizing Reservoir Network for Action Recognition

1
Faculty of Engineering and Technology, Multimedia University, Jalan Ayer Keroh Lama, Melaka 75450, Malaysia
2
Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, Universiti Malaya, Kuala Lumpur 50603, Malaysia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 1905; https://doi.org/10.3390/s22051905
Submission received: 19 January 2022 / Revised: 18 February 2022 / Accepted: 19 February 2022 / Published: 1 March 2022
(This article belongs to the Special Issue Computational Methods in Imagery (CMI))

Abstract

:
This work proposes a novel unsupervised self-organizing network, called the Self-Organizing Convolutional Echo State Network (SO-ConvESN), for learning node centroids and interconnectivity maps compatible with the deterministic initialization of Echo State Network (ESN) input and reservoir weights, in the context of human action recognition (HAR). To ensure stability and echo state property in the reservoir, Recurrent Plots (RPs) and Recurrence Quantification Analysis (RQA) techniques are exploited for explainability and characterization of the reservoir dynamics and hence tuning ESN hyperparameters. The optimized self-organizing reservoirs are cascaded with a Convolutional Neural Network (CNN) to ensure that the activation of internal echo state representations (ESRs) echoes similar topological qualities and temporal features of the input time-series, and the CNN efficiently learns the dynamics and multiscale temporal features from the ESRs for action recognition. The hyperparameter optimization (HPO) algorithms are additionally adopted to optimize the CNN stage in SO-ConvESN. Experimental results on the HAR problem using several publicly available 3D-skeleton-based action datasets demonstrate the showcasing of the RPs and RQA technique in examining the explainability of reservoir dynamics for designing stable self-organizing reservoirs and the usefulness of implementing HPOs in SO-ConvESN for the HAR task. The proposed SO-ConvESN exhibits competitive recognition accuracy.

1. Introduction

Human action recognition (HAR) has been an active research field to interpret human intentions. HAR studies aim to develop real-world and reliable applications to perceive, study and identify human actions in videos [1]. The advancement in depth data acquisition hardware technologies such as the ASUS Xtion Pro or the Microsoft Kinect has led to the rise of noticeable HAR research outcomes in both commercial products and studies. Some prominent HAR applications include visual surveillance [2,3], human computer interaction [4,5], physical rehabilitation [6], and autonomous driving vehicles [7].
The simplicity and accuracy of extracting three-dimensional (3D)-skeleton-joints from depth images [8] have driven our work to focus on 3D-skeleton-joints-based HAR. The research works on 3D-skeleton-joints-based HAR still remains demanding. Current studies consider skeleton-joints human actions as multivariate time-series of five body parts (left arm, right arm, left leg, right leg and a central trunk) and attempt to identify and model the dynamical temporal features in 3D space. Echo State Networks (ESNs) [9] are such a popular Reservoir Computing (RC) method which is suitable for learning the temporal context. Its simplicity by randomly initializing and fixing the ESN’s input and reservoir weights during training, as compared with backpropagation through time of Recurrent Neural Networks (RNNs), diminishes the computational complexity. Furthermore, short-term memory property which ensures history information of the reservoir is not broadcasted to other neurons too rapidly, makes ESN suitable to capture the dynamical temporal features in HAR. Recently, Ma et al. [10] proposed an integrated architecture for HAR tasks, the Convolutional Echo State Network (ConvESN), by combining the RC with convolutional deep learning. Owing to ESN multiscale memory, echo state representations (ESRs) of input action series produced by ConvESN contains the history information which make it suitable to characterize and capture temporal dynamics in 3D skeleton series. Additionally, in ConvESN, the Convolutional Neural Network (CNN) [11] substitutes the linear regression in ESN output layer to understand the complex action echo states. In spite of ESN-based approaches achieved encouraging recognition performance, in this work we have identified three research avenues for our investigation: (1) Despite the random initialization of the ESN’s input and reservoir weights may reduce computational cost, on the other hand, this may rise instability and variance in generalization and hence diminish reproducibility [12]. In the context of HAR, a stable and reproducible multiscale feature extraction mechanism is needed to guarantee the performance of action recognition, in particular, when 3D skeleton joints are considered as multivariate time-series. Randomly fixed neuron weights may diversify the recognition performance of ESN-based approaches even in performing the same task with the identical set of hyperparameter configurations [13]. It hardly reproduces the same performance due to the randomized input and reservoir weights in different repeated runs. (2) Building an ESN model requires a set of hyperparameters to be configured. Hyperparameters are external to the model and are commonly tuned based on rule of thumb or empirically fixed via trial-and-error by researchers’ past experiences. Moreover, ESN remains as a black box algorithm. Particularly, it lacks of explainability consideration to understand the input-dependent reservoir dynamics for HAR. Using the explanatory information about the knowledge learned by ESN in tuning the hyperparameters seems to be promising. (3) Following the body of work of ConvESN, it has incorporated modelling dynamics and multiscale temporal feature in a unified framework. However, the model may be very sensitive to the selection of hyperparameters in CNN. Notably, ConvESN implemented Adam [14] optimizer in which the learning rate monotonically decreases based on the training iteration index. Careful selection of initial learning rate is required to alleviate training traps in local minima. Moreover, finding the optimal set of CNN hyperparameters can become computational costly [15], particularly, ConvESN implemented manual hyperparameter tuning for CNN stage.
To address these problems, in this work we propose a novel reservoir design approach known as the Self-Organizing Reservoir Network with Explainability (SORN-E) which is characterised by (i) the integration of Adaptive Resonance Theory (ART) [16] architecture and topology construction based on Instantaneous Topological Mapping (ITM) [17] for the self-organization of the input weights and reservoir weights, and ii) hyperparameter tuning based on the explainability of self-organizing reservoir through Recurrent Plots (RPs) and Recurrence Quantification Analysis (RQA) technique [18]. Input-driven and self-organization have been proven to be crucial for the cortex to adapt the neurons in accordance with input topology or distribution [19]. This previous effort motivates our work to take unsupervised self-organizing learning into account as a potential biologically plausible approach for self-organizing reservoir design. Combining the advantages of ART and ITM, SORN-E has similar network architecture to handle plasticity and stability dilemma and the number of network nodes is not required to be defined prior the learning. The unsupervised learning process of SORN-E is composed of best-matching node selection, vigilance test, and node learning. SORN-E encodes human actions as multivariate time-series signals. It performs unsupervised learning from training dataset to generates a self-organizing clustered topology of nodes which maintains the topological properties of the input at a greatly reduced dimensionality. The generated maps are represented by clustered node centroids and interconnectivity maps. Input samples with sufficiently high similarity are often characterized by a single node or a cluster of nodes.
Furthermore, to guarantee stability and echo state property (ESP) [9], we exploit a state-of-the-art qualitative stability criterion for reservoirs of ESN called maximum diagonal line length, L M A X [18]. This RQA qualitative metric is used to quantify the reservoir stability. By measuring this metric at different configurations of input scaling ( I S ) and spectral radius ( S R ) of the reservoir, the optimal values for stability can be identified. With the aim to provide explanatory information of the dynamics and improve insights of SORN-E, RPs and RQA are used as descriptive approach to examine the dynamics of self-organizing reservoir and to gain a clear idea of the echo state representations (ESRs).
Referring to the body of work of ConvESN, the proposed SORN-E is cascaded with a simple CNN. The feature maps generated by SORN-E are applied to initialize the input weights and recurrent hidden weights in the ESN to yield optimized self-organizing reservoirs. CNN then learns the multiscale temporal features from ESRs for action recognition. This resultant novel implementation is named Self-Organizing Convolutional Echo State Network (SO-ConvESN). With respect to the learning of the CNN stage of the proposed approach, sequential, parallel, sequential-parallel hyperparameter optimization (HPO) algorithms are investigated with the intention of obtaining optimal HAR performance. SO-ConvESN has also been deployed to HAR tasks to demonstrate the feasibility and applicability.
In a nutshell, our main contributions can be summarized as follows:
  • We propose an unsupervised self-organizing network for learning node centroids and interconnectivity maps which are compatible for the deterministic initialization of ESN reservoir weights. To ensure stability and ESP in a self-organizing reservoir, we further exploit the RQA technique for explainability and characterization of the dynamics of self-organizing reservoir and hence tuning two critical ESN hyperparameters: input scaling ( I S ) and spectral radius ( S R ).
  • Cascading the stable and optimized self-organizing reservoirs with a simple CNN. Self-organizing reservoir ensures that the activation of ESN internal ESRs echoes similar topological qualities and temporal features of the input time-series and CNN efficiently learns the dynamics and multi-scale temporal features from the ESRs for action recognition.
  • Adopting three different categories of HPO algorithms, namely Sequential method: Bayesian Optimization (BO) [20], Parallel method: Asynchronous Successive Halving Algorithm (ASHA) [21], and Parallel-Sequential method: Population-based Training (PBT) [22] to search for optimal hypermeters of CNN stage in SO-ConvESN for HAR tasks.
  • Conducting experiments by using several publicly available 3D-skeleton-based action recognition datasets to examine the explainability of self-organizing reservoirs dynamics, investigate the recognition accuracy of SO-ConvESN and the feasibility of implementing HPOs in SO-ConvESN for the HAR task.
The rest of the paper is organized as follows: Reviews of the related works on ESN-based approaches for 3D-skeleton-based HAR and self-organizing approaches for clustering are first presented followed by the concise descriptions of explainability methods for ESN and HPO for CNN. Next, we describe the development framework of SO-ConvESN and the details of each stage: SORN-E, explainability methods, and HPO. The simulation experiments and results are then presented based on several publicly available benchmarking datasets. Last of all, concluding remarks are presented.

2. Related Works

Before proceeding further, this section first reviews the existing methods for HAR tasks, particularly the ESN-based approaches from which we identify suitable for 3D-skeleton-based HAR, and highlights the method’s limitations. We note that self-organization potentially fits reservoir design; we next highlight the overview of unsupervised self-organizing learning approaches. Finally, we provide a brief discussion of techniques for explainability of reservoir design and hyperparameter optimization of CNN.

2.1. ESN-Based Approaches for HAR Tasks

Current studies on 3D-skeleton-joints-based HAR making use of machine learning techniques, such as Support Vector Machines [23,24,25], Multilayer Perceptrons [25], Dynamic Time Warping [26,27,28], Hidden Markov Models [27], and Decision Trees [24,25] irregularly neglect the temporal features of human actions in 3D space. Temporal features could be containing helpful information over the intervals between activities. This problem drives the study of the HAR approaches to adequately modeling the temporal features in human actions data. Adopting memory mechanisms like the implementation in RNNs [29] is practical in handling temporal data for HAR. However, due to the gradient-based training of RNN architectures with numerous layers, it is well known to be experiencing the consequences of exploding or vanishing gradients. Long Short-Term Memory (LSTM) [30] networks address this error-prone gradient-based training by coupling gating mechanisms into RNN architecture and successfully enhancing the learning process of the temporal features for 3D-skeleton-joints-based HAR.
On the other hand, RC emerges as an alternative approach for conventional RNN training. ESNs [9] are a widespread special RC implementation for training RNN. The main characteristics that make ESNs distinguished are the short-term memory property contributed by the sparsely connected reservoir and reduced computational complexity compared to backpropagation through time yielded by the random initialization of the input and reservoir weights. These make ESNs a suitable method that fits for handling temporal context. Nevertheless, to the best knowledge of the literature review, the implementation of the ESN-based approach in HAR using 3D-skeleton joints has not been fully explored. Significatory strategies include bidirectional Leaky Integrator ESNs (LI-ESNs) [31], canonical ESN by integrating 3D body joints and objects category [13], and ConvESN [10]. Bidirectional LI-ESNs [31] perform direct processing on temporal 3D body joints without extra feature extraction procedures during the encoding process. Each input sequence must be completely available and go through the bidirectional reservoir to produce better state representation. Like canonical training-free ESN, LI-ESNs apply random initialization of input weights but describe a permutation matrix for the recurrent weights. On the other hand, Mici et al. [13] integrated additional contextual ideas with 3D body joints to enhance the performance of the HAR. This approach considers the participation between the performed action and the manipulated object to tackle ambiguities during the activity. Hence, the learning of ESN needs to combine the 3D body joints and the object labels. Object labels may impact the internal reservoir representation. Mici et al. [13] demonstrated the proposed approach to handle a multi-label classification task considering the amount and variety of manipulated objects that change in different actions. Mici et al. [13] also takes the advantage of randomly initialized and fixed input weight and recurrent weight in the reservoir.
The approaches mentioned above focused on the manipulation of mapping input data into the reservoir. However, linear regression, as in canonical ESN, may limit the decoding capability in the output layer [32]. Moreover, extracting dynamical, locally related features in 3D skeleton sequences is crucial to improve HAR performance. Existing feature extraction approaches in HAR by a heuristic handcrafted mechanism could be insufficient [33,34]. It greatly depends on researchers’ experience and domain knowledge. Ma et al. [10] introduced ConvESN as another ESN-based approach for 3D-skeleton-joints-based HAR to address these problems. ConvESN significantly impacted our works positively due to its superior structures in extracting multiscale temporal features from human actions data. In recent years, HAR research has applied deep learning, particularly CNN, for superior and automated high-level feature extraction [35]. The ConvESN is such an approach that bridged CNN and RC areas for the HAR task. It substitutes the ESN readout layer with a CNN that consists of a convolution layer followed by a max-pooling layer. By treating 3D-skeleton-joints input as multivariate time-series and projecting the series onto reservoir to generate ESRs, CNN effectively decodes the ESRs at all time steps and extracts dynamics and multiscale features. Ma et al. [10] demonstrated the effectiveness of ConvESN for HAR. In ConvESN, we can view the significance of CNN as complementary to the ESN stage in HAR. CNN decodes the multiscale features, which may be lacking by the ESN stage.
The ultimate goal of implementing ESN in the HAR approaches, as mentioned earlier, is to reduce the computational cost. However, the random initialization of the ESN’s reservoir may generate uncertainty and diversity in generalization. This variance may weaken reproducibility [12]. In a nutshell, repeating the same task with the identical set of hyperparameter configurations hardly reproduces the same recognition performance due to the randomly initialized and fixed neuron weights [13]. Palangi et al. [36] have shown that adapting one or both ESN’s input and recurrent weights to the input data offers more reliable classification performance. In biological analogy, self-organization and input-driven learning mechanism are also critical for the cortex to adapt the weights of the neurons to the input properties [19]. These notions show that adapting weights to the input data seems to be desirable. Besides, most of the existing ESN-based HAR methods also lack consideration of explainability in expressing the input-dependent reservoir dynamics. ESN often stays as a black-box algorithm. Expressing and understanding the explanatory information about the knowledge learned by ESN, especially reservoir echo state, could be helpful to set ESN’s hyperparameters configuration [37]. Focusing on the ConvESN, careful selection of hyperparameters in CNN is essential. During CNN weights learning, poor selection in the initial learning rate of the optimizer may leave the learning process stuck at local minima or saddle points. Currently, ConvESN implements manual hyperparameter tuning for the CNN stage. Finding the optimal set of CNN hyperparameters can be computationally costly [15].
With the issues mentioned above, we hypothesize that implementing a deterministic and self-organizing reservoir in ESN may ensure the activation of internal ESRs echoes similar topological qualities and temporal features of the input time-series. The explainability and characterization of the dynamics of the reservoir could also be helpful in tuning the ESN hyperparameters. Lastly, applying the HPO algorithms during CNN training could improve the HAR performance to a certain extent. We next extend the discussion to include the overview of unsupervised self-organizing learning approaches and a brief introduction of techniques for explainability of reservoir design and hyperparameter optimization of CNN.

2.2. Unsupervised Self-Organizing Learning Approaches

More recently, Kohonen[38] introduced an approach to initialize the canonical ESN’s reservoir based on unsupervised learning strategies. In essence, this approach applied Self-Organizing Map (SOM) [39] and Neural Gas Network (NG) [40,41] for the adaption of the reservoir input weights. The reservoir weight matrix was also learned from NG. However, the robustness of this approach was influenced by the dimensional of the input space. This limitation could be due to SOM and NG being characterized by a predefined and fixed-size topological map that cannot be changed over time. Growing Neural Gas (GNG) [42] has been introduced as another important algorithm of topological self-organizing clustering to address this limitation. GNG has extended the adaptive abilities to add new neurons, create and delete connections between neurons. SOM, NG, and GNG heavily assume that the input data are statistically independent. In particular, the GNG network grows at a fixed rate regardless of the variation of the input distribution. The limitation makes them challenging in manipulating trajectory data, such as 3D-skeleton joints. ITM [17] represents one approach to overcome this constraint. It handles input data with strong correlation by adapting a formation of a topological map. On the other hand, SOM, NG, and GNG algorithms also experience from plasticity-stability dilemma [43]. ART [16] is a well-known self-organizing approach that is capable to address this dilemma. The ART network has similar architecture as the learning process in the human cortex. Proposing a novel self-organizing reservoir design based on the integration of ART and ITM for the context of HAR seems to be motivating.

2.3. Explainability Methods for ESN

When developing an ESN for particular applications, respectively, for the HAR task, it is crucial knowing if the generated reservoir is optimal for the problem at hand. The area of explainability developed for ESN has been studied but at a knowingly smaller collection of published works. When developing an ESN for the particular applications, respectively, for the HAR task, it is crucial knowing if the generated reservoir is optimal for the problem at hand. Interpreting the reservoir’s explanatory information could improve and guide the development of ESN [37]. Bianchi et al. [18] address this issue by applying RPs [44,45,46] and heuristic complexity measures known as RQA to analyze the dynamics of the neuron activations in a reservoir. This technique is used as analysis tool to visualize and characterize the reservoir dynamics based on the echo state matrix. Alternatively, under the Explainable Artificial Intelligence (xAI) paradigm, Arrieta et al. [37] proposed a set of novel post hoc xAI techniques to discover the strengths and weaknesses for ESN-based models. It focuses on the explainability of Deep Echo State Networks (Deep ESNs) [47]. Comparing both mentioned techniques, RPs and RQA-based techniques generally seem to be well-suited for guiding self-organizing reservoir design. The reasons are twofold: RQA measure provides the stability criteria that are crucial to designing a stable reservoir. Moreover, RPs can be used to visualize the reservoir dynamics for better explainability. This work utilizes RPs and RQA-based techniques to extract the input-driven explanatory information captured over time for understanding self-organizing reservoirs and assess the reservoir stability properties for tuning the ESN hyperparameters. We include the comprehensive discussion of the RPs and RQA-based procedures to guide the design of the self-organizing reservoir in the later section. Applying explainability technique can provide transparency to understand and express the reservoir dynamics and use it in tuning the ESN’s hyperparameters configuration.

2.4. HPO Algorithms for CNN Training

As mentioned earlier, our work integrates SORN-E with CNN to yield SO-ConvESN to address the HAR task. Similar to most machine learning models, CNN possesses a set of external network configurations known as hyperparameters [48] that influence the learning process. Hyperparameters cannot be estimated or learned from the training data and must be carefully tuned to unlock the network performance fully. However, selecting the suitable configurations of the hyperparameters could be challenging. Some existing CNN models apply default implementation, which could not be optimal for the problem on hand. Besides, manually finding the optimal set of CNN hyperparameters, such as ConvESN, is common. The manual search process may require thorough hyperparameter tuning experiments that could be computationally costly [15]. In the worst case, the network training suffers from the sensitivity to the poor selection of some hyperparameters, such as the initial learning rate, which may influence the performance to be inadequate. Therefore, applying HPO may help reduce human efforts to search for the ideal CNN hyperparameters in an automated and structured manner.
Hyperparameters of CNN are often viewed as the variable of the network structure in which finding the optimal configurations leads to an optimization problem [15]. There is a wide variety of works on algorithms for HPO, which are often problem-dependent, and there is no single HPO method that suits all models [49,50,51,52]. We limit our focus on the well-established or/and state-of-the-art algorithms well-suited for CNN hyperparameter tuning. In general, two main categories of hyperparameter tuning algorithms exist: sequential methods and parallel methods [22]. These methods are distinguished by the use of computational resources and time elapsed to reach optimal configuration.
Sequential methods run several training processes consecutively. It progressively performs hyperparameter tuning. Each run uses information from the earlier training process to select new hyperparameters and retrain the model in the subsequent new runs. Typically, sequential methods employ minimum computational resources but require high hyperparameter tuning time. BO [20] is a sequential method that adaptively selects configurations by identifying proper search space to discover candidate hyperparameters to evaluate next. The inherently sequential optimization of BO makes this method commonly used for the small-scale regime. As an indicative algorithm in sequential methods, exploring and investigating the feasibility of BO in optimizing the hyperparameters of CNN in SO-ConvESN could be worth trying.
On the other side of the coin, parallel methods initialize network weights of multiple training processes and run with different hyperparameters in parallel. Parallel methods aim to obtain one single best-performing configuration from the multiple HPO processes. Generally, hyperparameter tuning time of parallel methods consume only one training process but employ more computational resources. Grid search and random search are representations of parallel HPO methods [49]. One of the state-of-the-art parallel methods is known as ASHA [21]. ASHA proposed a parallelization scheme by combining adaptive configuration selection as well as evaluation. It primarily addressed the issues of parallelizing Successive Halving Algorithm (SHA) in which the optimization time scales linearly with the number of workers [53]. ASHA asynchronously allocates more computational resources to the promising configuration and performs early stopping on underperformed training processes whenever possible to minimize stragglers. The parallelism of ASHA makes this method suitable for massive parallel optimization processes that require aggressive early stopping. Implementing ASHA could minimize the common overfitting problem in neural networks and hence suits for optimizing the hyperparameters of CNN in SO-ConvESN.
Recently, Population-Based Training (PBT) [22] has been introduced as a state-of-the-art HPO method that links and enhances parallel methods and sequential methods. PBT is similar to parallel methods [49]. It begins training by randomly selecting hyperparameters and initializing network weights. Additionally, it maintains a population of asynchronous training processes, periodically evaluated based on the chosen hyperparameters. In particular, it implements a novel explore-exploit paradigm to allocate the computational resources to the training process, which has the highest possibility to produce good performance meanwhile adaptively tune the hyperparameters during training. In PBT, an underperformed population member exploits other members of the population and replaces itself with a member with better performance. It also explores new hyperparameters to modify the hyperparameters of the better performing member. PBT is designed initially for neural networks. The jointly tuning of weight parameters and hyperparameters make PBT suited for the HPO of CNN.
Based on the potentials of each of the abovementioned techniques, it is interesting to implement these HPO algorithms and investigate the analysis results in optimizing the CNN to discover the most suitable method. Notably, in this work, applying these HPO algorithms during CNN training of SO-ConvESN, to a certain extent, could improve HAR performance.

3. Materials and Methods

This section first presents the overview of the development framework of SO-ConvESN for HAR. The paradigm of the proposed SORN-E approach and the approach of cascading SORN-E with a CNN to yield SO-ConvESN is then thoroughly introduced.

3.1. Development Framework of the Proposed SO-ConvESN

This work aims to propose a novel ESN-based network known as SO-ConvESN, tackling HAR task based on 3D-skeleton single-view and single-person-based scenario. Two key components: SORN-E stage and SO-ConvESN stage, play crucial roles in the framework. Figure 1 depicts the overview of the workflow used to develop the proposed SO-ConvESN.
During data preparation, each skeleton series from training set is separated into five individual channels correspond to 3D-skeleton joint coordinate trajectories of five body parts; left arm (LA), right arm (RA), central trunk (CT), left leg (LL) and right leg (RL). Considering that the movement of head coordinates are synchronized with the movement of center of shoulders, this work discarded the head coordinates. Each of the body parts of the 3D skeleton joint was treated as multivariate time-series. They were used to train and construct the corresponding reservoirs. Self-organizing clustering was performed by SORN separately for each channel and the corresponding reservoir was created to obtain the action echo states. RQA is then implemented on the echo states to tune the hyperparameters of self-organizing reservoir for optimal stability. This process will be iterated until the satisfied stability is achieved.
Following the direction of ConvESN, the developed SORN-E is afterward cascaded with a simple CNN to formulate the SO-ConvESN. Optimized self-organizing reservoirs generates the ESRs from the corresponding channels. CNN then extracts the multiscale temporal features from these ESRs. Hyperparameter optimization algorithms are applied for searching the optimal hyperparameters of the CNN. Optimized SO-ConvESN is then deployed for HAR applications.

3.2. Self-Organizing Reservoir Network with Explainability (SORN-E)

This section introduces the SORN-E for self-organizing reservoir design via the exploitation of explainability techniques. The proposed approach consists of two phases: Self-Organizing Reservoir Network (SORN) learning and RQA-based hyperparameter tuning. Both steps are applied separately to each of the five channels corresponding to the body parts over time. Figure 2 shows SORN generates five self-organizing reservoirs by clustering the five channels of skeleton data, followed by the implementing RQA on ESRs of each channel.

3.2.1. First Phase: The Self-Organizing Reservoir Network (SORN) Learning

SORN has similar architecture as the ART network and an ITM-based topology construction process. SORN learning aims to learn a set of node centroids and interconnectivity maps from input skeleton data. During the topology construction process, SORN implements edge adaptation of ITM and the least-recently-used node pruning policy. Topological connections, the edges, link the nodes with similar neighboring information and form the cluster of nodes. In a nutshell, the SORN learning algorithm is composed of (i) best-matching node selection, (ii) node matching using vigilance test, (iii) node learning, and (iv) topology construction.
First of all, let us represent the 3D skeleton joint coordinates of a channel at a time instant, t as u ( t ) . In order to accelerate the convergence of SORN clustering [54], noise is added to the input a priori according to Equation (1).
z ( t ) = u ( t ) + ( t 2 η ) I
The degree of the noise is governed by η = [ 0 , 1 ] . I is an identity matrix and it is same size as u ( t ) . The noise in z ( t ) would gradually reduce over time, t. The feasibility of designing self-organizing reservoirs using SORN is explored and analyzed in this work by going through:
  • Best-matching node selection. The learning process starts with a new and empty SORN. Each noisy input z ( t ) at instant t is inserted into the SORN for self-organizing learning. For an empty SORN, z ( t ) is inserted as a new node as
    K K + 1
    w K = z ( t )
    where weight, w K is the weight of node K. Otherwise, the best matching node b and the second-best matching node s are selected based on the similarity comparison with z ( t ) as follows
    k l ( z ( t ) , w j ) = z ( t ) w j 2
    b = argmin j J [ k l ( z ( t ) , C ) ]
    s = argmin j J , j b [ k l ( z ( t ) , C ) ]
    k l is the Euclidean distance measured between the z ( t ) and a node j with weight, w j , both b and s are the indices of the best matching node and the second-best matching node respectively, and C is the self-organizing clustered centroids with J nodes. SORN selects the second-best matching node for the later ITM-based topology construction process.
  • Node matching using vigilance test. Next, a node matching between sample z ( t ) and node b will be succeeding the nomination of the best-matching node. The vigilance test evaluates if the sample z ( t ) stays inside the vigilance region of node b as follows
    k l ( z ( t ) , w b ) V
    where V is the vigilance threshold. Using Hebbian rules, it decides whether to perform node learning or add a new node. If the sample z ( t ) and current node b do not satisfy the vigilance test, SORN will nominate the next candidate of the best matching node to fulfill the condition in Equation (7). If all the current J nodes fail the vigilance test, a sample z ( t ) is added as a new node into the SORN following Equations (2) and (3).
  • Node learning. On the other hand, if the sample z ( t ) satisfies the vigilance test in Equation (7), then SORN performs node learning to update the weight of the best-matching node as follows
    w b = w b + ϵ b ( z ( t ) w b )
    where ϵ b is the learning rate.
  • Topology construction. After node learning, if the second-best matching node satisfies the matching condition in Equation (7), incrementing edges between the best-matching node b and second-best matching node s will be created if it does not exist priorly, as follows:
    Δ E ( b , s ) = 1 .
    SORN constructs the interconnectivity matrix E, which defines the topology of nodes. Taking advantage of ITM, creating a topological map does not require the implementation of edge aging. After creating the connecting edge between node b and node s, vigilance region through node b and node s is compared with vigilance region through node b and every n-th member of N neighborhood nodes of best-matching node b, as follows:
    k l ( w b , w s ) k l ( w b , w n ) .
    Suppose the vigilance region comparison in Equation (10) is satisfied; it implies that node s has a closer neighborhood to node b. Hence, edge deletion removes the connecting edge between node n and node b. For every λ learning cycle, node pruning based on least-recently-used policy prunes any node which does not connect to any topological edges. Although SORN does not require us to predefine the number of nodes, N, before the learning, we use it as the stopping criterion for generating a self-organizing reservoir with N neurons.
The SORN learning process stimulates the formation of a cluster of neighboring nodes that are more specialized and activated by a specific input skeleton data, meanwhile maintaining a certain degree of diversity of the dynamical behavior represented by a different cluster of more distant nodes. SORN defines the topological structure of the reservoir via the formation of interconnectivity maps E. Each neuron interacts with the neighbors according to interconnectivity maps E. It could be a recipe that is compatible with adapting ESN reservoir weights [38]. On the other hand, node centroids adapt the reservoir input weights from input skeleton data. Node centroids, up to a certain extent, capture the topological distribution of the input patterns. Similar input stimuli tend to activate the same cluster of neighboring nodes. Pre-establishing the ESN input weights with node centroids by SORN learning may form the action cluster in the ESN reservoir [38].
Upon SORN generates node centroids C and the interconnectivity maps E, we adopt these weights for the deterministic initialization of the reservoir parameters: input weights, W i n and recurrent hidden weights, W r e s . Input weights and recurrent hidden weights are crucial elements for the existence of echo states and the emergence of suitable dynamics. It is essential to accomplish a few adjustments in these ESN weights during the creation of the self-organizing reservoir. The rescaling procedure is similar to canonical ESN [9]. Firstly, clustered node centroid weights C are normalized and adjusted by the hyperparameter known as input scaling, I S . We rescale each j-th element of node centroid weights to range between I S to I S as in Equation (11) for j J .
c j = I S . 2 c j m i n ( C ) m a x ( C ) m i n ( C ) 1
In the meantime, for the interconnectivity matrix E, we rescale it by the hyperparameter known as spectral radius, S R as in Equation (12).
E r e s c a l e d = S R E λ m a x ( E )
λ m a x ( E ) is the largest eigenvalue of interconnectivity matrix E. We adapt input weights and recurrent weights using Equations (13) and (14) to yield a self-organizing reservoir with N neurons.
W i n = C
W r e s = E r e s c a l e d
Rather than randomly fixed as in canonical ESN, SORN deterministically initializes the reservoir’s input weights and recurrent weights. This reservoir acting as a feature map, generates internal echo states that represent the similar topological conditions of the input action series and yield self-organization.

3.2.2. Second Phase: The RQA-Based Hyperparameter Tuning

The self-organizing reservoirs generated by SORN dramatically depend on the configurations of input scaling I S and spectral radius S R . Unfortunately, the exemplary configuration of these two crucial hyperparameters could often be relying on luck and require the experience of the researchers [12]. Mainly, literature works usually highlight that we must set S R close to but less than 1 to ensure the ESP [9]. It is necessary to have systematic guidance for hyperparameter tuning and a technique well-matched to express the explainability of the sound reasoning of already trained SORN. Given the above context, in this work, we demonstrate the applicability and showcasing of the RQA technique in designing stable self-organizing reservoirs and, by this means, properly facilitate the hyperparameters configuration process for the problem at hand. Combining SORN with the RQA technique to tune self-organizing reservoir hyperparameters and provide post hoc explainability led to the reservoir design approach we called SORN-E.
RQA technique requires building RPs and measurement criteria for tuning reservoir hyperparameters and supporting the explainability of the self-organizing reservoir. Therefore, after SORN learning and generating the self-organizing reservoirs, we construct RPs from ESRs and define the necessary measures.
Assuming u ( t ) represents a D-dimensional 3D skeleton joint coordinates of a single body part at a single time instance t and x ( 0 ) R N is an initial echo-state in the self-organizing reservoir. Similar to canonical ESN, projecting the time-series input at time instant t [ 0 , T 1 ] into the self-organizing reservoir generates an echo state according to the updated equation as follows:
x ( t + 1 ) = f ( W r e s x ( t ) + W i n u ( t + 1 ) ) .
For the T-length time-series input, ESRs are represented by an N-by-T dimensional matrix as follows:
X = ( x ( 0 ) , , x ( T 1 ) ) T ,
where an ESR state x ( t ) is a multivariate time-series with N state variables. Formation of RP requires the creation of a T-by-T binary matrix R based on ESR. Its element R i , j is determined by the Ref. [18]:
R i , j = Θ ( τ R P d ( x [ i ] , x [ j ] ) ) ,
for 1 i , j T , d ( · , · ) is a Euclidean-based dissimilarity measure, Θ ( · ) is the Heaviside function, and τ R P > 0 is a threshold for detecting recurrences. For a more detailed description, we direct the readers to the Ref. [18].
After binary matrix R has been generated, we proceed to prepare the histograms that count the number of diagonal lines with length l and vertical lines with length v as follows
P ( l ) = i , j = 1 T l ( 1 R i 1 , j 1 ) ( 1 R i + l , j + l ) k = 0 l 1 R i + k , j + k
P ( v ) = i , j = 1 T v ( 1 R i , j ) ( 1 R i , j + v ) k = 0 v 1 R i , j + k
Based on the histograms, we then measure the maximum diagonal line length, L M A X as follows
L M A X = m a x { l i } i = 1 N l
where l i is the length of i-th diagonal line, N l is the total number of diagonal lines, and 1 L M A X 2 T . This measure is a state-of-the-art stability measure of ESN reservoirs suggested by the Ref. [18]. It quantifies the degree of reservoir stability based on the diagonal lines in RPs of ESRs: the higher the L M A X , the more stable the reservoir. In the following process, we will start tuning input scaling and spectral radius subject to this reservoir stability criterion.
When tuning input scaling and spectral radius for reservoir stability, primarily, we must first ensure the ESP condition for ESR existence. Hence, we split the hyperparameter tuning process into two stages. In the first stage, considering the previous literature that suggested that we must set S R close to but less than 1 to ensure the ESP [9], we hence fix S R at a constant boundary value of 0.99 [10]. Meanwhile, we iteratively adjust the values of I S and measure its corresponding degree of stability measured by L M A X . Observing the response of L M A X for different values of I S , the value of the highest L M A X indicates the optimal configuration for a stable self-organizing reservoir. In the second stage, applying the empirically tuned input scaling, we fix input scaling and iteratively vary the values of S R and measure its corresponding degree of stability indicated by L M A X . In particular, we are interested in investigating the stability of the self-organizing reservoir when violating ESP, which is setting S R to be greater than unity. Similarly, the value of the highest L M A X indicates the optimal configuration of S R for a stable self-organizing reservoir. The proposed two-stage hyperparameter tuning approach guarantees optimal configuration of I S and S R for stable reservoir design and the satisfaction of ESP. Hence, optimized self-organizing reservoirs are generated.
Besides, we further measure additional RQA metrics and perform several explainability analyses by visualizing the dynamical behavior of the self-organizing reservoir in the context of the HAR problem. In this work, we focus on three dynamical behaviors: laminarity, time dependence and chaoticity [18]. Laminarity, L A M [ 0 , 1 ] quantifies the existence of laminar phase in a reservoir as follows
L A M = v = v m i n T v P ( v ) v = 1 T v P ( v )
where v m i n is the minimal vertical line length used as a threshold. The higher the value of L A M , the more significant a reservoir poses laminar phase, and its echo state changes very slowly for several adjacent time steps. In RP of a reservoir, the occurrence of large black rectangles exhibits laminar phase.
Determinism level, D E T [ 0 , 1 ] quantifies the existence of time dependence in a reservoir as follows
D E T = l = l m i n T l P ( l ) l = 1 T l P ( l )
where l m i n is the minimal diagonal line length used as a threshold. The higher the value of D E T , the more significant a reservoir poses time dependence. Non-uniformly distributed RPs indicates the presence of time dependence.
Recurrence rate, R R quantifies the chaoticity in a reservoir as follows
R R = 1 T 2 i , j = 1 T R i , j
It measures the recurrences density, and a low value of RR indicates the existence of chaoticity in the reservoir. In RP, the occurrence of short and erratic diagonal lines indicates chaoticity.
In HAR, human actions are considered as multivariate time-series data. We hypothesize that projecting human action data onto the self-organizing reservoir generates ESR showing significant laminarity and time dependence dynamics in RPs. Moreover, a stable reservoir offers less chaoticity. Figure 3 summarizes the workflow to generate RP for hyperparameter tuning and explainability analysis using RQA measures.

3.3. Self-Organizing Convolutional Echo State Network (SO-ConvESN)

In this section, we present the applicability and feasibility of the proposed SORN-E. Following the direction of ConvESN, the proposed SORN-E is cascaded with a simple CNN to yield a SO-ConvESN for the HAR task. SORN-E is composed of the optimized self-organizing reservoirs and echo state representations of actions series. CNN consists of convolutional-pooling layers and a classification stage. Figure 4 depicts the general framework of the SO-ConvESN for human action recognition. For ease of visualization, we display the SO-ConvESN with only three filters.
In the context of HAR, each frame of the skeleton series is first split into five individual channels. Each channel corresponds to each 3D-skeleton joint coordinate trajectories of five body parts; left arm (LA), right arm (RA), central trunk (CT), left leg (LL) and right leg (RL).
Each frame of a body part, u ( t ) , is treated as multivariate time-series in the SORN-E stage. For each time instant, t [ 0 , T 1 ] , u ( t ) is mapped into the respective optimized self-organizing reservoir, and echo states are extracted according to Equation (15) and collected into the ESR matrix as in Equation (16). ESR matrix is the complete representation of the T-length input skeleton series. CNN then extracts the discriminative multiscale temporal features from ESR and determines the action class.
In the CNN stage, the process starts with the multiscale convolution operations along the rows of the ESR matrix. In other words, the convolutions are along the direction of time. Multiple convolution filters with bias terms extract temporal features according to each time scale, k. Figure 4 depicts three examples with k = 1 , 2, and 3, and each time scale uses F = 3 filters. Hence, a convolution filter is of size k × N . The central trunk is considered as general multidimensional time-series without any correlation with another channel. However, the spatial correlation between the skeleton joints of left channel and right channel of the arms as well as the left channel and right channel of the legs are highly associated when a subject is performing an action. Hence, they are considered as multivariate time-series with spatial correlation. The single channel multiscale convolution process is not suitable to tackle such scenario. Following the baseline ConvESN, we separately apply dual-channel multiscale convolution operation on the ESRs of left channel and right channel of the arms as well the ESRs of the left channel and right channel of the legs. Figure 5a,b illustrate the single channel and dual-channel multiscale convolution process with F = 3 and k = 1 , 2 .
Similar to ConvESN, we use max-over-time pooling [55] to extract the maximum from the feature maps, which selects the most discriminative local feature regardless of which time step it locates. Max-over-time pooling maintains multiscale temporal invariance. The pooled features are collected into one scalar and then concatenated as one single vector. The output vector of the pooling layer enters a fully connected layer followed by the softmax layer for classification. Assuming C s represents the s-th class of actions, the softmax layer yields conditional distribution P ( C s | u ) over action classes.
After defining the architecture of the proposed SO-ConvESN, deploying for the HAR task requires prior weights training of CNN to be performed. However, the training of a CNN requires the specifications of multiple primary hyperparameters such as number of filters, learning rate, batch size, kernel size, optimizers, activation functions, and the number of epochs. Some of these hyperparameters greatly influence the network performance, and finding the optimal configurations can be viewed as a search problem. Based on preliminary experiments, hand tuning the optimizer and activation function do not show significant impact on the network performance. Therefore, we predefined rectified linear unit (ReLU) [56] as the non-linear activation function in CNN because of its computational efficiency and fast convergence. Meanwhile, the convolution kernel weights are optimized by Adam [14] which is essentially designed to adjust the learning rate adaptively for deep neural network training. But the network is still highly sensitive to the selection of initial learning rate. The cross-entropy error function is used as the training loss during weights optimization as for HAR is a classification problem.
On the other hand, this work explores the optimal configuration of three crucial hyperparameters of CNN: the learning rate, number of filters, and batch size, by applying HPO algorithms. We focus on exploring the feasibility of three HPO algorithms: BO, ASHA, and PBT, for the sake of investigating and selecting the most well-suited method to configure these CNN hyperparameters. We refer the readers to the works of [20,21,22] for further detailed descriptions of BO, ASHA, and PBT, respectively.
During the HPO of CNN in SO-ConvESN, it accepts ESR generated from the projection of action series onto the self-organizing reservoir according to Equations (15) and (16) as inputs and returns the recognition accuracy as numerical output as follows
A c c u r a c y = N u m b e r o f c o r r e c t l y r e c o g n i z e d s a m p l e s T o t a l n u m b e r o f s a m p l e s
HPO aims to find the hyperparameters that maximize the HAR accuracy efficiently. Figure 6 summarizes the general flow of the search loop process.
In general, an HPO algorithm first selects the possible valid configurations of hyperparameters from the search space. It then trains the CNN using the training dataset and evaluates the recognition accuracy using the validation dataset in every optimization run. The ultimate goal of HPO is to determine a set of best-performing hyperparameters. The search space of the CNN hyperparameters is problem-dependent. The detailed configuration of the search space is empirically demonstrated during simulation experiments.

4. Results and Discussion

Intending to demonstrate the capability of the SORN-E in generating stable self-organizing reservoirs and investigate the potential of HPO algorithms for optimizing the recognition performance of the SO-ConvESN, several simulation experiments are conducted using three openly available skeleton-based action recognition benchmark datasets: MSR-Action 3D (MSRA3D) [57], Florence3D-Action (Florence3D) [58], and AHA3D [59]. Each dataset includes a different collection of actions and gestures. AHA3D dataset is merely used for the deployment of the proposed framework for rehabilitation application.
We divide the simulation experiments into several parts. In the first experiments, we present the results in evaluating the feasibility and applicability of the proposed SORN-E to generate stable self-organizing reservoirs. Applying RQA techniques helps to investigate the sensitivity of two crucial hyperparameters of the reservoir and hence select the optimal configurations that ensure stability and ESP. We also aim to achieve more insight into the dynamics and reveal the explainability of the self-organizing reservoir by visualizing the ESRs via RPs and heatmaps.
In addition, the proposed SORN-E is cascaded with a CNN to deal with 3D-skeleton-based action recognition tasks. In the second experiment, we demonstrate the applicability and showcasing of the HPO algorithms in hyperparameter tuning for the CNN of SO-ConvESN. Results of different HPO algorithms are compared and investigated to determine the well-suited HPO algorithm for optimizing the CNN of SO-ConvESN. The last part shows the performance comparison with state-of-the-art approaches and the deployment of SO-ConvESN for HAR application.

4.1. Introduction of Benchmark Datasets

As mentioned before, we adopt three public skeleton-based action recognition benchmark datasets. The first dataset is MSRA3D, composed of 567 sequences with 23,797 skeleton frames recorded at 15 fps with each action performed by ten different subjects 2 or 3 times. It is one of the most famous HAR benchmark datasets used by researchers, which employed Kinect-like sensor to acquire 20 skeleton joints for 20 different activities: high-arm wave, horizontal arm wave, hammer, hand catch, forward punch, high throw, draw X, draw tick, draw a circle, hand clap, two-hand wave, side boxing, bend, forward kick, side kick, jogging, tennis swing, tennis serve, golf swing, and pick-up and throw. Instead of discarding some of the skeleton frames with excessive noise which are missing or corrupted, like in the Refs. [60,61], this work uses the entire dataset intending to assess the noise handling capability of the proposed self-organizing learning to generate reservoirs.
The second dataset is Florence3D, composed of 215 sequences with each action performed by ten different subjects two or three times. It used a Kinect sensor to record 15 skeleton joints for nine various activities: wave, drink from a bottle, answer phone, clap, tight lace, sit down, stand up, read watch, and bow. In this dataset, the same action is performed with both left and right hands, and it consists of actions with high similarities, such as drinking from a bottle and answering the phone. This high intraclass variation makes the recognition task more challenging.
The last dataset is AHA3D, composed of 79 different 3D skeletal videos with 171,753 skeleton frames, with each exercise action being performed one to three times by 21 subjects. It used a Kinect sensor to capture 21 skeleton joints for four different standard fitness exercises: 30 s chair stand, 8 ft up and go, two-minute step test, and unipedal stance. Similar to Florence3D, high intraclass variation makes the recognition task more challenging. The same exercises were performed by the mixture of 11 young subjects and ten elderly subjects. Five of the subjects were male, and 16 were female.
Adopting these three datasets during the reservoir design, explainability analysis, and the HPO of CNN, the proposed SORN-E and SO-ConvESN can be efficiently and empirically evaluated and compared with the state-of-the-art HAR approaches.

4.2. Implementation Details

To begin with, we performed preprocessing of all the raw skeletal sequences in all datasets before using the datasets to train and evaluate the SORN-E and SO-ConvESN. Since the origin references of the raw skeleton joints in a given sequence are different from each other, we first transformed the raw skeleton joints into a normalized coordinate system. For each skeleton sequence, we determined the average of the center, left, and right hip joints and set the origin of every frame in the skeleton sequence to this average point. Then, we applied a Savitzky–Golay smoothing filter [62] to smoothen the normalized skeleton sequences containing different smoothness levels. To ensure each sample to have the same length trajectories, we padded the skeleton sequences up to the maximum length with zeros.
We adopted different training and validation protocols according to the standard proposed in each benchmark dataset to support unbiased comparison with the state-of-the-art performances. In the MSRA3D experiment, we implemented the standard validation protocols as in the Ref. [57]. We created three sets of training and validation configurations where samples from half of the subjects were used as training dataset, and the balance was used as a validation dataset. Meanwhile, in the Florence3D experiment, ten-fold cross-validation protocols as used by the Ref. [58] were adopted for the training and validation process. As for AHA3D experiment, we followed the protocol proposed in the Ref. [59]. We split the available 79 skeletal videos into 39 videos for training, 20 videos for validation, and 20 videos for testing. The performance accuracy was measured and averaged over 100 runs. During performance evaluation and comparison, we employed accuracy as in Equation (24) as the metric for the context of the HAR problem. Better performance is indicated by a higher value of the computed accuracy.
In term of the implementation of SORN-E, we conducted preliminarily experiment of SORN learning to determine the number of self-organizing reservoir neurons. Even though SORN learning adaptively grows the network size, this work used the number of neurons as stopping criterion to fix the size of the reservoir. Whereas for the implementation of RPs and RQA for reservoir explainability, we adjusted the threshold τ R P and ensure τ R P > 0 for detecting recurrences. For SORN reservoir, the configurations of input scaling, I S and spectral radius, S R were tuned by RPs and RQA based techniques and later were fixed to optimal values of 0.1 ( 0.09 for Florence3D) and 0.99 , respectively to ensure stability and ESP for the deployment into SO-ConvESN.
In term of the implementation of HPO algorithms for CNN in SO-ConvESN, we configured the same search space of number of filters, learning rate, batch size for BO, ASHA, and PBT according to the benchmark dataset used. We essentially consolidated the optimization efforts on the learning rate with continuous range. We fixed the convolutional kernel size as 2, 3, and 4 for multiscale feature extraction [10]. PBT is no well-suited to tune hyperparameters that vary the network architecture. Perturbing these hyperparameters may void the weights inherited from best-performing generation. Hence, we fixed the number of filters to adapt PBT during HPO process. Table 1, Table 2 and Table 3 show the search space setups of the MSRA3D, Florence3D, and AHA3D datasets, respectively.
We ran BO, ASHA, and PBT in this experiment with the same respective configuration across all datasets. The hyperparameters of the CNN were optimized with BO algorithm with the expected improvement (EI) as the acquisition function. This process was repeated for 5 iterations for each run with a total of 25 independent runs. For ASHA, we set reduction factor, η to 4, minimum resource, r to 1, maximum resource, R to 64 and minimum early-stopping rate, s to 0. For PBT, we ran HPO with a population size of 10 with five generations. We implemented truncation selection for the exploit stage. During exploration, inherited learning rate were perturbed by either 0.8 or 1.2.
All experimental simulations were conducted in an Intel Core i5-9300H, 2.40 GHz CPU. The algorithms were developed and executed in the Python platform. The time elapsed to produce a self-organizing reservoir by the SORN learning consumed average of 1.6 ms per frame, 1.0 ms per frame, and 3.5 ms per frame for MSRA3D, Florence3D, and AHA3D datasets, respectively.

4.3. Effects of Hyperparameters in Self-Organizing Reservoir Network (SORN)

The effects of the hyperparameters configuration were examined by two benchmark datasets: MSRA3D and Florence3D. The results were obtained by empirically experimenting with different hand tuned settings over multiple SORN learning sessions.
As mentioned earlier, the proposed SORN expands the network size adaptively during the learning process. In order to define the size of the self-organizing reservoir, this work used the user-defined maximum number of neurons as stopping criterion. In addition, the learning cycle was set as 10 % of the max number of nodes. It implies that the pruning process will be regularly performed for every iteration reaching 10 % of the predefined maximum number of nodes. Preliminary experiments showed that as long as the number of reservoir neurons is set to at least three times the input dimension, the reservoir possesses sufficient complexity to capture and hold the information of the input samples. Considering the length of one body part with three skeleton joints, each of them has three coordinate values, contributing to an input length of nine features. Hence, throughout the simulations, we fixed the number of neurons at 36 for MSRA3D. For Florence3D, we set the number of neurons at 27. For AHA3D, we set the number of neurons at 50. Arbitrarily setting the learning rate showed no noticeable impact on the performance. We fixed it at 0.5 to a moderate learning rate during the update of winner nodes.
In terms of the vigilance threshold and reservoir perturbation of SORN, we conducted different training sessions by using a range of values and insert the generated self-organizing reservoirs into SO-ConvESN for HAR tasks using only MSRA3D and Florence3D. We manually tuned the vigilance threshold for a range from 0.05 to 0.95 , and the initial noise distribution scales at 0, 0.1 , 0.01 , and 0.001 to observe the SORN in performing clustering for HAR task. As shown in Figure 7, the results show that configuring the vigilance threshold to 0.05 produced optimal clustering by SORN. Compared to other configurations, setting the vigilance threshold to a low value resulted in the highest validation accuracy in both MSRA3D and Florence3D datasets. Whereas configuring the noise perturbation at 0.1 , SO-ConvESN outperformed the noise-less version, and the results showed no noticeable impact for the magnitudes set at 0.01 and 0.001 .
The results suggest that setting the vigilance threshold at low value is essentially to produce self-organizing clustered topology of nodes with high-granularity. These clusters at a significantly reduced dimensionality can efficiently represent the topological properties of the complete set of input skeleton joint samples. Besides, implementing a tunable noise perturbation as in Equation (1) to control the amount of the initial noise to be added to the input sample before SORN learning offers adjustment for fast convergence clustering [54]. Configuring the noise perturbation to 0.1 enabled SORN-generated self-organizing reservoirs unlocked better performance of the SO-ConvESN.

4.4. RQA-Based Hyperparameter Tuning for Self-Organizing Reservoirs

These additional experiments demonstrate the applicability of the proposed SORN-E to adjust the spectral radius, S R and the input scaling, I S in generating reservoirs that are stable yet satisfied ESP. We conducted the simulation experiments by using both MSRA3D and Florence3D datasets. As described previously, we divided the hyperparameter tuning process into two stages.
In the first stage, we fixed the spectral radius, S R at a constant boundary value of 0.99 on the basis of the previous study showing that this value ensures the ESP [9,10]. We then generated ESR and measured L M A X that indicates the degree of quantified reservoir stability based on Equation (20) against the variation of the input scaling, I S . A higher value of the L M A X represents a more stable configuration of input scaling. We hand tuned the input scaling for a range from 0.07 to 0.5 and measure the respective L M A X . We repeated the measurements using MSRA3D and Florence3D datasets. The measured values of L M A X against the different input scaling settings are depicted in Figure 8a and Figure 9a, respectively. Setting input scaling at a range less than or equal to 0.1 for the MSRA3D dataset and less than or equal to 0.09 for the Florence3D dataset shows optimal stability in which L M A X stays at the highest value.
In the second stage, we applied the optimal value determined in the previous step and fixed the input scaling. We then repeated the measurement of L M A X against the adjustment of the spectral radius, S R . In particular, we investigated the stability of the self-organizing reservoir when violating ESP, which is setting S R to be greater than unity. Similarly, the value of the highest L M A X indicates the optimal configuration of S R for a stable self-organizing reservoir. The measured values of L M A X against the different spectral radius settings for MSRA3D and Florence3D datasets are depicted in Figure 8b and Figure 9b, respectively. Interestingly, setting S R to be slightly greater than unity to violate ESP, self-organizing reservoir stays stable but only sustains up to a certain extend. In both MSRA3D and Florence3D scenarios, the results prove that setting spectral radius less than unity guarantees ESP and ensures stability.
We have applied the proposed SORN-E in generating stable self-organizing reservoirs for all five channels in correspondence to five body parts of the skeleton joints. For simplicity, this section solely showed the results of a single channel self-organizing reservoir. Based on the findings of this experiment, both chosen spectral radius and input scaling affects the stability of the reservoirs. The proposed SORN-E is feasible and applicable to guarantee optimal configuration of I S and S R for stable reservoir and the satisfaction of ESP. We demonstrated the applicability and showcasing of the RQA technique in designing stable self-organizing reservoir. Explainability of the L M A X metric essentially quantifies the reservoir stability and is one of the helpful tools to guide the reservoir design, especially tuning the hyperparameters. In a nutshell, the proposed SORN-E is capable of generating optimized self-organizing reservoirs that are stable and yet fulfilled ESP.

4.5. Comparison between the Self-Organizing Reservoirs and Randomly Initialized Reservoirs Based on Explainability

In the following, we considered that the proposed SORN-E had generated stable self-organizing reservoirs as described previously. We applied the same hyperparameter configuration to generate randomly initialized reservoirs. We compared the self-organizing reservoirs and randomly initialized reservoirs by quantifying the reservoir dynamics via RQA metrics and the visualization of the RPs of ESRs.
We measured RQA metrics and visualized the RPs of ESRs produced by the reservoirs using the MSRA3D and Florence datasets. We aimed to reveal the explainability of the reservoir dynamics for the context of HAR. We focused on three essential classes of reservoir dynamics: laminarity, time dependence, and chaoticity. Figure 10a,b depict the recurrent plots generated using the MSRA3D dataset. Figure 11a,b depict the recurrent plots generated using the Florence3D dataset. We had discovered similar findings of all five channels of action sequences. For visualization simplicity, we merely included the RPs of a single channel.
Comparing the stability measure via L M A X , self-organizing reservoir scores higher values at 7601 and 5874 in MSRA3D and Florence3D datasets, respectively as compared to randomly initialized reservoir. In terms of laminarity, the RPs of both self-organizing reservoir and randomly initialized reservoir shows that echo state varies very gradually over a number of adjacent time steps which can be observed by the presence of large black rectangles. In both MSRA3D and Florence3D, RPs visualizes the existence of laminarity. Comparing the measure L A M , self-organizing reservoir scores higher values at 0.999905 and 0.999980 in MSRA3D and Florence3D dataset, respectively as compared to randomly initialized reservoir. Besides, the RPs are non-uniformly distributed that visualize the time dependence dynamics. It means both reservoirs have captured the correlation of the action sequences. Self-organizing reservoir has superiority in capturing time dependency which can be observed by the measured D E T which is closer to 1 as compared to randomly initialized reservoir in both datasets. The measured R R of self-organizing is also closer to 1 which indicates lesser chaoticity than the randomly initialized reservoir.
Based on the experimental findings, the proposed SORN-E has successfully generated self-organizing reservoirs that are more stable, pose a higher laminarity phase, higher time dependence, and lesser chaoticity. The self-organizing reservoirs have significantly preserved these essential signature dynamics for HAR, which can be understood via the explanatory information extracted from ESRs. Applying SORN-E for deterministic initialization on the ESN’s weight not only ensures stability and ESP, but self-organization also ensures the action sequences’ dynamics are better reflected and captured by the reservoir neuron activations. The SORN-E can be considered one feasible and biologically plausible self-organizing reservoir design approach, notably, encode human actions’ temporal feature for HAR.

4.6. Comparison between the Self-Organizing Reservoirs and Randomly Initialized Reservoirs Based on Reproducibility

In the following, we additionally presented the visualization of ESRs using the MSRA3D dataset. We aimed to investigate the reproducibility of the self-organizing reservoirs generated by SORN-E. For simplicity, we visualized the heatmaps for ESRs of the left arm, right arm, central trunk, left leg, and right leg trajectories of a person performing two-hand waving. We projected this same action input sequence of 40 frames onto the ESN reservoirs to repeat three different trial runs independently. Figure 12, Figure 13, Figure 14, Figure 15, Figure 16 and Figure 17 show the heatmaps of the ESRs for self-organizing reservoirs and randomly initialized reservoirs, respectively.
The visualizations of ESRs show that a self-organizing reservoir reproduces the same heat maps for the identical action sequences in different runs. Conversely, randomly initialized reservoirs produce different ESRs in separate runs. The generated ESRs do not show regular neuron activations behaviors. It is interesting to note that the neuron activations of the self-organizing reservoir along the time length are highly specific. Only a particular group of signature neurons are activated at a time. This group of neurons relevant to a particular part of the body exhibit higher activations at any one time. Whereas for randomly initialized reservoirs, the behaviors are unpredictable and act stochastically. Randomly initialized reservoir failed to reproduce same ESRs in different runs. Moreover, the activations of the neurons do not show any sign of body part relevancy. In other words, randomly initialized reservoirs hardly achieve reproducibility even using the same set of hyperparameter configurations.
Based on the findings of the visualization experiments, SORN-E generates self-organizing reservoirs that preserve the reproducibility for the same input action sequence due to its deterministic initialization of the ESN’s reservoirs. Additionally, a particular body part explicitly activates a specific group of neurons during an action. This finding shows that neurons of the self-organizing reservoirs could be task-specific neurons. This result may also justify that SORN-E produces self-organizing reservoirs that follow the biological mechanism of adaptation of neuron excitability.

4.7. Optimizing the Performance of SO-ConvESN

As highlighted earlier, we cascaded the optimal self-organizing reservoir generated by SORN-E with a simple CNN to yield the SO-ConvESN. In this section, to ensure optimal HAR performance, we implemented BO, AHSA, and PBT algorithms to optimize the HAR accuracy and compared the results of different HPO techniques for the CNN stage that performs the multiscale convolutional process in SO-ConvESN using the MSRA3D and Florence datasets. We also included the baseline SO-ConvESN hand-tuned manually to explore the optimal number of kernels in CNN by specifying the scales from 16 to 256 and the learning rate from 0.001 and 0.003. Table 4 depicts the HAR results achieved by the implemented HPO algorithms. Figure 18 and Figure 19 show the normalized confusion matrices of the recognition accuracy achieved by SO-ConvESN-ASHA for MSRA3D dataset and Florence3D dataset, respectively.
We demonstrate the applicability and showcasing of the HPO algorithms in hyperparameter tuning of CNN in SO-ConvESN to justify that HPO is required to further improve the recognition accuracy in the baseline SO-ConvESN. Based on the experimental observations, all implemented HPO algorithms have improved the recognition accuracy of SO-ConvESN. In MSRA3D models, as compared to the baseline performance, BO improved the accuracy by 0.63%, ASHA achieved improvement of 0.93%, and PBT showed 0.30% better accuracy. Whereas in the Florence3D models, BO showed 4.83% improvement, ASHA achieved 7.10% improved accuracy, and PBT improved the performance by 2.47%. The ASHA significantly showed the best accuracy improvement in optimizing SO-ConvESN in both datasets, whereas BO outperformed PBT in both experiments.
The results demonstrate the effectiveness of ASHA in optimizing the proposed SO-ConvESN for HAR. The inherently sequential optimization of BO makes this method unsuitable for optimizing CNN in SO-ConvESN. For PBT, it could be due to the nature of hyperparameters that variances in the network architecture needs to be fixed. Manual setting of the number of filters in this experiment may degrade the optimization effect when applying PBT. It is also interesting to note that the recognition accuracy in MSRA3D experiments using either of the HPO algorithms exhibit a smaller factor of improvement as compared to Florence3D experiments. We inferred that including the noisy data from the MSRA3D dataset during SORN learning seems to be challenging for SO-ConvESN. We developed the SORN learning based on the ART model. It could be potentially sensitive to noise and makes the SORN-E unable to tolerate unmanageable, noisy MSRA3D data.
In the next section, we further compared the performances of the optimized SO-ConvESNs with the existing HAR approaches using the MSRA3D and Florence datasets. We abbreviated the SO-ConvESN optimized by BO, ASHA, and PBT as SO-ConvESN-BO, SO-ConvESN-ASHA, and SO-ConvESN-PBT, respectively. Table 5 and Table 6 tabulate the state-of-the-art HAR performance of cross-subject test on MSRA3D dataset and cross-validation on the Florence3D dataset.
For MSRA3D experiment, even our best-performing SO-ConvESN, that was optimized with ASHA, exhibited 95.09% overall accuracy. Compared to ConvESN, all of optimized SO-ConvESNs shows lower accuracy. The MSRA3D dataset could be a challenging dataset to SO-ConvESN. In particular, many actions are similar to each other, such as the “draw circle” action which was frequently misinterpreted as the “side boxing” action. Noisy data appearing in the MSRA3D dataset also make the noise-sensitive SORN-E fail to cope. However, considering SORN-E ensures stability and ESP in the self-organizing reservoir, reproducibility is more promising than the randomly initialized reservoir in ConvESN. For the Florence3D experiment, SO-ConvESNs outperformed ConvESN. Our best-performing SO-ConvESN, that was optimized with ASHA, exhibited 96.07% overall accuracy. SORN-E could be good at clustering data with low noise and a small number of action classes, as in Florence3D datasets.
Simulation experiments show that our proposed SO-ConvESN achieves competitive HAR performance with respect to the state-of-the-art approaches. The SORN-E considers self-organization inspired by cortex neuron adjustment mechanism and explainability for tuning ESN’s hyperparameters during the learning of reservoir weights ensures the stable self-organizing reservoirs capture the dynamics and topological properties of the input action sequences. The findings also show that the HPO algorithms are necessary to warrant improved recognition accuracy. The SO-ConvESN can be considered as one feasible and biologically plausible self-organizing reservoir design approach for the HAR problem.

4.8. Deployment of SO-ConvESN for Rehabilitation Application

Previous experimental studies have proven the applicability of the proposed SORN-E to generate stable self-organizing reservoirs, and yet satisfied ESP. Integrating SORN-E with a simple CNN has also been optimized via HPO algorithms to yield optimized SO-ConESN. Considering the outperformance of SO-ConvESN optimized by the ASHA, in this section, we demonstrate the deployment of the proposed approach for an assisted living-oriented performance assessment. We aimed to demonstrate the potential and usefulness of the SO-ConvESN-ASHA in an empirical application that recognizes physical fitness exercises for the elderly rehabilitation application using the AHA3D dataset.
This demonstration followed the proposed development framework as shown in Figure 1. First, the 79 skeletal videos of the AHA3D dataset were randomly split into 39 videos as the training set, 20 videos as the validation set, and 20 videos as the testing set. The training set was used for SORN learning and reservoir hyperparameter tuning. SORN-E generated node centroids and interconnectivity matrix and the optimal configuration of input scaling and spectral radius, which were then used to perform deterministic initialization of the recurrent weight and input weight in SO-ConvESN. The same training set and the previously selected 20 videos of the validation set were used to train and optimize SO-ConvESN using ASHA. Once trained, 20 videos of the testing set were used to evaluate the physical fitness exercise recognition performance of the SO-ConvESN-ASHA. This development and assessment process was iteratively conducted for 100 runs using 20 videos of testing set to recognize four fitness exercise actions: unipedal stance, 8 ft up and go, 30 s chair stand, and 2 min step.
From the 100-run recognition test, the SO-ConvESN-ASHA achieved an average HAR accuracy of 97.1%, with a median of 100%. More than 50 runs of the tests achieved 100% accuracy. Compared to the baseline approach [59], which revealed an average accuracy of 91%, SO-ConvESN-ASHA accomplished about 6.7% improvement. Figure 20 shows the normalized confusion matrix of the overall recognition accuracy for the 100-run experiment. The Class 1 row shows that all 448 videos of “Unipedal Stance” were accurately classified with 100% accuracy. The Class 2 row achievs an accuracy of 95.64%, where 834 of 872 videos representing an “8 ft up and go” were correctly classified, and SO-ConvESN-ASHA wrongly classified 38 videos as “30 s chair stand”. The Class 3 row indicates an accuracy of 96% where 384 of 400 videos for “30 s chair stand” were correctly classified, and SO-ConvESN-ASHA wrongly classified 16 videos as a “2 min step”. The Class 4 row exhibits 98.57% accuracy, with 276 out of 280 were correctly classified as “2 min step”, and SO-ConvESN-ASHA wrongly classified four videos as “30 s chair stand”.
The results of the 100-run performance evaluation show that our proposed development framework, as shown in Figure 1, has successfully trained and optimized SO-ConvESN to recognize the fitness exercise actions with promising performance. In addition to the outperformed recognition accuracy, false recognition of SO-ConvESN-ASHA only shows up to one class. This achievement demonstrates the practicality of the proposed SO-ConvESN.

5. Conclusions

This paper presented a new method for 3D-skeleton-joints-based HAR by cascading a SORN-E into a simple CNN to yield SO-ConvESN. Current studies consider skeleton-joint human actions as multivariate time-series and attempt to identify and model the dynamical temporal features in 3D space. ESNs and their variants are a popular reservoir-computing method suitable for learning the temporal context. Following the body of work of ConvESN, it has incorporated modeling dynamics and multiscale temporal features in a unified framework. Despite the random initialization of the ESN’s input and reservoir, weights may reduce the computational cost, and on the other hand, this may raise instability and variance in generalization and hence diminish reproducibility. Moreover, hyperparameters of the ESN model are commonly tuned based on the rule of thumb or empirically fixed via trial-and-error by researchers’ past experiences. ESN also remains a black-box algorithm. Particularly, it lacks of explainability consideration to understand the input-dependent reservoir dynamics for HAR. To address these problems, we propose SORN-E, a novel self-organizing reservoir design approach that integrates ART architecture and topology construction based on ITM for learning node centroids and interconnectivity maps. To ensure stability and ESP in self-organizing reservoirs, we further exploit the RQA technique to explainability and characterize the dynamics of self-organizing reservoirs, hence tuning two critical ESN hyperparameters: input scaling and spectral radius. SORN-E is compatible with the deterministic initialization and self-organization of the ESN’s input and reservoir weights. The feature maps generated by SORN-E are applied to initialize the input weights and recurrent hidden weights in the ESN to yield optimized self-organizing reservoirs. We cascade the optimized self-organizing reservoirs with a simple CNN to learn the multiscale temporal features from ESRs for action recognition. This resultant novel implementation is named SO-ConvESN. With respect to the learning of the CNN stage, we also adopt three HPO algorithms: BO, ASHA, and PBT to optimize SO-ConvESN for HAR tasks. With the purpose of demonstrating the feasibility and applicability of the proposed approaches, we conduct experiments by using several publicly available 3D-skeleton-based action recognition datasets to examine the explainability of self-organizing reservoir dynamics, investigate the recognition accuracy of SO-ConvESN and the feasibility of implementing HPOs in SO-ConvESN for HAR task. The experimental results have demonstrated the applicability of SORN-E in deterministic initialization of ESN’s weights for reproducibility and generating optimized self-organizing reservoirs and the potentiality of the proposed SO-ConvESN in achieving competitive HAR performance compared with state-of-the-art approaches. Future works may consider unlocking the potential of SO-ConvESN for general time-series classification applications. Improvement of the robustness of the SO-ConvESN in noise handling capability may be another avenue for exploration and investigation.

Author Contributions

Conceptualization, C.K.L. and G.C.L.; methodology, C.K.L. and G.C.L.; software, G.C.L.; validation, C.K.L. and G.C.L.; formal analysis, C.K.L. and G.C.L.; investigation, G.C.L.; resources, C.K.L. and G.C.L.; data curation, G.C.L.; writing—original draft preparation, C.K.L. and G.C.L.; writing—review and editing, G.C.L.; visualization, G.C.L.; supervision, C.K.L.; project administration, C.K.L.; funding acquisition, C.K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the COVID-19 Special Research Grant under Project CSRG008-2020ST and Impact Oriented Interdisciplinary Research Grant Programme (IIRG), IIRG002C-19HWB from Universiti Malaya.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs. Publicly available datasets were analyzed in this study. This data can be found here: [https://sites.google.com/view/wanqingli/data-sets/msr-action3d] https://www.micc.unifi.it/resources/datasets/florence-3d-actions-dataset/; https://vislab.isr.tecnico.ulisboa.pt/datasets/ (accessed on 24 September 2019).

Acknowledgments

The authors would like to thank the anonymous reviewer for his/her comments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHA3DAugmented Human Assistance 3D
ARTAdaptive Resonant Theory
ASHAAsynchronous Successive Halving Algorithm
BOBayesian Optimization
ConvESNConvolutional Echo State Network
CTCentral Trunk
DETDeterminism level
EIExpected Improvement
ESNsEcho State Networks
ESPEcho State Property
Florence3DFlorence 3D Actions
GNGGrowing Neural Gas
HPOHyperparameter Optimization
HARHuman Action Recognition
ITMInstantaneous Topological Mapping
LALeft Arm
LAMLaminarity
LI-ESNsLeaky Integrator-Echo State Networks
LLLeft Leg
L M A X Maximum diagonal line length
LSTMLong Short-Term Memory
MSRA3DMicrosoft Research Action 3D
NGNeural Gas
PBTPopulation-based Training
RARight Arm
RCReservoir Computing
ReLURectified Linear Unit
RLRight Leg
RNNsRecurrent Neural Networks
RPsRecurrent Plots
RQARecurrence Quantification Analysis
RRRecurrence Rate
SHASuccessive Halving Algorithm
SO-ConvESNSelf-Organizing Convolutional Echo State Network
SOMSelf-Organizing Map
SORNSelf-Organizing Reservoir Network
SORN-ESelf-Organizing Reservoir Network with Explainability

References

  1. Zhang, Z.; Ma, X.; Song, R.; Rong, X.; Tian, X.; Tian, G.; Li, Y. Deep learning based human action recognition: A survey. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; pp. 3780–3785. [Google Scholar]
  2. Cichy, R.M.; Khosla, A.; Pantazis, D.; Torralba, A.; Oliva, A. Deep neural networks predict hierarchical spatio-temporal cortical dynamics of human visual object recognition. arXiv 2016, arXiv:1601.02970. [Google Scholar]
  3. Onofri, L.; Soda, P.; Pechenizkiy, M.; Iannello, G. A survey on using domain and contextual knowledge for human activity recognition in video streams. Expert Syst. Appl. 2016, 63, 97–111. [Google Scholar] [CrossRef]
  4. Presti, L.L.; Cascia, M.L. 3D Skeleton-based Human Action Classification: A Survey. Pattern Recognit. 2016, 53, 130–147. [Google Scholar] [CrossRef]
  5. Papadopoulos, G.T.; Axenopoulos, A.; Daras, P. Real-time skeleton-tracking-based human action recognition using Kinect data. In Proceedings of the 20th Anniversary International Conference on MultiMedia Modeling, Dublin, Ireland, 6–10 January 2014; pp. 473–483. [Google Scholar]
  6. Huang, J.D. Kinerehab: A Kinect-based system for physical rehabilitation: A pilot study for young adults with motor disabilities. In Proceedings of the 13th international ACM SIGACCESS Conference on Computers and Accessibility, Dundee, UK, 24–26 October 2011; pp. 319–320. [Google Scholar]
  7. Kong, Y.; Fu, Y. Human action recognition and prediction: A survey. arXiv 2018, arXiv:1806.11230v2. [Google Scholar]
  8. Shotton, J.; Fitzgibbon, A.; Cook, M.; Sharp, T.; Finocchio, M.; Moore, R.; Kipman, A.; Blake, A. Real-time human pose recognition in parts from single depth images. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 1297–1304. [Google Scholar]
  9. Jaeger, H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. Tech. Rep. Ger. Natl. Res. Cent. Inform. Technol. GMD 2001, 148, 172–185. [Google Scholar]
  10. Ma, Q.; Shen, L.; Chen, E.; Tian, S.; Wang, J.; Cottrell, G.W. WALKING WALKing walking: Action Recognition from Action Echoes. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2457–2463. [Google Scholar]
  11. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  12. Wu, Q.; Fokoue, E.P.; Kudithipudi, D. On the statistical challenges of echo state networks and some potential remedies. arXiv 2018, arXiv:1802.07369. [Google Scholar]
  13. Mici, L.; Hinaut, X.; Wermter, S. Activity recognition with echo state networks using 3D body joints and objects category. In Proceedings of the European Symposium On Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 27–29 April 2016; pp. 465–470. [Google Scholar]
  14. Kingma, D.; Lei-Ba, J. Adam: A method for stochastic optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  15. Bochinski, E.; Senst, T.; Sikora, T. Hyper-parameter optimization for convolutional neural network committees based on evolutionary algorithms. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3924–3928. [Google Scholar]
  16. Grossberg, S. Competitive learning: From interactive activation to adaptive resonance. Cogn. Sci. 1987, 11, 23–63. [Google Scholar] [CrossRef]
  17. Jockusch, J.; Ritter, H. An instantaneous topological mapping model for correlated stimuli. In Proceedings of the International Joint Conference on Neural Networks, Washington, DC, USA, 10–16 July 1999; Volume 1, pp. 529–534. [Google Scholar]
  18. Bianchi, F.M.; Livi, L.; Alippi, C. Investigating Echo-State Networks Dynamics by Means of Recurrence Analysis. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 427–439. [Google Scholar] [CrossRef]
  19. Nelson, C.A. Neural plasticity and human development: The role of early experience in sculpting memory systems. Dev. Sci. 2000, 3, 115–136. [Google Scholar] [CrossRef]
  20. Snoek, J.; Rippel, O.; Swersky, K.; Kiros, R.; Satish, N.; Sundaram, N.; Patwary, M.M.A.; Prabhat.; Adams, R.P. Scalable Bayesian Optimization Using Deep Neural Networks. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 7–9 July 2015; Volume 37, pp. 2171–2180. [Google Scholar]
  21. Li, L.; Jamieson, K.; Rostamizadeh, A.; Gonina, E.; Hardt, M.; Recht, B.; Talwalkar, A. A System for Massively Parallel Hyperparameter Tuning. arXiv 2020, arXiv:1810.05934. [Google Scholar]
  22. Jaderberg, M.; Dalibard, V.; Osindero, S.; Czarnecki, W.M.; Donahue, J.; Razavi, A.; Vinyals, O.; Green, T.; Dunning, I.; Simonyan, K.; et al. Population Based Training of Neural Networks. arXiv 2017, arXiv:1711.09846. [Google Scholar]
  23. Biswas, K.; Basu, S.K. Gesture recognition using Microsoft Kinect. In Proceedings of the 5th International Conference on Automation, Robotics and Applications, Wellington, New Zealand, 6–8 December 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 100–103. [Google Scholar]
  24. Bhattacharya, S.; Czejdo, B.; Perez, N. Gesture classification with machine learning using Knect sensor data. In Proceedings of the 3rd International Conference on Emerging Applications of Information Technology, Kolkata, India, 29 November–1 December 2012; pp. 348–351. [Google Scholar]
  25. Patsadu, O.; Nukoolkit, C.; Watanapa, B. Human gesture recognition using Kinect camera. In Proceedings of the 2012 Ninth International Conference on Computer Science and Software Engineering (JCSSE), Bangkok, Thailand, 30 May–1 June 2012; pp. 28–32. [Google Scholar]
  26. Dupont, M.; Marteau, P.F. Coarse-dtw for sparse time series alignment. In Advanced Analysis and Learning on Temporal Data; Lecture Notes in Computer Science; Douzal-Chouakria, A., Vilar, J., Marteau, P.F., Eds.; Springer: Cham, Swizerland, 2016; Volume 9785, pp. 157–172. [Google Scholar]
  27. Ibanez, R.; Soria, A.; Teyseyre, A.; Camp, M. Easy gesture recognition for Kinect. Adv. Eng. Softw. 2014, 76, 171–180. [Google Scholar] [CrossRef]
  28. Marteau, P.F.; Gibet, S.; Reverdy, C. Down-sampling coupled to elastic kernel machines for efficient recognition of isolated gestures. In Proceedings of the 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 363–368. [Google Scholar]
  29. Lev, G.; Sadeh, G.; Klein, B.; Wolf, L. RNN Fisher Vectors for Action Recognition and Image Annotation. arXiv 2016, arXiv:1512.03958v1. [Google Scholar]
  30. Zhang, P.; Lan, C.; Xing, J.; Zeng, W.; Xue, J.; Zheng, N. View adaptive neural networks for high performance skeleton-based human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1963–1978. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Gallicchio, C.; Micheli, A. A reservoir computing approach for human gesture recognition from Kinect data. In Proceedings of the Workshop Artificial Intelligence for Ambient Assisted Living, Genova, Italy, 28 November 2016; Volume 1803, pp. 33–42. [Google Scholar]
  32. Ma, Q.; Zhuang, W.; Shen, L.; Cottrell, G.W. Time series classification with Echo Memory Networks. Neural Netw. 2019, 117, 225–239. [Google Scholar] [CrossRef] [Green Version]
  33. Bengio, Y. Deep learning of representations: Looking forward. In Proceedings of the International Conference on Statistical Language and Speech Processing, Tarragona, Spain, 29–31 July 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 1–37. [Google Scholar]
  34. Nweke, H.F.; Wah, T.Y.; Al-garadi, M.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
  35. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  36. Palangi, H.; Deng, L.; Ward, R. Learning Input and Recurrent Weight Matrices in Echo State Networks. arXiv 2013, arXiv:1311.2987. [Google Scholar]
  37. Arrieta, A.B.; Gil-Lopez, S.; Laña, I.; Bilbao, M.N.; Ser, J.D. On the Post-hoc Explainability of Deep Echo State Networks for Time Series Forecasting, Image and Video Classification. arXiv 2021, arXiv:2102.08634. [Google Scholar]
  38. Boccato, L.; de Faissol Attux, R.; Zuben, F.J.V. Self-organization and lateral interaction in echo state network reservoirs. Neurocomputing 2014, 138, 297–309. [Google Scholar] [CrossRef]
  39. Kohonen, T. Self-organized formation of topologically correct feature maps. Biol. Cybern. 2004, 43, 59–69. [Google Scholar] [CrossRef]
  40. Fritzke, B. A Growing Neural Gas Network Learns Topologies. In Proceedings of the 7th International Conference on Neural Information Processing Systems, 1 January 1994. [Google Scholar]
  41. Martinetz, T.M.; Schulten, K.J. A “Neural Gas” Network Learns Topologies. In Proceedings of the International Conference on Artificial Neural Networks, Espoo, Finland, 24–28 June 1991; Elsevier: Amsterdam, The Netherlands; New York, NY, USA, 1991; pp. 397–402. [Google Scholar]
  42. Fritzke, B. Growing cell structures: A self-organizing network for unsupervised and supervised learnin. Neural Netw. 1994, 7, 1441–1460. [Google Scholar] [CrossRef]
  43. Carpenter, G.A.; Grossberg, S. The ART of adaptive pattern recognition by a self-organizing neural network. Computer 1988, 3, 77–88. [Google Scholar] [CrossRef]
  44. Marwan, N.; Romano, M.C.; Thiel, M.; Kurths, J. Recurrence plots for the analysis of complex systems. Phys. Rep. 2007, 438, 237–329. [Google Scholar] [CrossRef]
  45. Marwan, N. How to Avoid Potential Pitfalls in Recurrence Plot Based Data Analysis. Int. J. Bifurc. Chaos 2011, 21, 1003–1017. [Google Scholar] [CrossRef] [Green Version]
  46. Eroglu, D.; Peron, T.K.D.M.; Marwan, N.; Rodrigues, F.A.; da Fontoura Costa, L.; Sebek, M.; Kiss, I.Z.; Kurths, J. Entropy of weighted recurrence plots. Phys. Rev. E 2014, 90, 042919. [Google Scholar] [CrossRef] [Green Version]
  47. Gallicchio, C.; Micheli, A.; Pedrelli, L. Deep reservoir computing: A critical experimental analysis. Neurocomputing 2017, 268, 87–99. [Google Scholar] [CrossRef]
  48. Xie, L.; Yuille, A.L. Genetic CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1388–1397. [Google Scholar]
  49. Bergstra, J.; Bengio, Y. Random Search for Hyper-Parameter Optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  50. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, CA, USA, 3–6 December 2012. [Google Scholar]
  51. Domhan, T.; Springenberg, J.T.; Hutter, F. Speeding Up Automatic Hyperparameter Optimization of Deep Neural Networks by Extrapolation of Learning Curves. In Proceedings of the IJCAI International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
  52. Eggensperger, K. Towards an Empirical Foundation for Assessing Bayesian Optimization of Hyperparameters. In Proceedings of the NIPS Workshop on Bayesian Optimization in Theory and Practice, Lake Tahoe, NV, USA, 10 December 2013; pp. 1–5. [Google Scholar]
  53. Jamieson, K.G.; Talwalkar, A.S. Non-stochastic Best Arm Identification and Hyperparameter Optimization. arXiv 2016, arXiv:1502.07943. [Google Scholar]
  54. Osoba, O.; Kosko, B. Noise-enhanced clustering and competitive learning algorithm. Neural Netw. 2013, 37, 132–140. [Google Scholar] [CrossRef] [PubMed]
  55. Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P.P. Natural Language Processing (Almost) from Scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
  56. Nair, V.; Hinto, G.E. Rectified Linear Units Improve Restricted Boltzmann Machine. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  57. Li, W.; Zhang, Z.; Liu, Z. Action recognition based on a bag of 3d point. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 9–14. [Google Scholar]
  58. Seidenari, L.; Varano, V.; Berretti, S.; Bimbo, A.D.; Pala, P. Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Portland, OR, USA, 23–28 June 2013; pp. 479–485. [Google Scholar]
  59. Antunes, J.; Bernardino, A.; Smailagic, A.; Siewiorek, D. AHA-3D: A Labelled Dataset for Senior Fitness Exercise Recognition and Segmentation from 3D Skeletal Data. In Proceedings of the Vision International Behaviour Understanding Workshop, British Machine Vision Conference, Newcastle upon Tyne, UK, 3–6 September 2018. [Google Scholar]
  60. Zhang, X.; Wang, Y.; Gou, M.; Sznaier, M.; Camps, O. Efficient temporal sequence comparison and classification using gram matrix embeddings on a riemannian manifold. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4498–4507. [Google Scholar]
  61. Bengio, Y. Practical recommendations for gradient-based training of deep architectures. In Neural Networks: Tricks Trade; Springer: Berlin/Heidelberg, Germany, 2012; pp. 437–478. [Google Scholar]
  62. Steinier, J.; Termonia, Y.; Deltour, J.; Chem, A. Smoothing and differentiation of data by simplified least square procedure. Anal. Chem. 1972, 44, 1906–1909. [Google Scholar] [CrossRef]
  63. Hussein, M.E.; Torki, M.; Gowayyed, M.A.; El-Saban, M. Human action recognition using a temporal hierarchy of covariance descriptors on 3d joint locations. In Proceedings of the Twenty-third international joint conference on artificial intelligence, Beijing, China, 3–9 August 2013; Volume 13, pp. 2466–2472. [Google Scholar]
  64. Vemulapalli, R.; Arrate, F.; Chellappa, R. Human action recognition by representing 3d skeletons as points in a lie group. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 588–595. [Google Scholar]
  65. Presti, L.L.; Cascia, M.L.; Sclaroff, S.; Camps, O.I. Hankelet-based dynamical systems modeling for 3d action recognition. Image Vis. Comput. 2015, 44, 29–43. [Google Scholar] [CrossRef] [Green Version]
  66. Gao, X.; Hu, W.; Tang, J.; Liu, J.; Guo, Z. Optimized Skeleton-based Action Recognition via Sparsified Graph Regression. arXiv 2019, arXiv:1811.12013. [Google Scholar]
  67. Li, C.; Cui, Z.; Zheng, W.; Xu, C.; Yang, J. Spatio-Temporal Graph Convolution for Skeleton Based Action Recognition. arXiv 2018, arXiv:1802.09834. [Google Scholar]
Figure 1. Development framework of the Proposed SO-ConvESN for HAR. It consists of two key components: the SORN-E stage and the SO-ConvESN stage.
Figure 1. Development framework of the Proposed SO-ConvESN for HAR. It consists of two key components: the SORN-E stage and the SO-ConvESN stage.
Sensors 22 01905 g001
Figure 2. Five channels from human skeleton data: left arm (LA), right arm (RA), left leg (LL), right leg (RL), and central trunk (CT) are first extracted during data preparation. Then, SORN learning generates the respective self-organizing reservoirs independently. RQA is applied on the corresponding ESRs for hyperparameter tuning.
Figure 2. Five channels from human skeleton data: left arm (LA), right arm (RA), left leg (LL), right leg (RL), and central trunk (CT) are first extracted during data preparation. Then, SORN learning generates the respective self-organizing reservoirs independently. RQA is applied on the corresponding ESRs for hyperparameter tuning.
Sensors 22 01905 g002
Figure 3. Generating RP for hyperparameter tuning and explainability analysis using RQA measure. u ( t ) is the action input at instant t. Projecting u ( t ) onto the self-organizing reservoir with N neurons generates an echo state x ( t ) .
Figure 3. Generating RP for hyperparameter tuning and explainability analysis using RQA measure. u ( t ) is the action input at instant t. Projecting u ( t ) onto the self-organizing reservoir with N neurons generates an echo state x ( t ) .
Sensors 22 01905 g003
Figure 4. Overview architecture of SO-ConvESN. Optimized self-organizing reservoirs are generated by SORN-E and cascaded in multi-scale CNN with three time-scales, three filters and five channels for human action recognition. HPO implementation is solely conducted for the CNN stage.
Figure 4. Overview architecture of SO-ConvESN. Optimized self-organizing reservoirs are generated by SORN-E and cascaded in multi-scale CNN with three time-scales, three filters and five channels for human action recognition. HPO implementation is solely conducted for the CNN stage.
Sensors 22 01905 g004
Figure 5. Multiscale convolution process extracts temporal features from ESR. (a) Single channel convolution is used for CT channel; (b) Dual-Channel Convolution is used for LA and RA as well as LL and RL channels.
Figure 5. Multiscale convolution process extracts temporal features from ESR. (a) Single channel convolution is used for CT channel; (b) Dual-Channel Convolution is used for LA and RA as well as LL and RL channels.
Sensors 22 01905 g005
Figure 6. A general search loop of HPO implementation in CNN stage of SO-ConvESN. Every optimization run loop will train a CNN based on the selected hyperparameters and evaluate the recognition accuracy based on the training and validation datasets. best-performing configuration and the corresponding accuracy are the outcomes of the search loop.
Figure 6. A general search loop of HPO implementation in CNN stage of SO-ConvESN. Every optimization run loop will train a CNN based on the selected hyperparameters and evaluate the recognition accuracy based on the training and validation datasets. best-performing configuration and the corresponding accuracy are the outcomes of the search loop.
Sensors 22 01905 g006
Figure 7. The graphs depict the impact of tuning the vigilance threshold from 0.05 to 0.95 on the performance of SO-ConvESN in HAR task based on MSRA3D dataset (Red Solid line) and Florence3D dataset (Blue Solid line). In both models, when the vigilance threshold increases, validation accuracy drops.
Figure 7. The graphs depict the impact of tuning the vigilance threshold from 0.05 to 0.95 on the performance of SO-ConvESN in HAR task based on MSRA3D dataset (Red Solid line) and Florence3D dataset (Blue Solid line). In both models, when the vigilance threshold increases, validation accuracy drops.
Sensors 22 01905 g007
Figure 8. Results on the MSRA3D dataset. Measured reservoir stability, L M A X against different hyperparameter settings.
Figure 8. Results on the MSRA3D dataset. Measured reservoir stability, L M A X against different hyperparameter settings.
Sensors 22 01905 g008
Figure 9. Results on the Florence3D dataset. Measured reservoir stability, L M A X against different hyperparameter settings.
Figure 9. Results on the Florence3D dataset. Measured reservoir stability, L M A X against different hyperparameter settings.
Sensors 22 01905 g009
Figure 10. Recurrence plots using MSRA3D dataset and setting optimal S R at 0.99 and I S at 0.1. (a) Self-organizing reservoir with RQA metrics: L M A X = 7601, L A M = 0.999905, D E T = 0.999976, R R = 0.999052; (b) Randomly initialized reservoir with RQA metrics: L M A X = 4419, L A M = 0.989145, D E T = 0.985305, R R = 0.931044.
Figure 10. Recurrence plots using MSRA3D dataset and setting optimal S R at 0.99 and I S at 0.1. (a) Self-organizing reservoir with RQA metrics: L M A X = 7601, L A M = 0.999905, D E T = 0.999976, R R = 0.999052; (b) Randomly initialized reservoir with RQA metrics: L M A X = 4419, L A M = 0.989145, D E T = 0.985305, R R = 0.931044.
Sensors 22 01905 g010
Figure 11. Recurrence plots using Florence3D dataset and setting optimal S R at 0.99 and I S at 0.09. (a) Self-organizing reservoir with RQA metrics: L M A X = 5874, L A M = 0.999980, D E T = 0.999999, R R = 0.998964; (b) Randomly initialized with RQA metrics: L M A X = 5075, L A M = 0.998944, D E T = 0.999626, R R = 0.992221.
Figure 11. Recurrence plots using Florence3D dataset and setting optimal S R at 0.99 and I S at 0.09. (a) Self-organizing reservoir with RQA metrics: L M A X = 5874, L A M = 0.999980, D E T = 0.999999, R R = 0.998964; (b) Randomly initialized with RQA metrics: L M A X = 5075, L A M = 0.998944, D E T = 0.999626, R R = 0.992221.
Sensors 22 01905 g011
Figure 12. Visualization of ESRs for an action sequence of a person performing two-hand waving with 40 frames that was projected onto the self-organizing reservoirs with 36 neurons generated by SORN-E. Vertical axis indicates the number of reservoir neurons and horizontal axis indicates the time frames. The results are produced by projecting the same action time-series onto self-organizing reservoirs to produce ESRs for three different trial runs. Self-organizing reservoirs ensure deterministic initialization of the reservoir weights for reproducibility.
Figure 12. Visualization of ESRs for an action sequence of a person performing two-hand waving with 40 frames that was projected onto the self-organizing reservoirs with 36 neurons generated by SORN-E. Vertical axis indicates the number of reservoir neurons and horizontal axis indicates the time frames. The results are produced by projecting the same action time-series onto self-organizing reservoirs to produce ESRs for three different trial runs. Self-organizing reservoirs ensure deterministic initialization of the reservoir weights for reproducibility.
Sensors 22 01905 g012
Figure 13. Visualization of ESRs of randomly initialized reservoir with 36 neurons for left arm trajectories of a person performing two-hand waving.
Figure 13. Visualization of ESRs of randomly initialized reservoir with 36 neurons for left arm trajectories of a person performing two-hand waving.
Sensors 22 01905 g013
Figure 14. Visualization of ESRs of randomly initialized reservoir with 36 neurons for right arm trajectories of a person performing two-hand waving.
Figure 14. Visualization of ESRs of randomly initialized reservoir with 36 neurons for right arm trajectories of a person performing two-hand waving.
Sensors 22 01905 g014
Figure 15. Visualization of ESRs of randomly initialized reservoir with 36 neurons for central trunk trajectories of a person performing two-hand waving.
Figure 15. Visualization of ESRs of randomly initialized reservoir with 36 neurons for central trunk trajectories of a person performing two-hand waving.
Sensors 22 01905 g015
Figure 16. Visualization of ESRs of randomly initialized reservoir with 36 neurons for left leg trajectories of a person performing two-hand waving.
Figure 16. Visualization of ESRs of randomly initialized reservoir with 36 neurons for left leg trajectories of a person performing two-hand waving.
Sensors 22 01905 g016
Figure 17. Visualization of ESRs of randomly initialized reservoir with 36 neurons for right leg trajectories of a person performing two-hand waving.
Figure 17. Visualization of ESRs of randomly initialized reservoir with 36 neurons for right leg trajectories of a person performing two-hand waving.
Sensors 22 01905 g017
Figure 18. The normalized confusion matrix for SO-ConvESN-ASHA on MSRA3D dataset.
Figure 18. The normalized confusion matrix for SO-ConvESN-ASHA on MSRA3D dataset.
Sensors 22 01905 g018
Figure 19. The normalized confusion matrix for SO-ConvESN-ASHA on Florence3D-Action dataset.
Figure 19. The normalized confusion matrix for SO-ConvESN-ASHA on Florence3D-Action dataset.
Sensors 22 01905 g019
Figure 20. The normalized confusion matrix for 100 runs using 20 videos of the testing set. Classes 1, 2, 3, and 4 indicate unipedal stance, 8 ft up and go, 30 s chair stand, and 2 min step, respectively.
Figure 20. The normalized confusion matrix for 100 runs using 20 videos of the testing set. Classes 1, 2, 3, and 4 indicate unipedal stance, 8 ft up and go, 30 s chair stand, and 2 min step, respectively.
Sensors 22 01905 g020
Table 1. Search Space Setup for MSRA3D Dataset.
Table 1. Search Space Setup for MSRA3D Dataset.
HyperparameterTypeValues
Number of filtersDiscrete128, 256
Learning rateContinuous0.001, 0.003
Batch sizeDiscrete1, 4
Table 2. Search Space Setup for Florence3D Dataset.
Table 2. Search Space Setup for Florence3D Dataset.
HyperparameterTypeValues
Number of filtersDiscrete256, 512
Learning rateContinuous0.001, 0.002
Batch sizeDiscrete1, 4
Table 3. Search Space Setup for AHA3D Dataset.
Table 3. Search Space Setup for AHA3D Dataset.
HyperparameterTypeValues
Number of filtersDiscrete64, 128, 256
Learning rateContinuous0.001, 0.003
Batch sizeDiscrete1, 4
Table 4. Performance Comparison of SO-ConvESN optimized by three different HPO Algorithms based on Validation Accuracy (%).
Table 4. Performance Comparison of SO-ConvESN optimized by three different HPO Algorithms based on Validation Accuracy (%).
AlgorithmsBaselineBOAHSAPBT
MSRA3D Dataset94.2194.8095.0994.49
Florence3D Dataset89.7094.0396.0791.92
Table 5. Recognition accuracy on cross-subject test of the MSRA3D dataset.
Table 5. Recognition accuracy on cross-subject test of the MSRA3D dataset.
ApproachesAverage (%)
Covariance [63]88.10
Skeletons Lie group [64]92.40
DHMM+SL [65]92.91
SO-ConvESN (Our approach)94.21
SO-ConvESN-PBT (Our approach)94.49
SO-ConvESN-BO (Our approach)94.80
SO-ConvESN-ASHA (Our approach)95.09
Gram matrices rep. [60]96.90
ConvESN [10]97.88
Table 6. Recognition accuracy on 10-fold cross-validation Florence3D-Action dataset.
Table 6. Recognition accuracy on 10-fold cross-validation Florence3D-Action dataset.
ApproachesAverage (%)
Multi-Part Bag-of-Poses [58]82.00
SO-ConvESN (Our approach)89.70
Skeletons Lie group [64]90.88
ConvESN [10]91.72
SO-ConvESN-PBT (Our approach)91.92
SO-ConvESN-BO (Our approach)94.03
SO-ConvESN-ASHA (Our approach)96.07
Complete GR-GCN [66]98.40
Deep STGC K  [67]99.10
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, G.C.; Loo, C.K. On the Post Hoc Explainability of Optimized Self-Organizing Reservoir Network for Action Recognition. Sensors 2022, 22, 1905. https://doi.org/10.3390/s22051905

AMA Style

Lee GC, Loo CK. On the Post Hoc Explainability of Optimized Self-Organizing Reservoir Network for Action Recognition. Sensors. 2022; 22(5):1905. https://doi.org/10.3390/s22051905

Chicago/Turabian Style

Lee, Gin Chong, and Chu Kiong Loo. 2022. "On the Post Hoc Explainability of Optimized Self-Organizing Reservoir Network for Action Recognition" Sensors 22, no. 5: 1905. https://doi.org/10.3390/s22051905

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop