Next Article in Journal
Hemodynamic Response Asymmetry During Motor Imagery in Stroke Patients: A Novel NIRS-BCI Assessment Approach
Previous Article in Journal
Intelligent Defect Detection of Ancient City Walls Based on Computer Vision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PG-Mamba: An Enhanced Graph Framework for Mamba-Based Time Series Clustering

1
Department of Computer Science and Technology, College of Computer and Information Engineering, Inner Mongolia Agricultural University, Hohhot 010011, China
2
Inner Mongolia Autonomous Region Big Data Center, Hohhot 010091, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(16), 5043; https://doi.org/10.3390/s25165043 (registering DOI)
Submission received: 12 June 2025 / Revised: 6 August 2025 / Accepted: 7 August 2025 / Published: 14 August 2025

Abstract

Time series clustering finds wide application but is often limited by data quality and the inherent limitations of existing methods. Compared to high-dimensional structured data like images, the low-dimensional features of time series contain less information, and endogenous noise can easily obscure important patterns. When dealing with massive time series data, existing clustering methods often focus on mining associations between sequences. However, ideal clustering results are difficult to achieve by relying solely on pairwise association analysis in the presence of noise and information scarcity. To address these issues, we propose a framework called Patch Graph Mamba (PG-Mamba). For the first time, the spatio-temporal patterns of a single sequence are explored by dividing the time series into multiple patches and constructing a spatio-temporal graph (STG). In this graph, these patches serve as nodes, connected by both spatial and temporal edges. By leveraging Mamba-driven long-range dependency learning and a decoupled spatio-temporal graph attention mechanism, our framework simultaneously captures temporal dynamics and spatial relationships and, thus, enabling the effective extraction of key information from time series. Furthermore, a spatio-temporal adjacency matrix reconstruction loss is introduced to mitigate feature space perturbations induced by the clustering loss. Experimental results demonstrate that PG-Mamba outperforms state-of-the-art methods, offering new insights into time series clustering tasks. Across the 33 datasets of the UCR time series archive, PG-Mamba achieved the highest average rank of 3.606 and secured the most first-place rankings (13).

1. Introduction

Research on time series clustering has gained popularity among academics because of its capacity for identifying patterns in the evolution of data and analyzing its intrinsic characteristics. However, the inherent mistakes and trends in time series data, such as periodicity and seasonality, surely present difficulties for clustering tasks. As a result, identifying important features in time series data and creating reliable representations have become important study areas that need more exploration. There are many different clustering methods available today, which we have intuitively divided into the following three primary categories: raw data-based methods, feature-based methods, and deep learning-based methods.
Raw data-based methods. As is well known, K-means [1] is appropriate for large-scale datasets due to its straightforward logic and significant cluster interpretability. However, Euclidean distance cannot be used to quantify the noise and scale variance found in time-series data. To accommodate time series data of different lengths or data of the same length but irregular in time, dynamic time warping (DTW) is applied. For example, Petitjean et al. [2] proposed Dynamic Time Warping Center Average (DBA) based on the concept of global averaging of cluster centers, addressing the issue that DTW aligns two time series data but cannot derive a global average as the cluster center. In addition, Yang et al. [3] proposed the K-Spectral Centroid (KSC). By varying scaling during clustering, KSC presents a novel distance computing method with appropriate translation and scaling factors to capture the time series data’s evolutionary processes. Instead of using DBA to choose average sequences at random, k-shape, a center-of-mass computation method that preserves translation invariance, was developed [4]. K-shape maintains a constant time series while calculating the Shape-Based Distance (SBD) by translating the other time series to determine how similar two time series at different time offsets are. The clustering center is determined to be the time series that is most similar to all the sequences. According to experiments, k-shape is less time-consuming than more sophisticated methods like DTW while remaining competitive.
Feature-based methods. Feature-based methods use conventional clustering methods to complete the clustering task while transforming the original features. Similarity Preserving Representation Learning (SPIRAL), for instance, was proposed by Lei et al. [5] and demonstrates that a similarity matrix may accurately mimic a low-rank matrix. Li et al. [6] proposed a novel clustering method named Symbolic Pattern Forest (SPF), which differs from conventional feature-based approaches. SPF corrects the Symbolic Pattern Tree (SPT) in a global viewpoint by biasing to cluster time series data with linear complexity, while SPT uses randomly chosen patterns to classify time series data in order to capture temporal properties in a local perspective. In conclusion, it is easy to see that scholars concentrate on cutting down on time. However, the features we extract play a major role in the clustering performance of time series data, and we might not obtain satisfactory results if we are unable to assess the time series-specific patterns and capture the deeper feature dependencies. Thus, time series data application to neural networks for deep learning has received increased attention.
Deep learning-based methods. The broad application of neural networks in the analysis of time-series data has led to the gradual emergence of a fundamental problem: how to successfully maintain the time series’ essential characteristics while training the model [7]. According to Guo et al. [7], clustering loss may cause the data’s feature space to be perturbed, producing meaningless feature representations and lowering clustering accuracy. As a result, the Improved Deep Embedded Clustering (IDEC) method was proposed. Experiments showed that the method’s performance is significantly enhanced by the introduction of a denoising encoder, which guarantees that the data with its original distribution during clustering training is only adjusted by clustering loss. For many years, there has been a lot of interest in applying end-to-end modeling to time series clustering. This allows for the comprehensive collection of more effective features in time series data by combining feature extraction with model training. In order to capture spatial and temporal feature relationships, Madiraju et al. [8] proposed Deep Temporal Clustering (DTC), which uses a bi-directional LSTM (BiLSTM) and a one-dimensional Convolutional Neural Network (1D CNN) as encoder layers. Furthermore, Minirocket [9] was proposed by Dempster et al. by replacing the random initialization of convolutional weights with specified rules, which significantly increases the model’s efficiency. Motivated by ROCKET [10], Jorge et al. [11] suggested R-Clustering as a mathematical computation to model the convolution process. This minimizes the time consumption of deep clustering and applies the convolution structure with random weights to time series for the first time. RandomNet is a further implementation of the concept of random weights [12]. RandomNet clusters separately using a combination of multi-branch CNN and LSTM modules and then integrates the multi-branch clustering using Hybrid Bipartite Graph Formulation (HBGF) [13] to generate the final clustering labels.
A growing number of methods based on the Mamba have been proposed in recent years [14], including clustering models [15,16], detection models [17,18] based on image and audio data, and prediction models based on time series data. The potential of the Mamba architecture in distant sensing change detection applications was initially examined by Chen et al. [17]. Additionally, they created a number of decoder designs for a variety of applications, like change detection and semantic recognition, based on the visual Mamba architecture. Zhang et al. [18] quickly learned features from many channels and applied frequency domain features to filter high-frequency features while successfully detecting tiny cracks using a cross-visual Mamba feature extraction module. The potential and effectiveness of Mamba in capturing global context in various task modes are essentially feedback from all of these methods. It is worthwhile to consider how to more effectively apply Mamba’s parallel operation capacity for the time series clustering work, as the clustering task for time series data is not observed. Furthermore, we discovered that the aforementioned CNN-based clustering methods concentrate on the relationship between samples and capture the spatial dependency of time series features, whereas other modules capture the temporal dependency. This separated learning approach frequently fails to learn temporal dependency effectively and ignores spatio-temporal concordance and dynamics. To solve these issues, we first investigate the possibilities of Mamba in time-series clustering tasks. Taking into account the dynamics and non-smoothness of spatio-temporal graph (STG) data [19], we propose PG-Mamba. By transforming time series into patch-based spatio-temporal graph structures and employing Mamba-driven state-space models to learn hierarchical representations of time series for clustering, the framework avoids concentrating on the links between time series.

Contributions of the Paper

Unlike conventional learning representation methods, we focus on the characteristic patterns of time series rather than the correlation between samples. Specifically, each time series is divided into multiple patches, which are then transformed into a two-dimensional feature matrix. In each graph, a patch is regarded as a graph node, with its corresponding vector representing the feature of that node. To synergistically capture the temporal and spatial characteristics of time series, spatial edges are designed using the K-nearest neighbor algorithm, while adjacent patches serve as temporal edges, thus transforming the time series into an STG. For the first time, we explore spatiotemporal patterns within a single sequence, addressing the issue that traditional methods fail to capture non-adjacent similar patterns.
We designed a module for representation learning. The module leverages Mamba to capture long-range dependencies across patches and applies these global dependencies to a decoupled STG attention module. This approach facilitates the undisturbed learning of local evolution and repetitive patterns in time series data. Finally, we introduce a ternary feature fusion mechanism that integrates the global relationships extracted by Mamba, the local evolution captured by temporal graph attention, and the global repetitive patterns identified by spatial attention. The fused features are then mapped to obtain the final latent representation.
In addition, to bolster the framework’s resilience to noise in time series data, we introduce noise and generate augmented data versions using random masks. A contrastive loss function is then employed to enhance the framework’s denoising capabilities.
To mitigate the interference of clustering loss with the time-series feature space, we propose integrating adjacency matrix reconstruction loss into the clustering process, thereby achieving dual graph reconstruction self-supervision. The performance of PG-Mamba was evaluated on the UCR time-series dataset [20] across 11 clustering algorithms. Experimental results demonstrate that PG-Mamba achieves optimal performance.

2. Materials and Methods

PG-Mamba is a deep clustering framework that builds time series as STG while paying close attention to spatio-temporal dynamics and concordance, thereby changing the conventional approach of viewing time series as a one-dimensional flow and pioneering the concept of a single-sequence internal STG. In the meantime, the multi-level Mamba modules that have been built enable shared pattern learning across patches. Figure 1 shows the overall structure of the framework, which includes three modules: (a) The STG building module, (b) feature representation learning module, and (c) clustering module. Among these, Figure 2 specifically displays the MCBlock module in (b).

2.1. Datasets

We evaluated PG-Mamba’s performance using 33 datasets from the popular UCR time series archive, which consist of 13 data types, including device, image, sensor, simulated, spectro, traffic, motion, ECG, EOG, EPG, HAM, hemodynamics, and spectrum. Each dataset contains a minimum of 17 and a maximum of 3601 samples in the train set, and a minimum of 28 and a maximum of 4000 samples in the test set. Furthermore, the time series length varies from 15 to 2844, and the number of time series classes ranges from 2 to 52. Specific information for these datasets is presented in Table 1.

2.2. The STG Building Module

We denote the time series as X   =   { x 1 , x 2 ,   , x n } , where n is the number of time series samples. Following Zhang et al. [21] and Abdel-Sater et al. [22], we divide the time series into patches and divide each item into a two-dimensional feature matrix in order to completely capture the time series’ deeper trends. Zhang et al.’s method treats the time series as essentially a one-dimensional sequence, merely preserving local information through segmentation. Abdel-Sater et al. reshape a time step of length L into multiple patches to satisfy the model’s input requirement. In contrast, our approach aims to transform the time-series data from a one-dimensional sequence into a graph structure. Using x 1 as an example, we divide x 1   =   { x 1 i } i = 1 t into X 1 , where t is the feature dimension. In this process, a sliding window strategy is adopted. Specifically, unfold is applied with a window size d and stride s in the feature dimension. As a result, m   =   ( t d ) / s   +   1 patches are extracted. Therefore, X 1 is expressed as follows:
X 1 = x 11 x 1 d x m 1 x md ,
where d is the dimension of each patch and m is the number of patches.
We regard the patches as nodes and the related vectors as the node characteristics, represented by V =   v p p   =   1 , ,   m ,   v p     R d , when building an STG G   =   ( V ,   E ) based on each time series. To concentrate on both temporal and spatial aspects of the time series, we build temporal and spatial edges when building edges between nodes. In particular, we created the temporal edge, represented by E t   =   { ( v p , v p + ) |     { - w ,   w } } , after defining the parameter w and extracting each patch’s neighboring patches during the closed interval [−w, w]. Additionally, spatial edges were constructed by applying k-nearest neighbors (k = k) to connect each patch with its k most spatially correlated patches, represented by E sp   =   p     P   p ,   q   q     N k p ,   p   =   1 , ,   m , where N k p is the set of the node p’s k nearest neighbor nodes. The temporal and spatial edges are combined and de-duplicated to produce the final set of edges E. As a result, every graph has node collection V and a maximum of 24   +   ( m - 6 )   ×   ( 2 w ) temporal edges and m   × k spatial edges. It is important to note that the connection of boundary nodes requires special consideration when constructing the temporal adjacency graph. Specifically, the first three nodes in the sequence have fewer adjacent nodes than others. For instance, the first node has only three subsequent adjacent nodes; therefore, the number of its temporal edges is 3. Similarly, the second and third nodes have temporal edges of 4 and 5, respectively. The same applies to the last three nodes in the sequence. Consequently, the sum of temporal edges for the first and last three nodes is 24, while 2 w temporal edges exist for all the remaining nodes. In this way, the concept of a single-sequence internal STG is first introduced.

2.3. Feature Representation Learning Module

In this section, we introduce MGAT, a feature representation learning module that enables both local-to-global generalized pattern learning and shared pattern learning over STG patches. First, a brief introduction to State Space Models (SSMs) is given because we use them in this section. SSMs are a framework for modeling the link between input-output sequences, while the updating process’s coefficient matrix is discretized to allow for deeper data analysis. The following ordinary differential equation [23] represents this:
h t   =   A ~ t h t 1 + B ~ t x t , y t   =   C t h t + D x t ,
where A ~ t is a discretized representation of the state matrix A, B ~ t is a discretized representation of the input projection matrix B, and Δ is a time-step discretization factor. The specific representations are as follows:
A ~ t   =   e Δ A , B ~ t   =   ( Δ A ) 1 ( e Δ A   I ) Δ B .
This selective scanning mechanism is also seen in Mamba, which focuses on dynamic learning. In particular, Mamba’s Δ , B, and C are separated following linear mapping of the input data. Δ and B can dynamically adjust the key information of the input data, and C can determine which aspects of the state are important to the output. Every training process concentrates more on the initial data patterns and employs real-time hidden states to fully understand the contextual relationships. The model’s comprehension of dependencies is enhanced by this method of selective focus, memorization, and application of contextual information. Simultaneously, sequence blocks’ and feature blocks’ parallel processing method significantly increases computing throughput and is a reliable sequence model.
The Mamba model principle is now fully realized, and we go back to our feature representation learning module. Before entering the MCBlock module, a STG feature first goes through a linear layer that changes the feature dimension d to the hidden dimension D, converting X     R n × m × d to X     R n × m × D . It then enters a multi-layer MCBlock, where the operations are specifically represented as follows for each layer:
X   =   ReLU ( MCBlock ( X ) ) .
The graph node features are initially put through the normalization process when they arrive at the MCBlock module. It is important to note that we carry out an independent normalizing operation on the patch features, which maintains the distributional disparities between the patches, in contrast to the prior normalization process. The following is the particular representation:
X b , d , m   =   X b , d , m     μ b , d σ b , d , μ b , d   =   1 M m = 1 M X b , d , m , σ b , d   =   1 M m = 1 M ( X b , d , m     μ b , d ) 2   +   ε   ,
where ε is a hyperparameter that guarantees stable values, we followed the common practice for normalization layers in deep learning and set the value to 1 × 10−5. Additionally, we perform a radial transformation to the normalized feature representation, which can be represented as follows, to improve the framework’s expressive power:
X b , d , m   =   γ m · X b , d , m   +   β m ,
where γ m and β m are the learnable weights and biases, respectively.
After normalizing the node data, we may divide it into convolution modules and the Mamba module. To begin, we use the convolution module to capture local patterns within patch interiors. Equation (7) provides the specifications for the following convolution module:
X =   Dropout ( LeakyReLU ( BatchNorm 2 d ( Conv 2 d ( X ) ) ) ) .
Concurrently, the STG is fed into the Mamba module to capture long-range dependencies across patches. Subsequently, the node representations from convolution modules and the Mamba module are concatenated. This design is motivated by two key aspects: (1) the Mamba module focuses on learning higher-order node representations, intentionally abstracting away localized fine-grained details at this stage; (2) conversely, convolution modules preserve these detailed features. The concatenation of these representations thus enriches feature diversity by combining complementary information.
Based on the learned global features, spatio-temporal decoupling within a single sequence is performed. Specifically, GATConv is applied to the temporal and spatial edges of the STG, denoted as GATConv_T and GATConv_S, respectively. GATConv_T captures changes in adjacent patches, representing a short-term pattern, while GATConv_S captures similar, non-adjacent patches, representing a long-term pattern. This spatio-temporal decoupling architecture allows for a greater focus on local evolution or repetitive patterns, enabling the capture of key information within the time series.
Meanwhile, we propose a ternary feature fusion mechanism that integrates the features extracted by MCBlock with those obtained after spatio-temporal edge learning. This fusion mechanism enables more effective learning of global feature dependencies, local-to-global dynamics, and pattern repetition consistency. Finally, the ternary features are processed by a multilayer perceptron (MLP), followed by sample-wise global pooling to obtain the latent representation of the time series.

2.4. Feature Representation Learning Pre-Training

To mitigate the effects of endogenous noise in time series, data augmentation is performed by injecting adversarial noise. Specifically, we generate a Bernoulli-distributed mask vector m m   ~   Bernoulli ( p   =   mask _ ratio ) , where each element independently indicates patch selection with a fixed probability mask _ ratio . The input data X 1     R m × d is then randomly masked and Gaussian noise is injected, yielding the augmented time series X ~ 1 :
X ~ 1   =   X 1 ( mask )   +   ε m , d , X 1 ( mask )   =   ( 1     m m ) X 1 , ε i , j   ~   N ( 0 , σ 2 ) ,   i     [ 1 , m ] ,   j     [ 1 , d ] ,
where σ denotes the variance and is a predefined hyperparameter. Given that we have standardized the data, resulting in a standard deviation of approximately 1, the morning scale is set to σ = 0.1 .
The pre-training objective aligns original samples with their augmented counterparts, equipping the framework with discriminative capability against unrelated samples and robustness to noise. Accordingly, we formulate the contrastive loss function (Equation (9)). During pre-training, we further implement early stopping to ensure computational efficiency.
L c = 1 2 N i = 1 N [ log e Z 1 i · Z 2 i / τ j = 1 N Z 1 i · Z 2 j / τ + log e Z 2 i · Z 1 i / τ j = 1 N Z 2 i · Z 1 j / τ ] ,
where N denotes the number of samples, Z 1 denotes the feature representation learned from the source data, Z 2 denotes the feature representation learned from the augmented data, and τ     R + denotes the temperature hyperparameters. We choose τ   =   0.1 to match the standard deviation of the injected noise, thereby maintaining robustness to augmented data.

2.5. Cluster Module

The clustering process lies in enhancing the intra-cluster closeness and widening the inter-cluster distance of samples with similar representation patterns. In each training process, the probability distribution q of the potential representation belonging to each cluster center is calculated. And q is normalized to obtain p . KL scattering is applied to quantify the difference between the two probabilities, respectively, pushing q towards a sharper probability distribution. The clustering loss function is shown in Equation (10).
L cl = 1 N i = 1 N D KL ( p i | | q i ) .
In addition, inspired by Guo et al. [7], in order to avoid the possible perturbation of the feature space caused by the clustering loss, we propose a new loss function based on the STG, i.e., the loss of neighbor matrix reconstruction, as shown in Equation (11). Based on the potential representation of the framework output, the source and target node features are concatenated, and the reconstruction of the STG adjacency matrix is realized by MLP mapping to a single feature.
L r = 1 N ( A A recon ) 2 ,
where N denotes the number of samples and A is the original adjacency matrix.
The loss function of the final framework is represented as follows:
L   =   L cl   +   L r .

2.6. Cluster Validity Indices

We used two clustering performance metrics, the Rand Index (RI) and the Adjusted Rand Index (ARI). The RI computes the agreement between predicted cluster assignments and ground-truth cluster assignments by comparing the co-membership of the sample pairs. The ARI [24] extends this by correcting for chance agreement through expectation normalization, using the expectation of a random assignment of the RI to make the statistic expected to have a value of 0.

3. Results

3.1. Training Setting

PG-Mamba makes full use of the temporal and spatial features of time series and employs contrast loss to align the source data with the augmented data with noise in the pre-training process to train the denoising stability of the representation learning module. We adopt the AdamW optimizer and set both the learning rate and weight decay to 1 × 10−4. In addition, an early stopping mechanism is set to enhance training efficiency. In the clustering module, we adopted the Adam optimizer and set the learning rate to 1 × 10−4. The clustering is dynamically corrected by the fusion of the KL scatter loss and the reconstruction loss, which maintains the distribution of the original time series during the clustering process and maximizes the advantages of combining the two types of losses. PG-Mamba is trained and optimized on a GPU (Tesla V100S-PCIE32GB) with 3.6 M parameters. For the other method implementations, we follow their respective architectures and hyperparameter values.

3.2. Experimental Results

To evaluate the ability of PG-Mamba in time-series clustering, we compared it with 11 more advanced methods. The specific details of each method are listed below:
  • K-Means [1] accomplishes the clustering task by assigning samples to the nearest clusters by calculating the Euclidean distance and continuously updating the cluster centers.
  • DBA [2] flexibly aligns the time series to the average series based on DTW path backtracking by calculating the DTW distance from each time series to the average series.
  • KSC [3] takes into account that the time series should still maintain the same shape when the time axis is panned and captures the evolutionary pattern of the time-series data through different scaling methods when clustering.
  • K-shape [4] aligns the nearest clusters by SBD values and uses Rayleigh Quotient maximization to find the center of the representation clusters that have the greatest similarity to all time series.
  • SPIRAL [5] obtains the similarity matrix based on the DTW distance of paired time-series data and solves it to obtain new features to accomplish the clustering task using K-Means [1].
  • SPF [6] converts the timing data into symbol patterns after time window division and randomly selects the symbol patterns with SPT, which together form the SPF to correct the deviation of SPT in a global view.
  • IDEC [7] fine-tunes the clustering loss by pre-training a denoising self-encoder to preserve the distribution of the data.
  • DTC [8] uses CNN and BiLSTM as encoder layers to capture spatial and temporal feature dependencies independently.
  • Minirocket [10] determines the initialization of the weighted convolution kernel by means of predefined rules to simulate the convolution operation in a linear computation to achieve the classification task.
  • R-Clustering [11] simulates inflated convolution to construct a multi-feature representation of each sample by combining inflated values with a convolution kernel. The learned diversity features are dimensionalized by PCA, and clustering is accomplished using K-Means [1].
  • Randomnet [12] utilizes a combined module of multi-branch CNN and LSTM to independently extract temporal and spatial features of time-series data and integrates them through multi-branch clustering to obtain the final clustering labels.
We compared the above 11 methods with PG-Mamba based on RI metrics, and the experimental results are shown in Table 2. We bolded the best results and calculated the average RI, the average ranking, and the number of best RIs for each method. If the average ranking has duplicate values, we adopt the averaging rule. Simply put, when there are three values that are all the same and should be ranked from two to four, all three values are standardized to three, and so on. The experimental results show that PG-Mamba achieves the highest average RI (0.736) and average ranking (3.606). It secured first place in RI on 13 datasets—outperforming the second-best method, which attained only three first-place RI rankings and an average ranking of 4.379. Notably, PG-Mamba exhibits more prominent performance on long sequence datasets. It also demonstrates a distinct advantage in high-noise datasets such as AllGestureWiimoteX and EOGHorizontalSignal. Further experiments were conducted on datasets with sequence lengths greater than 500, revealing that the average RI reached 0.781, exceeding the overall 0.736, and the average ranking was 2.71, surpassing the second-best score of 4.03.
In addition, to mitigate dataset-specific randomness and category bias, we categorized all data into 10 distinct time series types, and the experimental results are shown in Table 3. It reports the average RI rankings per category (lower values indicate superior clustering), along with each method’s overall average ranking and first-place count. Addi-violin plots for 9 categories (Figure 3) visualize RI ranking distributions, assessing the impact of time series type on method performance. Specifically, because the distribution cannot be effectively represented when a category contains only one dataset, we excluded it from Figure 3. PG-Mamba demonstrates statistically significant superiority across time series dataset types, achieving the highest average RI ranking and the most first-place RI results among all baselines. This robust performance confirms that PG-Mamba’s STG representation learning effectively captures critical features.
Meanwhile, in order to facilitate a clear comparison of whether the various algorithms have significant differences on the full dataset, we plotted a key difference plot [25], as shown in Figure 4. The key difference plot calculates the global algorithmic difference by the Friedman test and calculates the p-value by combining the Wilcoxon test of two-by-two algorithms in order to assess the statistical significance of each pair of algorithms on the RI metrics. The horizontal line in the figure is used to connect algorithms without significant differences, and the value is the average ranking of the algorithm across all datasets. Results confirm PG-Mamba’s unique dominance (avg. rank:3.6515).
While the RI-based experimental results demonstrate the advantages of PG-Mamba, we also consider the effect of random assignment on chance consistency. We selected two newer clustering methods, R-Clustering [11] and RandomNet [12], eight datasets were randomly selected for plotting circular spider web diagrams based on the ARI, as shown in Figure 5. The figure shows that PG-Mamba never displays an ARI value less than 0. Notably, these datasets are all from different data types, indicating that PG-Mamba’s advantage lies in its ability to capture the data structure rather than random assignment. Additionally, Figure 6 demonstrates that the three algorithms differ significantly and that PG-Mamba performs better than the other two, with an average ARI ranking of 1.9375. In conclusion, PG-Mamba learns the important patterns in the time series and achieves global modeling while maintaining details. This results in a fresh perspective for the time series clustering task.

3.3. Analyzing Parameter Sensitivity

The clustering procedure optimizes cluster separation through the joint minimization of intra-cluster distances and maximization of inter-cluster distances. A robust clustering framework exhibits minimal performance variance under parameter perturbations while maintaining consistent clustering quality. To validate the robustness of the proposed framework, we conducted parameter sensitivity analyses across three heterogeneous time series datasets. Given the implementation of early-stopping mechanisms during training, epoch sensitivity was excluded from consideration. Figure 7 presents the RI metric fluctuations of PG-Mamba under varying batch sizes on the MedicalImages dataset and GunPointAgeSpan dataset, demonstrating its operational stability.
Also, the more important hyperparameters in our framework include patch and stride, we perform a sensitivity analysis on the remaining two datasets as shown in Figure 8. As illustrated in the figure, the MedicalImages dataset (top) and GunPointAgeSpan dataset (bottom), differing by several times in the number of features, were analyzed. With stride values ranging from two to six and patch values from two to four, almost no significant changes in the results were observed. Hyperparameter stability is quantitatively measured using the relative standard deviation (RSD = σ/μ). After traversing all combinations of patch     { 2 , 4 , 6 } and stride     { 2 , 4 , 6 , 8 , 10 , 12 , 18 , 20 } , the RSD for the MedicalImages dataset is only 0.03%, while it is 2.6% for the GunPointAgeSpan dataset. We also observed that increasing the stride from 2 to 20 results in a decrease of only 1.4% in the RI of MedicalImages, whereas the RI of GunPointAgeSpan decreases by as much as 10%. Notably, the RI of GunPointAgeSpan does not exhibit significant fluctuations when the stride is less than or equal to eight. Considering the characteristics of the dataset, the GunPointAgeSpan dataset contains a large amount of sensor noise and domain shift redundancy, and an excessively large stride will skip key information, leading to adverse effects.
Evidently, the inherent temporal redundancy of different datasets directly determines their sensitivity to the same hyperparameters. In addition, a sensitivity analysis was conducted on the hyperparameters k and w used for STG construction, as detailed in Figure 9. Specifically, w was varied from 4 to 6, and k values of 4 and 6 were tested. After ten experimental validations, the results for the SmoothSubspace dataset (top) and UMD dataset (bottom) were obtained. Ultimately, k = 4 and w = 6 were selected as the parameter values for our framework. In summary, parameter sensitivity analysis allows us to derive the mapping rules between data and parameters, thus offering a valuable reference for minimizing parameter tuning costs across various datasets.

3.4. Analyzing the Time Complexity

In real-world applications, the impact of the dataset sample size on clustering performance is non-negligible. To verify the scalability of the framework, a detailed analysis was conducted on the TwoPatterns dataset, as shown in Figure 10. The number of time series was selected, ranging from 1200 to 5000, to analyze the time complexity and its impact on the RI. As the figure illustrates, our framework maintains a relatively stable RI value without significant fluctuations. Furthermore, the time consumption, when fitted against the timeline, demonstrates a linear time complexity.

3.5. Ablation Study

To verify the effectiveness of each component in PG-Mamba, we compared PG-Mamba with three variants on 33 datasets, evaluating them based on the RI metric and the number of floating-point operations (FLOPs), as detailed in Table 4. PG-Mamba w/o Conv signifies the variant with convolution removed, PG-Mamba w/o Mamba signifies the variant with Mamba removed, and PG-Mamba without Conv and Mamba signifies the variant with both removed. As demonstrated in Table 4, the full PG-Mamba outperforms all three variants in terms of the Average Rand Index and exhibits the highest return per unit of computation.

4. Discussion

Reviewing the existing clustering methods, more attention is paid to finding the relationship between two time series to accomplish the clustering task. However, the inherent intrinsic noise as well as the low-dimensional nature of time series tend to have redundant data and limited information. Therefore, when faced with massive amounts of time series data, only focusing on the relationship between the two may not be able to obtain better clustering results. This study provides another way of thinking, where we focus on the time series itself, aiming to find the key patterns within it, and clustering the time series data based on the key patterns carried by the time series to achieve a better performance.
To comprehensively capture essential temporal patterns and leverage the full feature potential of time series data, we propose PG-Mamba—a novel framework that advances beyond conventional independent feature learning paradigms. Our method pioneers a collaborative spatio-temporal feature fusion scheme, introducing the concept of single-sequence spatio-temporal patterns. Instead of simply treating a time series as a one-dimensional flow, we decouple it into temporal evolution and pattern space through carefully designed temporal and spatial edges. This allows us to construct a spatio-temporal graph structure, enabling dynamic modeling of spatio-temporal consistency. At the core of our architecture lies Mamba-driven long-range dependency learning and decoupled spatio-temporal graph attention. Mamba captures global patterns within the spatio-temporal graph (STG), while spatio-temporal graph attention focuses on key pattern features. Furthermore, a ternary feature fusion mechanism effectively preserves critical information. This research marks the first exploration into the effectiveness of a hybrid Mamba and spatio-temporal GAT architecture for time series clustering tasks.
This framework employs adversarial noise injection to complete the pre-training process, thereby enhancing the framework’s denoising capability. To address potential distribution perturbations caused by the clustering objective, a dual-loss optimization strategy is adopted: (1) KL divergence is utilized for clustering alignment; (2) a novel neighbor matrix reconstruction loss is introduced to maintain the integrity of the spatio-temporal structure. This comprehensive approach ensures both clustering performance and framework stability, while preserving the integrity of the data distribution. We verified that PG-Mamba is not randomized for a particular data type by comparing the average rankings of different time series types. In order to exclude the inflated RI caused by randomness, we also compared ARI with better algorithms on all datasets, and the experimental results showed that PG-Mamba achieved the highest average ranking and average value. Overall, PG-Mamba focuses on the time series itself, learns the key patterns, and fills the gap where the Mamba model is not applied to clustering tasks. This exploration provides a new way of thinking about the time series clustering task.
However, our framework still has some limitations: (1) We are working with manually set hyperparameters, and although it has been shown that PG-Mamba has only small fluctuations for hyperparameter variations, in the future we will find a way to automatically extract hyperparameters that are suitable for a specific type of time series and adaptively accomplish the clustering task. (2) Inspired by the “in-cluster sharing” fine-tuning strategy for edge devices proposed by Abdel-Sater et al. [22], we recognize that techniques such as low-rank fine-tuning can be adopted to reduce the framework’s tuning cost. Therefore, our future work will focus on designing fine-tuning strategies tailored for graph neural networks or time series frameworks to balance clustering effectiveness and framework performance.

5. Conclusions

This study introduces PG-Mamba, a deep learning framework specifically designed for time series clustering. In contrast to traditional time feature learning methods, the framework transforms time series into patch-based STG structures and employs Mamba-driven state space models to capture global dependencies. Concurrently, a spatio-temporal GAT module is utilized to decouple spatio-temporal features, ultimately achieving clustering through multivariate feature fusion. To validate PG-Mamba’s performance, 33 datasets were selected from the UCR time series archive and compared against 11 state-of-the-art methods, with RI and ARI used as evaluation metrics. The experimental results demonstrate PG-Mamba’s superior performance in time series clustering tasks. While this study has achieved promising results, we acknowledge that the manual setting of hyperparameters can influence the framework’s performance. In the future, we will focus on designing a method that can automatically learn hyperparameters based on time series characteristics to achieve better clustering performance.

Author Contributions

Conceptualization, Y.S.; methodology, Y.S.; data preprocessing, Y.S. and D.Z.; writing—original draft, Y.S.; editing, Y.S.; review, J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the following grants: Inner Mongolia Autonomous Region Science and Technology Major Project-Research and Development of Cloud Computing Application Technology (2019ZD016) and Inner Mongolia Autonomous Region Science and Technology Department-Research and Development and Application of Key Technology for Intelligent Control of Potato Industry Chain (2021ZD0005).

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in https://github.com/aniani6aa/PG-Mamba.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Los Angeles, CA, USA, 1 January 1967; University of California Press: Los Angeles, CA, USA, 1967. [Google Scholar]
  2. Petitjean, F.; Ketterlin, A.; Gançarski, P. A global averaging method for dynamic time warping, with applications to clustering. Pattern Recognit. 2011, 44, 678–693. [Google Scholar] [CrossRef]
  3. Yang, J.; Leskovec, J. Patterns of temporal variation in online media. In Proceedings of the fourth ACM international conference on Web search and data mining, Hong Kong, China, 9–12 February 2011. [Google Scholar]
  4. Lines, J.; Bagnall, A. Time series classification with ensembles of elastic distance measures. Data. Min. Knowl. Disc. 2015, 29, 565–592. [Google Scholar] [CrossRef]
  5. Lei, Q.; Yi, J.; Vaculin, R.; Wu, L.; Dhillon, I.S. Similarity preserving representation learning for time series clustering. arXiv 2017, arXiv:1702.03584. [Google Scholar]
  6. Li, X.; Lin, J.; Zhao, L. Linear Time Complexity Time Series Clustering with Symbolic Pattern Forest. In Proceedings of the Nineth International Joint Conference on Artificial Intelligence, Macau, China, 10–16 August 2019. [Google Scholar]
  7. Guo, X.; Gao, L.; Liu, X. Improved deep embedded clustering with local structure preservation. In Proceedings of the 17th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  8. Madiraju, N.S. Deep Temporal Clustering: Fully Unsupervised Learning of Time-Domain Features. Master’s Thesis, Arizona State University, Tempe, AZ, USA, 2018. [Google Scholar]
  9. Dempster, A.; Schmidt, D.F.; Webb, G.I. Minirocket: A very fast (almost) deterministic transform for time series classification. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual Event Singapore, 14–18 August 2021. [Google Scholar]
  10. Dempster, A.; Petitjean, F.; Webb, G.I. ROCKET: Exceptionally fast and accurate time series classification using random convolutional kernels. Data. Min. Knowl. Disc. 2020, 34, 1454–1495. [Google Scholar] [CrossRef]
  11. Jorge, M.B.; Rubén, C. Time series clustering with random convolutional kernels. Data. Min. Knowl. Disc. 2024, 8, 1862–1888. [Google Scholar] [CrossRef]
  12. Li, X.; Xi, W.; Lin, J. Randomnet: Clustering time series using untrained deep neural networks. Data. Min. Knowl. Disc. 2024, 38, 3473–3502. [Google Scholar] [CrossRef]
  13. Fern, X.Z.; Brodley, C.E. Solving cluster ensemble problems by bipartite graph partitioning. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004. [Google Scholar]
  14. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023. [Google Scholar] [CrossRef]
  15. He, Y.; Tu, B.; Liu, B.; Li, J.; Plaza, A. 3DSS-Mamba: 3D-spectral-spatial mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5534216. [Google Scholar] [CrossRef]
  16. Wu, R.; Liu, Y.; Liang, P.; Chang, Q. Ultralight vm-unet: Parallel vision mamba significantly reduces parameters for skin lesion segmentation. arXiv 2024, arXiv:2403.20035. [Google Scholar] [CrossRef]
  17. Chen, H.; Song, J.; Han, C.; Xia, J.; Yokoya, N. Changemamba: Remote sensing change detection with spatio-temporal state space model. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4409720. [Google Scholar]
  18. Zhang, Z.; Peng, B.; Zhao, T. An ultra-lightweight network combining Mamba and frequency-domain feature extraction for pavement tiny-crack segmentation. Expert. Syst. Appl. 2025, 264, 125941. [Google Scholar] [CrossRef]
  19. Li, L.; Wang, H.; Zhang, W.; Coster, A. Stg-mamba: Spatial-temporal graph learning via selective state space model. arXiv 2024, arXiv:2403.12418. [Google Scholar]
  20. Dau, H.A.; Bagnall, A.; Kamgar, K.; Yeh, C.C.M.; Zhu, Y.; Gharghabi, S.; Ratanamahatana, C.A.; Keogh, E. The ucr time series archive. J. Autom. Sin. 2019, 6, 1293–1305. [Google Scholar] [CrossRef]
  21. Zhang, Y.; Ma, L.; Pal, S.; Zhang, Y.; Coates, M. Multi-resolution time-series transformer for long-term forecasting. In Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2–4 May 2024. [Google Scholar]
  22. Abdel-Sater, R.; Ben Hamza, A. A federated large language model for long-term time series forecasting. In Proceedings of the 27th European Conference on Artificial Intelligence, Santiago de Compostela, Spain, 19–24 October 2024. [Google Scholar]
  23. Huang, L.; Chen, Y.; He, X. Spectral-spatial mamba for hyperspectral image classification. arXiv 2024, arXiv:2404.18401. [Google Scholar] [CrossRef]
  24. Milligan, G.W.; Cooper, M.C. A study of the comparability of external criteria for hierarchical cluster analysis. Multivar. Behav. Res. 1986, 21, 441–458. [Google Scholar] [CrossRef] [PubMed]
  25. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. PG-Mamba workflow. (a) The STG building module: splitting the time series into equal-sized patches, using the patches as nodes, and then using the proposed temporal edges—where the blue nodes are located—and spatial edges—where the red nodes are located—to build the STG associated with each time series. (b) Feature representation learning module. To improve the denoising capabilities of the framework, the augmented time series are fed into the MGAT module together with the source data. (c) Clustering module: training the framework to jointly learn spatiotemporal features by calculating the reconstruction loss and clustering loss to obtain a possible representation of the time series.
Figure 1. PG-Mamba workflow. (a) The STG building module: splitting the time series into equal-sized patches, using the patches as nodes, and then using the proposed temporal edges—where the blue nodes are located—and spatial edges—where the red nodes are located—to build the STG associated with each time series. (b) Feature representation learning module. To improve the denoising capabilities of the framework, the augmented time series are fed into the MGAT module together with the source data. (c) Clustering module: training the framework to jointly learn spatiotemporal features by calculating the reconstruction loss and clustering loss to obtain a possible representation of the time series.
Sensors 25 05043 g001
Figure 2. MCBlock module detailed process.
Figure 2. MCBlock module detailed process.
Sensors 25 05043 g002
Figure 3. Violin plots on the average RI ranking metric for nine data types.
Figure 3. Violin plots on the average RI ranking metric for nine data types.
Sensors 25 05043 g003
Figure 4. Critical difference diagram of PG-Mamba with advanced methods based on RI.
Figure 4. Critical difference diagram of PG-Mamba with advanced methods based on RI.
Sensors 25 05043 g004
Figure 5. ARI comparison of PG-Mamba with two methods on eight datasets.
Figure 5. ARI comparison of PG-Mamba with two methods on eight datasets.
Sensors 25 05043 g005
Figure 6. Critical difference diagram of PG-Mamba with two methods based on ARI.
Figure 6. Critical difference diagram of PG-Mamba with two methods based on ARI.
Sensors 25 05043 g006
Figure 7. Variation in RI metrics of PG-Mamba with batch-size.
Figure 7. Variation in RI metrics of PG-Mamba with batch-size.
Sensors 25 05043 g007
Figure 8. Variation in RI metrics of PG-Mamba with different patches and strides on MedicalImages dataset (top) and GunPointAgeSpan dataset (bottom).
Figure 8. Variation in RI metrics of PG-Mamba with different patches and strides on MedicalImages dataset (top) and GunPointAgeSpan dataset (bottom).
Sensors 25 05043 g008
Figure 9. Variation in RI metrics of PG-Mamba with different k and w on SmoothSubspace dataset (top) and UMD dataset (bottom).
Figure 9. Variation in RI metrics of PG-Mamba with different k and w on SmoothSubspace dataset (top) and UMD dataset (bottom).
Sensors 25 05043 g009
Figure 10. Running time of PG-Mamba on the different number of instances.
Figure 10. Running time of PG-Mamba on the different number of instances.
Sensors 25 05043 g010
Table 1. Descriptions of datasets for experiments.
Table 1. Descriptions of datasets for experiments.
Dataset NameData TypeTrain SizeTest SizeLengthClass
AllGestureWiimoteXSensor30070050010
ChinatownTraffic20343242
CoffeeSpectro28282862
CricketYMotion39039030012
DistalPhalanxOutlineCorrectImage600276802
DodgerLoopGameSensor201382882
ECGFiveDaysECG238611362
EOGHorizontalSignalEOG362362125012
EOGVerticalSignalEOG362362125012
EthanolLevelSpectro50450017514
FordASensor360113205002
FungiHRM1818620118
GesturePebbleZ1Sensor1321724556
GunPointMotion501501502
GunPointAgeSpanMotion1353161502
HerringImage64645122
InlineSkateMotion10055018827
InsectEPGSmallTrainEPG172496013
MeatSpectro60604483
MedicalImagesImage3817609910
OliveOilSpectro30305704
PhonemeSensor2141896102439
PigAirwayPressureHemodynamics104208200052
RefrigerationDevicesDevice3753757203
RockSpectrum205028444
ScreenTypeDevice3753757203
SemgHandMovementCh2Spectrum45045015006
SemgHandSubjectCh2Spectrum45045015005
SmoothSubspaceSimulated150150153
UMDSimulated361441503
WineSpectro57542342
WormsMotion181779005
TwoPatternsSimulated100040001284
Table 2. Comparison of PG-Mamba with state-of-the-art methods.
Table 2. Comparison of PG-Mamba with state-of-the-art methods.
DatasetsData Typek-MeansKSCk-ShapeSPFSPIRALKDBAIDECDTCMiniRRandomNetR-ClusterPG-Mamba
AllGestureWiimoteXSensor0.830.0990.8540.8360.8350.8190.8240.8170.820.8340.8120.878
ChinatownTraffic0.5270.5260.5260.6330.7870.5690.5130.5270.5820.5920.5790.489
CoffeeSpectro0.7510.8050.7510.8340.8050.7770.4920.4920.7510.7770.815
CricketYMotion0.8540.5140.8760.8710.8560.8050.8580.8480.8690.8680.8760.884
DistalPhalanxOutlineCorrectImage0.4990.4990.4990.50.5010.5020.5260.5210.5010.5010.5000.499
DodgerLoopGameSensor0.5030.6390.5850.5170.5290.5150.5560.5970.5020.5290.5100.513
ECGFiveDaysECG0.50.8710.8790.5780.530.5110.5060.5960.7610.5310.5120.507
EOGHorizontalSignalEOG0.8570.4220.8770.8690.8660.7970.0820.8550.5690.8740.8570.859
EOGVerticalSignalEOG0.8560.6030.8760.870.8510.80.8550.8380.7860.8610.8400.871
EthanolLevelSpectro0.6230.6210.6230.6260.6060.5530.6130.6890.6210.6270.6730.741
FordASensor0.50.5050.5780.5010.50.50.50.510.50.5010.5030.507
FungiHRM0.9380.7940.7980.990.9930.3980.9590.9260.9990.9761.0000.930
GesturePebbleZ1Sensor0.8020.2130.870.9040.7950.750.8380.8410.8320.8180.7960.826
GunPointMotion0.4970.5070.4970.4970.4980.4970.4980.530.4970.5070.4970.512
GunPointAgeSpanMotion0.6280.5180.530.5140.5460.5180.4990.5590.4990.4990.5190.575
HerringImage0.50.4990.5040.5040.5060.4990.5040.4890.5020.5080.5070.777
InlineSkateMotion0.7360.760.7490.7590.6930.6690.7490.7630.7380.7590.7490.832
InsectEPGSmallTrainEPG0.5640.5740.7070.7320.7220.6290.6280.6350.77510.7680.749
MeatSpectro0.7850.7850.7290.830.8520.80.3280.5780.7680.860.8540.872
MedicalImagesImage0.66500.6680.6670.6680.6840.6820.6510.6780.6770.6740.694
OliveOilSpectro0.7390.8450.7450.8750.8720.8280.2880.2880.7750.8150.8820.816
PhonemeSensor0.9110.4910.9290.930.9280.7890.9220.0820.9280.9320.9290.921
PigAirwayPressureHemodynamics0.9140.0160.9030.9360.960.840.9380.8830.9670.9690.9670.925
RefrigerationDevicesDevice0.5550.3320.5560.5580.5870.5880.5540.5180.5380.580.5780.662
RockSpectrum0.6640.380.6890.7190.6570.6750.660.7220.7340.6950.7470.777
ScreenTypeDevice0.5620.3320.5570.5680.5660.5250.5620.6350.5590.5690.5910.642
SemgHandMovementCh2Spectrum0.7350.6380.7390.7560.6040.7320.7430.7830.7620.7430.7430.822
SemgHandSubjectCh2Spectrum0.7340.6450.7210.7340.5680.6610.730.7970.6980.70.6960.782
SmoothSubspaceSimulated0.7090.3330.6820.6450.8960.6290.5850.6380.6310.8160.6620.741
UMDSimulated0.5570.5590.6120.6220.6260.5570.6140.6820.6160.6210.7620.739
WineSpectro0.4960.4960.4960.50.4950.4960.4960.4960.5030.5020.5070.660
WormsMotion0.6460.520.6560.6830.6660.6440.2630.7010.6480.6870.6550.724
TwoPatternsSimulated0.6280.5370.6750.6930.6560.9450.630.7050.6380.7250.7430.741
Average Rand Index 0.6750.5110.6950.7050.6980.6520.6060.6420.6830.7170.7050.736
Average rank 8.3649.2126.3034.8946.3188.5458.0766.3037.0304.3794.9703.606
Number Best 1141211204313
Best results are highlighted in bold.
Table 3. RI of PG-Mamba vs. state-of-the-art methods on 10 distinct dataset types.
Table 3. RI of PG-Mamba vs. state-of-the-art methods on 10 distinct dataset types.
Datasetsk-MeansKSCk-ShapeSPFSPIRALKDBAIDECDTCMiniRRandomNetR-ClusterPG-Mamba
Sensor8.808.002.403.907.009.606.405.808.105.007.605.40
Device7.7512.008.508.504.506.508.256.509.504.004.001.00
Image10.0011.178.177.335.505.173.338.335.674.005.673.67
Simulated8.8311.337.006.004.337.509.675.338.334.003.002.67
Spectro8.206.608.503.806.907.7010.708.907.703.603.002.40
Motion7.807.406.306.306.7010.108.603.608.605.605.601.40
EOG6.0011.501.003.005.509.509.008.5010.503.007.003.50
EPG12.0011.007.005.006.009.0010.008.002.001.003.004.00
HRM7.0011.0010.004.003.0012.006.009.002.005.001.008.00
Spectrum7.1711.337.004.1711.679.337.172.334.676.505.331.33
Average rank9.0510.306.604.955.908.907.956.457.253.254.052.70
Num. Top-1002000100115
Best results are highlighted in bold.
Table 4. Ablation results of PG-Mamba on 33 UCR datasets.
Table 4. Ablation results of PG-Mamba on 33 UCR datasets.
Model ConfigurationAverage Rand IndexAverage FLOPs (G)
PG-Mamba0.736 (+0.040)49.851 (+6.828)
PG-Mamba w/o Conv0.715 (+0.019)44.751 (+1.728)
PG-Mamba w/o Mamba0.708 (+0.012)49.188 (+6.165)
PG-Mamba w/o Conv and Mamba0.69643.023
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, Y.; Zuo, D.; Gao, J. PG-Mamba: An Enhanced Graph Framework for Mamba-Based Time Series Clustering. Sensors 2025, 25, 5043. https://doi.org/10.3390/s25165043

AMA Style

Sun Y, Zuo D, Gao J. PG-Mamba: An Enhanced Graph Framework for Mamba-Based Time Series Clustering. Sensors. 2025; 25(16):5043. https://doi.org/10.3390/s25165043

Chicago/Turabian Style

Sun, Yao, Dongshi Zuo, and Jing Gao. 2025. "PG-Mamba: An Enhanced Graph Framework for Mamba-Based Time Series Clustering" Sensors 25, no. 16: 5043. https://doi.org/10.3390/s25165043

APA Style

Sun, Y., Zuo, D., & Gao, J. (2025). PG-Mamba: An Enhanced Graph Framework for Mamba-Based Time Series Clustering. Sensors, 25(16), 5043. https://doi.org/10.3390/s25165043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop