Next Article in Journal
Multi-Target Sensing with Interference Analysis for the Multi-UAV-Enabled ISAC Network
Previous Article in Journal
Attention-Driven Time-Domain Convolutional Network for Source Separation of Vocal and Accompaniment
Previous Article in Special Issue
Assessing Chatbot Acceptance in Policyholder’s Assistance Through the Integration of Explainable Machine Learning and Importance–Performance Map Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adapting a Previously Proposed Open-Set Recognition Method for Time-Series Data: A Biometric User Identification Case Study †

by
András Pál Halász
1,*,
Nawar Al Hemeary
1,
Lóránt Szabolcs Daubner
1,
János Juhász
1,2,
Tamás Zsedrovits
1 and
Kálmán Tornai
1
1
Faculty of Information Technology and Bionics, Pázmány Péter Catholic University, 1083 Budapest, Hungary
2
Institute of Medical Microbiology, Semmelweis University, 1089 Budapest, Hungary
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in Halász, A.; Daubner, L.; Al-Hemeary, N.; Juhász, J.; Zsedrovits, T.; Tornai, K. Adapting Open-Set Recognition Method to Various Time-Series Data. In Proceedings of the 19th International Conference on Web Information Systems and Technologies, Rome, Italy, 15–17 November 2023.
Electronics 2025, 14(20), 3983; https://doi.org/10.3390/electronics14203983
Submission received: 28 August 2025 / Revised: 29 September 2025 / Accepted: 9 October 2025 / Published: 11 October 2025

Abstract

Conventional classifiers are generally unable to identify samples from classes absent during the model’s training. However, such samples frequently emerge in real-world scenarios, necessitating the extension of classifier capabilities. Open-Set Recognition (OSR) models are designed to address this challenge. Previously, we developed a robust OSR method that employs generated—“fake”—features to model the space of unknown classes encountered during deployment. Like most OSR models, this method was initially designed for image datasets. However, it is essential to extend OSR techniques to other data types, given their widespread use in practice. In this work, we adapt our model to time-series data while preserving its core efficiency advantage. Thanks to the model’s modular design, only the feature extraction component required modification. We implemented three approaches: a one-dimensional convolutional network for accurate representation, a lightweight method based on predefined statistical features, and a frequency-domain neural network. Further, we evaluated combinations of these methods. Experiments on a biometric time-series dataset, used here as a case study, demonstrate that our model achieves excellent open-set detection and closed-set accuracy. Combining feature extraction strategies yields the best performance, while individual methods offer flexibility: CNNs deliver high accuracy, whereas handcrafted features enable resource-efficient deployment. This adaptability makes the proposed framework suitable for scenarios with varying computational constraints.

1. Introduction

Machine learning has made significant strides in various classification and recognition tasks, often surpassing human performance. For example, the current benchmark is a model that achieves an exceptionally low error rate of just 0.21 % on the MNIST handwritten digit image dataset [1]. Although the field may appear to have overcome all challenges, these accomplishments are primarily confined to closed-set scenarios, where all classes are known during training. However, new classes can appear during real-world applications at the time of testing, necessitating models to make informed rejections in open-set scenarios.
Our previous research tackled this fundamental challenge by introducing an effective Open-Set Recognition (OSR) methodology [2]. The core of our approach was generating synthetic samples from actual data instances to represent the unknown space. A key finding was that training the model to identify and reject these artificial samples significantly enhances its ability to reject genuine unknown samples during testing.
Our approach diverges from traditional methods by generating synthetic features within a hidden layer rather than entirely new inputs. This strategy not only improves accuracy but also reduces computational overhead. The generative model for these features is simpler than the input layer, optimizing computational efficiency. Moreover, placing the synthetic samples in a hidden layer bypasses the initial model segments, saving substantial computational resources. Despite this departure, we still use Generative Adversarial Networks (GANs) [3], including refined and simplified generator and discriminator networks.
One application of OSR is authentication, because, besides recognizing known subjects, the model also needs to reject unknown subjects. The data serving as a base for classification is not necessarily in the form of images. Other types of data can also occur, for example, time series. We are unaware of any OSR model tailored for different data types besides images. Some existing methods appear to be easy to adapt for various kinds of data, but none have been applied to or tested on those. Hence, it is necessary to develop OSR models capable of processing time-series data.
Initially designed for image datasets using convolutional networks, our OSR model is highly adaptable to various data types. This adaptability is due to the model’s modular architecture, specifically the feature extraction module located just before the hidden layer, where synthetic samples are generated. Once the necessary features are extracted, the generative and feature-classifier components work seamlessly together.
In this work, we adapt this model to classify multi-channel time-series data, focusing on biometric signals. We aim to accurately identify users based on the vibrational patterns of their hands, which are captured by accelerometer and gyroscope sensors in mobile phones [4]. Our preliminary results, published in an earlier paper, demonstrate promising outcomes using one-dimensional convolutional networks for feature extraction. The model also retains its favorable time complexity, a significant advantage of its original design [5].
The focus of this work is not on advancing biometric methods per se, but rather on the broader machine learning challenge of Open-Set Recognition (OSR)—enabling classifiers to detect and reject inputs from previously unseen classes reliably. The biometric dataset serves as a case study to demonstrate the model’s adaptation to time-series data, but the methodology itself is domain-independent.
In this work, we conducted a comprehensive experiment, implementing and evaluating several feature extraction methods. We combined these methods into a single model, where the resulting feature vector is the concatenated output of the individual techniques. This approach enhances the model’s performance, albeit at the cost of additional computational resources due to the simultaneous execution of multiple feature-extraction models.
This paper is structured as follows: We commence with an exhaustive literature review, providing a comprehensive background to contextualize our work. Next, we present an overview of the original OSR model. Following this, we discuss in detail how our model was adapted to accommodate the new data type, including data preprocessing and feature extraction methodologies. Finally, we present our results, analyzing the model’s performance and cost-efficiency using various combinations of feature extraction methods.

2. Literature Review

In this section, we provide a succinct overview of the relevant literature. First, we introduce the concept of Open-Set Recognition, followed by an overview of the most relevant existing solutions. Then, we introduce our previous solution to the problem, which serves as the basis for the work described in this article. Lastly, we present the dataset utilized to evaluate our model’s performance.

2.1. Theory of Open-Set Recognition

Many algorithms have long been used to solve classification tasks where only some samples belong to any known class [6,7] or the model lacks sufficient confidence [8,9] in classifying a sample. These models have addressed problems similar to OSR, but without laying down the theoretical background for it. OSR itself was finally formalized by Scheirer et al. [10]:
Let O denote the open space (i.e., the space far from any known data). The open space risk is defined as follows:
R O ( f ) = O f ( x ) d x S O f ( x ) d x
where S O denotes the space containing both the positive training examples and the positively labeled open space, and f is the recognition function, with f ( x ) = 1 if the sample x is recognized as a known class, and f ( x ) = 0 otherwise. Intuitively, R O ( f ) measures the proportion of the function’s support that lies in open space, i.e., the extent to which the classifier incorrectly labels regions far from the training data as known.
Definition 1.
Open-Set Recognition Problem: Let V be the set of training samples, R O the open-space risk, and R ϵ the empirical risk (i.e., the closed-set classification risk, associated with misclassifications). Then, Open-Set Recognition is employed to find an f H , where H is the hypothesis space of measurable recognition functions, such that f ( x ) > 0 indicates assignment to a known class, and f minimizes the open-set risk R O :
arg min f H { R O ( f ) + λ r R ϵ ( f ( V ) ) }
where λ r is a regularization parameter balancing open-space risk and empirical risk.
Definition 2.
The openness of an Open-Set Recognition problem is defined as follows:
O = 1 2   ×   | C T R | | C T A |   +   | C T E |
where C T R ,   C T A , and C T E denote the sets of training, target, and test classes, respectively.

2.2. Existing Approaches

After introducing the concept of Open-Set Recognition, Scheirer et al. [10] proposed the 1-vs-Set Machine as an initial solution. This specialized Support Vector Machine (SVM) model is designed to address open-set challenges. After training, the model incorporates a second hyperplane parallel to the first. Inputs classified between these hyperplanes are labeled as positive. The authors argue that by comparing the volume of a d-dimensional ball to the positively labeled slab inside it, the open-space risk of the model approaches zero as the ball’s radius increases. Despite this approach, the positively labeled space remains unbounded.
In a subsequent study, ref. [11] applied the Radial Basis Function (RBF) kernel to the SVM model. They noted that the radial kernel function K satisfies lim d ( x , x ) K ( x , x ) = 0 , where d(x, x′) represents the distance between feature vectors x and x . The study identified a negative bias term as a necessary and sufficient condition for ensuring a bounded positively labeled open space. This was achieved by adding a regularization term for the bias in the objective function.
Bendale et al. proposed OpenMax [12], a method that adapts a neural network classifier trained initially in a closed-set scenario using a SoftMax layer. OpenMax modifies this setup to allow for the rejection of open-set samples.
Traditional SVMs and SoftMax classifiers are initially tailored for closed-set scenarios, where all classes are known during training. To effectively reject open-set samples, these models must be adapted, which involves rethinking how the input or feature space is divided using hyperplanes or other methods. Such adaptations require fundamentally different approaches to improve the results.
In Open-Set Recognition, the likelihood that a sample belongs to a known class can be estimated through the distribution of its distances to training data in the feature space.
Distance-based classifiers naturally align with the open-set framework. Besides identifying the most similar class for a given input, they yield a similarity score that can be thresholded to decide whether the input should be assigned to that class or rejected as unknown.
Ref. [13] adapted the nearest neighbor method for open-set conditions. To classify a sample s, its nearest neighbor t is located, followed by another neighbor u that belongs to a different class than t. The ratio R = d ( t , s ) / d ( u , s ) is then compared against a threshold T: if R < T , the label of t is assigned to s; otherwise, s is rejected.
Miller et al. [14] proposed an alternative that uses predefined class centroids instead of pairwise sample comparisons. Their method projects inputs into a logit space and evaluates Euclidean distances between the input logit and each class mean for classification.
A central difficulty in open-set learning is the lack of negative (unknown) samples during training, as these appear only at the time of testing. Introducing synthetic data to approximate the distribution of unknowns can help mitigate this limitation.
Generative Adversarial Networks (GANs), introduced by Goodfellow et al. [3], provide a way to produce realistic artificial data. A GAN is composed of two competing models: a generator that maps noise to candidate samples, and a discriminator that attempts to separate real from generated instances.
Kong and Ramanan [15] proposed an OSR-specific GAN framework that conditions the generator on feature vectors to synthesize class-related data, thereby improving rejection of unknowns.
Other works, such as those by Jo et al. and Ge et al. [16,17], also employed GAN-based augmentation strategies for Open-Set Recognition.
Generative modeling has been successfully applied in biometric and physiological signal contexts as well, for tasks such as classification [18] as well as for data augmentation and validation [19,20].
Generating negative samples enhances the performance of an OSR model. However, their substantial additional computational overhead represents a significant drawback.
Most OSR models are primarily designed for processing images, highlighting the need for algorithms working on time-series data. Among the pioneers in this field were Tornai et al. [21], who, following the extraction of statistical features from data collected via smartphones’ sensors, applied the P I S V M [22] and EVM [23] models to time-series data. Their work demonstrated the feasibility of OSR on time series, albeit with room for improvement in the results.
The latter solution was similar to that employed in this work in that it utilized OSR on biometric data for authentication purposes. As Maiorana et al. discussed, another source of such biometric data can be the user’s keystroke dynamics when typing various kinds of texts, e.g., PINs or passwords [24].
Wandji Piugie et al. [25] converted the time-series data of the HAR dataset [26] into images and processed the converted data with the standard convolutional network-based models. We argue that images are not inherently better suited for efficient feature extraction than time series; the reason for having better results on images is that models working on them received significantly more attention. Therefore, developing more time-series-optimized solutions remains a pressing need.

2.3. Previous Work

We previously implemented an OSR method that employs a distance-based approach, diverging from the traditional softmax-based structure to better accommodate open-set scenarios. In this method, training is simplified into a quadratic regression by using fixed class centers. To prepare the model for future unknown inputs, synthetic samples were generated within a hidden feature layer rather than in the input space. The neural network was divided into two segments: the first segment extracts features from samples, with its output serving as the layer where synthetic features are generated. This configuration enabled the training procedure described in Algorithm 1.
Initially, both segments of the model were pre-trained as a unified network. Subsequently, the outputs from the pre-trained first segment were recorded and used as real inputs for training the generative model. The genuine features, along with those produced by the generative model, were then used to train the second segment of the network further. Figure 1 illustrates the overall model architecture.
Algorithm 1 Procedure for training the model. Here, N 1 : feature extractor; N 2 : classifier; N G : generator; N D : discriminator; Y: fixed class centers.
Require:  X = ( x 1 , , x n ) training samples, numbers of iterations n 1 ,   n 2
Ensure:  ( N 1 ,   N 2 ,   Y ) trained networks and fixed class centers
  1: Initialize N 1 ,   N 2 ,   N G ,   N D with random parameters; initialize class centers Y = ( y 1 , , y k )
  2: for  i = 1  to  n 1  do▹Train feature extractor and classifier
  3:    for each batch x j X  do
  4:         o u t N 2 ( N 1 ( x j ) )
  5:         l o s s quadratic loss ( o u t , Y )
  6:        Update N 1 and N 2 with the gradient of l o s s
  7:    end for
  8: end for
  9: f 1 ( X ) = ( N 1 ( x 1 ) , , N 1 ( x n ) ) ▹ Extract features
10: ( N G ,   N D ) G A N ( N G ,   N D , f 1 ( X ) ) ▹ Train the generative model using f 1 ( X ) as real samples
11: z random noise
12: X G N G ( z ) ▹ Generate synthetic features
13: for  i = 1  to  n 2  do▹ Refine classifier with generated samples
14:    for each batch j do
15:         o u t N 2 ( f 1 ( X ) j )
16:         l o s s quadratic loss ( o u t , Y )
17:         o u t N 2 ( ( X G ) j )
18:         l o s s l o s s + quadratic loss ( o u t , Y )
19:        Update N 2 with the gradient of l o s s
20:    end for
21: end for
22: return  ( N 1 ,   N 2 , Y )
The model outperformed most competing methods on commonly used image datasets. For instance, on CIFAR10, it achieved an open-set detection AUC of 0.839 and a closed-set accuracy of 0.914 . Both values were the highest among the evaluated OSR algorithms, with the closed-set accuracy trailing only behind that of highly optimized closed-set classifiers [2].
Like most OSR approaches, however, the original model was tailored for image data, using convolutional architectures optimized for spatial features. Direct application to time series or other modalities is not feasible due to fundamental differences in data structure. This limitation necessitated adapting the method to accommodate sequential data. Fortunately, the modular design of the model significantly simplifies this process: only the feature extraction component requires modification, while the remaining architecture—including the generative module and the distance-based classification mechanism—remains unchanged. This flexibility allows the core principles of the original method to be preserved across different data types.

2.4. Dataset

The primary motivation behind this study is to demonstrate the adaptability of our previously proposed OSR model to time-series data. To this end, we evaluate the method in a user identification setting based on hand-gesture signals, which serves as a biometric case study. A dedicated database tailored to this purpose is currently under development. The software that runs on smartphones and performs the measurements is complete; however, the necessary quantity of data has yet to be collected. Therefore, we conduct our tests on a publicly available dataset, that of Jiokeng et al., who devised a distinct biometric authentication system in a related effort. In their work, the basis of the classification is the subject’s heart signal, detected through the vibration of the hand and measured by a phone held in the hand. The data, when collected in this manner, exhibits a poor signal-to-noise ratio, with the meaningful component being very faint. Still, with extensive preprocessing efforts, the authors achieved high-accuracy results, although only in a closed-set scenario with 112 users [4,27].
The dataset contains measurements from 112 participants (93 male, 19 female), aged between 20 and 60 years. Each user contributed 10 measurement sessions of 30 s each, recorded with a Samsung Galaxy S8 smartphone. This resulted in a total of 1120 raw recordings. The dataset’s challenging properties—weak signals embedded in substantial noise—make it particularly well-suited for testing the robustness and adaptability of OSR approaches.

3. Adapting the Previously Proposed OSR Model to Time-Series Data

As shown in Figure 1 and described above, the model is composed of distinct components. The first part extracts suitable features for training the generative model. The second component classifies these features. Separately, a generative model produces negative inputs.
The latter two components operate on extracted features and are therefore independent of the input data type. As a result, adapting the model to time-series data requires modifying only the feature extraction component.
Three different feature extraction methods were implemented: a convolutional network with 1D convolutional layers reflecting the distinct nature of the data compared to images, predefined statistical features that do not need training, and a lightweight neural network operating in the frequency domain. Their combinations were also tested; multiple of these can be used in the same model, running in parallel. The outputs of the individual methods were then concatenated into the output of the model’s first part. This, along with the individual methods, each with different complexities and levels of performance, gives plenty of room to calibrate the model, optimizing it more for either resources or performance.

3.1. Convolutional Network

Convolutional Neural Networks (CNNs) have demonstrated exceptional performance in handling images due to their ability to process two-dimensional spatial data. When dealing with time-series data, which typically exists in a single dimension (not counting the channels of multivariate time series comparable to the channels of colored images), it is logical to utilize convolutional layers designed for one-dimensional data. Our research identified the architecture illustrated in Figure 2 as the most effective among the various configurations tested. This model comprises five stages, each doubling the number of channels from the previous stage. Each stage contains several 1D convolutional layers activated by the ReLU function, followed by a max pooling layer. The design of our network closely parallels the VGG19 network architecture [28]. Ultimately, the network generates a feature vector matching the size of those produced in our earlier image-based projects [2], ensuring consistency and compatibility with our existing methodologies. We have already implemented this method in [5]. In this work, the method is tested with slightly improved hyperparameters and compared to other methods.

3.2. Predefined Statistical Features

Jiokeng and colleagues [4] demonstrated promising results with a Support Vector Machine (SVM) model that leveraged predefined features. Their findings suggest that these features efficiently encapsulate the essential information needed for the task. Consequently, a similar approach has been adopted in this study, with notable advantages. By employing predefined features, we can bypass the initial phase of model training, requiring only a single extraction of features from each data element.
The selected features include statistical measures such as the mean, median, variance, and standard deviation, alongside other data aggregates like total signal energy, binned entropy, and permutation entropy. Additionally, ref. [4] employed Fourier transform, incorporating all Fourier coefficients and the complete set of auto-correlation and cross-correlation values between signal axes. However, incorporating all these coefficients resulted in their overwhelming dominance in the feature vector, effectively overshadowing the other aggregate features. Our final feature vector excluded these coefficients to maintain a balanced and representative feature set. The exhaustive list of features—covering statistical, entropy-based, and frequency-related descriptors—is provided in the Appendix A.

3.3. Neural Network in Frequency Domain

In the frequency domain, the data can reveal significant insights that might not be immediately apparent. A notable benefit of this representation is that it does not necessitate using complex Convolutional Neural Networks (CNNs). Instead, more straightforward and cost-effective fully connected networks (FCNs) can be utilized. This is because translational invariance—a property where a slight shift in an image or time series does not change the input’s essence—does not apply in the frequency domain. Each value in a specific position retains its unique significance regardless of any shift. Consequently, we employed a relatively small, Fully Connected Network to extract features from the Fast Fourier Transform (FFT) values effectively. Specifically, the FCN consists of three hidden layers of 512 neurons each, with ReLU activations, followed by a linear output layer of 256 neurons. The output of this final layer serves as the feature vector that is passed on to the OSR model.
Recurrent networks such as LSTMs [29] could also be considered for feature extraction from sequential data. They are particularly advantageous for modeling long-term dependencies or handling variable-length sequences. However, in our case, each input window is of fixed length and relatively short, where convolutional or fully connected approaches can already capture the relevant temporal or spectral patterns. LSTMs would therefore introduce additional computational cost without a clear performance benefit. For this reason, we reserve their exploration for future work.

Preprocessing

Each acquisition session was stored in a single file containing measurements from multiple sensors. For our experiments, only the accelerometer and gyroscope data were retained, as these form the basis of the classification task. The raw signals from these two sensors were sampled at irregular timestamps and with non-uniform sampling rates, which required resampling and synchronization. Following the procedure in [4], the data were resampled to a uniform frequency of 200 Hz, aligning the accelerometer and gyroscope streams. Since both sensors operate on three axes, the result was a synchronized six-channel time series.
To focus on the relevant signal components, a fourth-order Butterworth bandpass filter was applied with cutoff frequencies of 0.5 Hz and 30 Hz. The filtered signals were then segmented into windows of a 1.5 s duration with a stride of 0.05 s, which produces a large number of overlapping segments while ensuring sufficient variability for training. Importantly, the division into training and test sets was performed at the session level: two out of ten sessions per subject (20%) were assigned to the test set, and all segments from those sessions were included there. This ensures that overlapping windows never span across training and test sets, preventing information leakage and maintaining strict separation between the two.
For each segment, three different representations were produced and stored together for unified access: the raw six-channel signal (for the CNN-based extractor), the Fourier transform (for the frequency-domain FCN), and the predefined statistical and entropy-based features (for the handcrafted extractor). The Fourier transform values were processed as real and imaginary coefficients, while the handcrafted features are detailed in Appendix A. All representations were normalized to zero mean and unit variance—per channel for the raw and FFT data, and per feature element for the predefined features.

4. Experimental Results

The model has been extensively tested on the dataset described in Section 2.4. We are unaware of any OSR experiments conducted on this dataset by others. Thus, only the different feature extraction methods and their combinations can be compared regarding the open-set detection performance. The closed-set performance (accuracy) can still be compared to that presented in [4]. The relevant hardware specifications were as follows:
  • Intel(R) Core(TM) i7-9800X CPU @ 3.80 GHz;
  • NVIDIA(R) GeForce RTX(TM) 2080 Ti;
  • 128 GB RAM.
For each experiment, the set of classes was randomly divided into known and unknown groups. The training subset of the known classes was used for model training, while the corresponding test subset was used for evaluation. For the unknown classes, only the test samples were included, serving as negative examples during evaluation. The training subsets of the unknown classes were not used at all. This procedure was repeated independently in each run, ensuring that the unknown classes represented truly unseen categories at the time of testing.

4.1. Evaluation Metrics

A suitable and commonly used metric for evaluating OSR methods is the F1 measure, which is capable of capturing both closed-set and open-set performance to some extent. However, this metric is highly sensitive to the model’s calibration—for example, the confidence threshold above which a sample is considered to belong to a known class. Hence, we instead chose the two metrics described below.
Closed-set accuracy: This metric considers only the results for the positive (known) samples. It measures the percentage of correctly classified samples. The model’s prediction is the class with the highest probability, regardless of whether this probability is above a threshold, or the sample would otherwise be classified as unknown. Since all classes in our dataset contribute an equal number of samples, overall accuracy is an appropriate and unbiased measure of closed-set performance. In the OSR setting, it is equally as important to ensure that known samples are not misclassified into other known classes as it is to detect unknowns. Reporting closed-set accuracy therefore complements the open-set detection metrics by indicating how well the model preserves discrimination among the enrolled classes.
AUC: The receiver operating characteristic (ROC) curve is obtained by plotting the true positive rate (sensitivity) against the false positive rate (1—specificity) at each relevant threshold setting. The area under this curve provides a calibration-free measure of open-set detection performance. While calculating it, the known samples are considered positive—regardless of their class—and the unknown samples are considered negative [30].
EER: The Equal Error Rate is the point on the ROC curve where the false acceptance rate (FAR) equals the false rejection rate (FRR). Like the AUC, it is derived from the ROC, but unlike the AUC, it corresponds to a specific operating point. While the AUC is the most widely used metric in OSR research, the EER is a standard measure in biometrics. Including it therefore allows us to position our results in the broader context of biometric evaluation.
The AUC provides a threshold-independent assessment of open-set detection capability, making it particularly suitable for comparing methods across different datasets and configurations. In practical biometric deployments, however, operating thresholds must be chosen to balance the rates of false acceptances and false rejections. This calibration can, for example, be performed by cross-class validation, as demonstrated in our previous work [2], where one testing scenario followed a protocol for outlier detection using the F1 measure. In that setup, the threshold was determined via cross-class validation under the assumption that both types of errors—misclassifying unknowns and rejecting known-class samples—carry equal weight.

4.2. Results

The model was evaluated with several known-class configurations ranging from 10 to 60, using all non-empty subsets of the three feature extraction strategies introduced earlier. Table 1 reports the open-set detection performance. Among the single-method setups, the 1D convolutional network consistently delivers the strongest results, closely followed by predefined statistical features. While the frequency-domain model alone is not competitive as a standalone solution, it provides a measurable boost when combined with other methods. Therefore, it remains a valuable option in scenarios where computational resources are plentiful and performance is the top priority, making the performance improvement worth the added cost. In fact, incorporating multiple feature extraction methods generally leads to clear performance improvements. The number of known classes does not noticeably degrade performance, and the low standard deviation values confirm the robustness of the approach with respect to class selection. To complement the AUC results, Table 2 presents the corresponding EER values. Including the EER allows our results to be related more directly to the biometrics literature, where it is a widely used measure. As expected, the EER values are strongly and almost linearly correlated with the AUC, so they convey largely the same information about model performance. Nonetheless, their inclusion facilitates comparison with prior biometrics studies.
The closed-set accuracy results (Table 3) remain high across all configurations. Although marginally below the upper bound reported in the original HandBCG study [4] in a purely closed-set scenario (98.27% to 99%), our approach maintains strong accuracy while introducing the capability to reject unknown classes—an essential open-set property. The ranking of the methods remains consistent with the AUC results, although the differences in accuracy are smaller.
Table 4 highlights the computational trade-offs between the examined methods during the first training phase. As expected, the convolutional network is the most resource-intensive option. However, the additional cost translates to a solid performance advantage. At the other end of the spectrum, predefined features entirely eliminate the first phase of training. Their extraction cost is negligible (performed once), and the subsequent model is extremely lightweight, making this setup ideal for constrained environments such as embedded systems. Frequency-domain features also offer a resource-efficient alternative, requiring significantly less computation than convolutional networks—though this comes at the price of reduced performance. In all cases, the flexibility of combining feature sets allows practitioners to balance performance and efficiency according to deployment constraints.
Overall, these results demonstrate a favorable performance–cost trade-off across all configurations, underlining the scalability of the proposed approach. A lightweight yet effective model can be built with predefined features for low-resource settings, while maximum accuracy can be achieved through convolutional networks or their combinations with other methods.
A major advantage of our design lies in generating samples within the hidden layer. This strategy almost eliminates the overhead of producing additional training data while retaining the performance benefits of augmentation. The hidden layer’s simpler structure requires only a lightweight generative model, avoiding the need to propagate generated samples through the entire network. Previous work [2] confirmed this as a key efficiency advantage, and our experiments reinforce this observation. On the time-series dataset, the generative model trains in only 5.2 ms per batch—comparable to the image-based results due to identical feature vector sizes. Notably, there are no established GAN architectures specifically designed for time-series data, which makes input-level sample generation not only computationally expensive but also practically challenging. Our approach sidesteps this limitation by working in the feature space, eliminating the need to design or train a specialized generative model for raw sequences. Furthermore, skipping the heavy first stage of the model provides substantial runtime savings: for example, when using the convolutional network as a feature extractor, 91% of the cost of processing a generated batch is eliminated (2.1 ms versus 0.23 ms). This efficiency, combined with strong open-set performance, makes the method not only practical but also uniquely suited for domains where data-level generative models are less feasible.

5. Conclusions

Our exploration of Open-Set Recognition (OSR) in time-series data has unveiled several critical insights with significant implications. OSR is a powerful enhancement of traditional classification techniques, proving particularly relevant in real-world applications. One of the most pressing areas of application is authentication, which serves as a valuable use case for our research.
Authentication systems must not only recognize legitimate users but also effectively identify unknown or unauthorized individuals. This is where OSR becomes indispensable, acting as a robust defense against potential security breaches and unauthorized access. Our research highlights the crucial role OSR can play in enhancing the security and reliability of authentication systems, emphasizing the urgency of integrating OSR into practical applications.
Additionally, our study identifies a significant gap in current research: the need for OSR methods specifically designed for time-series data. Our model demonstrates that it is feasible to adapt OSR techniques to time-series datasets, suggesting that this area will likely attract increasing research attention in the future.
In our experiments, we evaluated various feature extraction methods for OSR. The 1D convolutional network emerged as the most effective, although it incurred the highest computational cost. Predefined features offered a more lightweight alternative, delivering performance only slightly inferior to that of the convolutional network. Combining multiple methods further improved performance but also increased computational demands. This adaptability enables our approach to be tailored to diverse scenarios, allowing users to prioritize either accuracy or, in constrained environments such as embedded systems, optimize the model for computational efficiency while maintaining satisfactory performance.
Looking ahead, several promising research avenues emerge. First, further evaluation on additional biometric modalities, such as voice, ECG, or gait, could demonstrate the generality of our OSR framework beyond the current case study. Sequential architectures like LSTMs also represent a natural extension, as they can explicitly model long-term temporal dependencies; while unnecessary for the fixed-length short segments considered here, they could become valuable for other settings. Similarly, multimodal or lightweight sensor-fusion approaches may increase robustness, especially in mobile or embedded applications. An interesting direction is the application of the adapted OSR framework to continuous monitoring scenarios, where long-term non-stationarity of signals may be addressed.
Another direction concerns the variability of real-world conditions. While the present dataset already contains faint signals embedded in significant noise, making it a challenging benchmark, future work could explore robustness across different acquisition environments and sensor qualities. Adaptation strategies such as transfer learning or online learning may also help maintain performance under varying signal conditions or during long-term monitoring. Finally, threshold optimization and calibration strategies tailored to open-set detection should be studied more explicitly to bridge the gap between research evaluation and deployment in real authentication systems.
In parallel with these extensions, we are developing a new biometric database in which participants are classified based on hand gestures recorded while picking up their phones, using measurements from the device’s sensors. This aims to support a continuous authentication system that operates passively, without requiring any explicit action from the user. Future research may also study how time-dependent changes in the signal—caused by shifts in user behavior or health conditions—affect classification performance.
In summary, our findings underscore the transformative potential of OSR in enhancing authentication systems, particularly when applied to time-series data. As we continue to refine these techniques, OSR is poised to become an increasingly vital component in developing secure and adaptable systems.

Author Contributions

Conceptualization: A.P.H. and K.T.; methodology: A.P.H.; formal analysis and investigation: A.P.H., L.S.D. and N.A.H.; writing—original draft preparation: A.P.H. and K.T.; writing—review and editing: A.P.H., K.T., L.S.D., N.A.H., J.J. and T.Z.; funding acquisition: K.T. and T.Z.; resources: K.T. and T.Z.; supervision: K.T. and T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Research, Development, and Innovation Office through the grant TKP2021-NVA-26.

Data Availability Statement

The dataset used for the experiments was created by Jiokeng et al. [27] and is publicly available at https://doi.org/10.5281/zenodo.5187910.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
OSROpen-Set Recognition
FFTFast Fourier Transform
SVMSupport Vector Machine
CNNConvolutional Neural Network
FCNFully Connected Network
GANGenerative Adversarial Network
RBFRadial Basis Function

Appendix A. Full List of Predefined Features

Unless otherwise noted, each feature is computed independently for all six channels (three-axis accelerometer and three-axis gyroscope). The input is represented as a matrix X R n × 6 , where n is the time dimension.
Table A1. Predefined statistical and spectral features extracted per channel.
Table A1. Predefined statistical and spectral features extracted per channel.
FeatureDefinition or Notes
Time-Domain Statistics
Mean μ = 1 n i = 1 n x i
MedianMedian of { x i } i = 1 n
Standard Deviation σ = 1 n i = 1 n ( x i μ ) 2
Variance σ 2
Skewness 1 n i = 1 n ( x i μ ) 3 1 n i = 1 n ( x i μ ) 2 3 / 2
Kurtosis 1 n i = 1 n ( x i μ ) 4 1 n i = 1 n ( x i μ ) 2 2
Minimum min i x i
Maximum max i x i
Range max i x i min i x i
RMS (Root-Mean Square) 1 n i = 1 n x i 2
Std-to-Range Ratio σ max i x i min i x i
Percentage Above Mean 1 n i = 1 n I ( x i > μ )
Percentage Above Median 1 n i = 1 n I ( x i > median )
Absolute Sum of Changes i = 2 n | x i x i 1 |
Entropy and Energy Features
Total Energy i = 1 n x i 2
Approximate EntropyStandard definition applied per channel
Binned Entropy j p j log p j , with p j = histogram bin probability
Permutation EntropyOrdinal-pattern-based entropy measure
Spectral and Correlation Features
FFT StatisticsComputed on magnitude spectrum; mean, median, variance, and standard deviation
Autocorrelation StatisticsBased on positive-lag coefficients ACF ( k ) = i = 1 n k x i x i + k ; mean, median, variance, and standard deviation
Cross-Correlation StatisticsBetween axis pairs (XY, XZ, YZ): CCF a , b ( k ) = i = 1 n k x a , i x b , i + k ; aggregated mean, median, variance, and standard deviation

References

  1. Wan, L.; Zeiler, M.; Zhang, S.; Cun, Y.L.; Fergus, R. Regularization of Neural Networks using DropConnect. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 17–19 June 2013; Dasgupta, S., McAllester, D., Eds.; Volume 28, pp. 1058–1066. [Google Scholar]
  2. Halász, A.P.; Al Hemeary, N.; Daubner, L.S.; Zsedrovits, T.; Tornai, K. Improving the Performance of Open-Set Recognition with Generated Fake Data. Electronics 2023, 12, 1311. [Google Scholar] [CrossRef]
  3. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 28th International Conference on Neural Information Processing Systems (NIPS’14), Montreal, QC, Canada, 8–13 December 2014; Volume 2, pp. 2672–2680. [Google Scholar]
  4. Jiokeng, K.; Jakllari, G.; Beylot, A.L. I Want to Know Your Hand: Authentication on Commodity Mobile Phones Based on Your Hand’s Vibrations. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022, 6, 1–27. [Google Scholar] [CrossRef]
  5. Halász, A.; Daubner, L.; Al-Hemeary, N.; Juhász, J.; Zsedrovits, T.; Tornai, K. Adapting Open-Set Recognition Method to Various Time-Series Data. In Proceedings of the 19th International Conference on Web Information Systems and Technologies-DMMLACS, Rome, Italy, 15–17 November 2023; INSTICC, SciTePress: Setúbal, Portugal, 2023; pp. 595–601. [Google Scholar] [CrossRef]
  6. Bodesheim, P.; Freytag, A.; Rodner, E.; Denzler, J. Local Novelty Detection in Multi-class Recognition Problems. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 5–9 January 2015. [Google Scholar] [CrossRef]
  7. Nguyen, A.; Yosinski, J.; Clune, J. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 427–436. [Google Scholar] [CrossRef]
  8. Fumera, G.; Roli, F. Support Vector Machines with Embedded Reject Option. In Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar] [CrossRef]
  9. Grandvalet, Y.; Rakotomamonjy, A.; Keshet, J.; Canu, S. Support Vector Machines with a Reject Option. Adv. Neural Inf. Process. Syst. 2008, 21, 537–544. [Google Scholar]
  10. Scheirer, W.J.; Rocha, A.; Sapkota, A.; Boult, T.E. Toward Open Set Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1757–1772. [Google Scholar] [CrossRef] [PubMed]
  11. Júnior, P.; Wainer, J.; Rocha, A. Specialized Support Vector Machines for open-set recognition. arXiv 2016, arXiv:1606.03802. [Google Scholar] [CrossRef]
  12. Bendale, A.; Boult, T. Towards Open Set Deep Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1563–1572. [Google Scholar] [CrossRef]
  13. Júnior, P.; Souza, R.; Werneck, R.; Stein, B.; Pazinato, D.; Almeida, W.; Penatti, O.; Torres, R.; Rocha, A. Nearest neighbors distance ratio open-set classifier. Mach. Learn. 2017, 106, 1–28. [Google Scholar] [CrossRef]
  14. Miller, D.; Sunderhauf, N.; Milford, M.; Dayoub, F. Class Anchor Clustering: A Loss for Distance-Based Open Set Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Virtual, 5–9 January 2021; pp. 3570–3578. [Google Scholar]
  15. Kong, S.; Ramanan, D. OpenGAN: Open-Set Recognition via Open Data Generation. IEEE Trans. Pattern Anal. Mach. Intell. 2025, 47, 3233–3243. [Google Scholar] [CrossRef] [PubMed]
  16. Jo, I.; Kim, J.; Kang, H.; Kim, Y.D.; Choi, S. Open Set Recognition by Regularising Classifier with Fake Data Generated by Generative Adversarial Networks. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2686–2690. [Google Scholar] [CrossRef]
  17. Zongyuan Ge, S.D.; Garnavi, R. Generative OpenMax for Multi-Class Open Set Classification. In Proceedings of the British Machine Vision Conference (BMVC); Tae-Kyun, K., Stefanos Zafeiriou, G.B., Mikolajczyk, K., Eds.; BMVA Press: Durham, UK, 2017; pp. 42.1–42.12. [Google Scholar] [CrossRef]
  18. Alzantot, M.; Garcia, L.; Srivastava, M. PhysioGAN: Training High Fidelity Generative Model for Physiological Sensor Readings. arXiv 2022, arXiv:2204.13597. [Google Scholar]
  19. Li, X.; Metsis, V.; Wang, H.; Ngu, A.H.H. TTS-GAN: A Transformer-based Time-Series Generative Adversarial Network. In Proceedings of the Artificial Intelligence in Medicine: 20th International Conference on Artificial Intelligence in Medicine, AIME 2022, Halifax, NS, Canada, 14–17 June 2022. [Google Scholar]
  20. Laganá, F.; Pellicanò, D.; Arruzzo, M.; Pratticò, D.; Pullano, S.; Fiorillo, A. FEM-Based Modelling and AI-Enhanced Monitoring System for Upper Limb Rehabilitation. Electronics 2025, 14, 2268. [Google Scholar] [CrossRef]
  21. Tornai, K.; Scheirer, W.J. Gesture-based User Identity Verification as an Open Set Problem for Smartphones. In Proceedings of the IAPR International Conference On Biometrics, Crete, Greece, 4–7 June 2019. [Google Scholar]
  22. Jain, L.P.; Scheirer, W.J.; Boult, T.E. Multi-class Open Set Recognition Using Probability of Inclusion. In Proceedings of the ECCV, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  23. Rudd, E.M.; Jain, L.P.; Scheirer, W.J.; Boult, T.E. The Extreme Value Machine. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 762–768. [Google Scholar] [CrossRef] [PubMed]
  24. Maiorana, E.; Kalita, H.; Campisi, P. Mobile keystroke dynamics for biometric recognition: An overview. IET Biom. 2020, 10, 1–23. [Google Scholar] [CrossRef]
  25. Wandji Piugie, Y.; Charrier, C.; Manno, J.; Rosenberger, C. Deep features fusion for user authentication based on human activity. IET Biom. 2023, 12, 222–234. [Google Scholar] [CrossRef]
  26. Reyes-Ortiz Jorge, A.D.G.A.O.L.; Parra, X. Human Activity Recognition Using Smartphones. In UCI Machine Learning Repository; Springer: Berli/Heildelberg, Germany, 2013. [Google Scholar] [CrossRef]
  27. Jiokeng, K.; Jakllari, G.; Beylot, A.-L. Hand-BCG & SCG Signals Dataset; Zenodo: Geneva, Switzerland, 2021. [Google Scholar] [CrossRef]
  28. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  29. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of the model. It remains effectively the same across various types of data–time series as well; only the feature extraction has to be modified to an appropriate model.
Figure 1. Schematic representation of the model. It remains effectively the same across various types of data–time series as well; only the feature extraction has to be modified to an appropriate model.
Electronics 14 03983 g001
Figure 2. The structure of the convolutional network responsible for feature extraction.
Figure 2. The structure of the convolutional network responsible for feature extraction.
Electronics 14 03983 g002
Table 1. Open-set detection AUC scores using different feature extraction methods and their combinations, averaged over five runs. Each configuration was evaluated using a varying number of known classes.
Table 1. Open-set detection AUC scores using different feature extraction methods and their combinations, averaged over five runs. Each configuration was evaluated using a varying number of known classes.
Known Classes 102030405060
Convolutional networkMean0.7500.7720.7970.8110.8280.798
Std0.0480.0450.0540.0830.0650.039
Predefined featuresMean0.7530.7470.7620.7750.7610.752
Std0.0380.0320.0330.0280.0350.040
Frequency domainMean0.7020.6170.7010.7320.6520.688
Std0.0720.1030.0650.0730.0540.059
Conv. + Predef.Mean0.8010.8240.8290.7940.7940.811
Std0.0420.0250.0340.0200.0360.064
Conv. + FFTMean0.7560.7880.7930.8220.7970.825
Std0.0780.0570.0530.0520.0470.060
Predef. + FFTMean0.7700.7880.7790.7530.7440.718
Std0.0790.0540.0630.0490.0570.132
All threeMean0.8230.7940.8040.8250.8420.820
Std0.0430.0450.0370.0410.0350.029
Table 2. Open-set detection performance in terms of EER score, measured from the same runs as in Table 1.
Table 2. Open-set detection performance in terms of EER score, measured from the same runs as in Table 1.
Known Classes 102030405060
Convolutional networkMean0.3020.2930.2780.2520.2170.245
Std0.0400.0330.0470.0410.0360.052
Predefined featuresMean0.3000.3100.2490.2920.2770.295
Std0.0200.0470.0260.0260.0450.023
Frequency domainMean0.3380.3880.3450.2930.3850.349
Std0.0870.0220.0450.0260.0570.013
Conv. + Predef.Mean0.2280.2030.2090.2570.2520.258
Std0.0250.0170.0200.0570.0460.054
Conv. + FFTMean0.3030.2650.2440.1800.2900.198
Std0.0500.0330.0530.0450.0410.057
Predef. + FFTMean0.3080.2500.2490.3080.5880.332
Std0.0530.0380.0310.0350.1090.047
All threeMean0.2250.2770.2690.3400.2010.239
Std0.0420.0280.0530.0540.0170.017
Table 3. Closed-set classification accuracy using different feature extraction setups. All results correspond to the same runs reported in Table 1.
Table 3. Closed-set classification accuracy using different feature extraction setups. All results correspond to the same runs reported in Table 1.
Known Classes 102030405060
Convolutional networkMean0.9500.9430.9060.9100.9240.900
Std0.0270.0280.0250.0320.0240.030
Predefined featuresMean0.9410.9270.9320.9130.9150.890
Std0.0180.0220.0240.0280.0240.028
Frequency domainMean0.8530.8630.7970.7520.7350.711
Std0.0320.0380.0400.0310.0430.045
Conv. + Predef.Mean0.9580.9410.9120.9140.9100.919
Std0.0250.0330.0300.0300.0380.035
Conv. + FFTMean0.9590.9270.9090.9180.8900.900
Std0.0410.0460.0350.0430.0500.037
Predef. + FFTMean0.9310.9190.9270.9230.8830.895
Std0.0330.0360.0360.0420.0400.040
All threeMean0.9550.9500.9220.8940.9220.915
Std0.0290.0410.0380.0290.0290.036
Table 4. Runtime performance using different feature extraction methods. The reported values (in ms) were measured during the first phase of training, except in the case of the predefined features, where the first phase is completely eliminated.
Table 4. Runtime performance using different feature extraction methods. The reported values (in ms) were measured during the first phase of training, except in the case of the predefined features, where the first phase is completely eliminated.
MethodFeature Extraction (ms)Whole Model (ms)Training a Batch (ms)
Convolutional network2.12.517.0
Predefined features0.23
Frequency domain0.361.23.1
All three combined3.34.023.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Halász, A.P.; Al Hemeary, N.; Daubner, L.S.; Juhász, J.; Zsedrovits, T.; Tornai, K. Adapting a Previously Proposed Open-Set Recognition Method for Time-Series Data: A Biometric User Identification Case Study. Electronics 2025, 14, 3983. https://doi.org/10.3390/electronics14203983

AMA Style

Halász AP, Al Hemeary N, Daubner LS, Juhász J, Zsedrovits T, Tornai K. Adapting a Previously Proposed Open-Set Recognition Method for Time-Series Data: A Biometric User Identification Case Study. Electronics. 2025; 14(20):3983. https://doi.org/10.3390/electronics14203983

Chicago/Turabian Style

Halász, András Pál, Nawar Al Hemeary, Lóránt Szabolcs Daubner, János Juhász, Tamás Zsedrovits, and Kálmán Tornai. 2025. "Adapting a Previously Proposed Open-Set Recognition Method for Time-Series Data: A Biometric User Identification Case Study" Electronics 14, no. 20: 3983. https://doi.org/10.3390/electronics14203983

APA Style

Halász, A. P., Al Hemeary, N., Daubner, L. S., Juhász, J., Zsedrovits, T., & Tornai, K. (2025). Adapting a Previously Proposed Open-Set Recognition Method for Time-Series Data: A Biometric User Identification Case Study. Electronics, 14(20), 3983. https://doi.org/10.3390/electronics14203983

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop