Next Article in Journal
A Hierarchically Structured Ni-NOF@ZIF-L Heterojunction Using Van Der Waals Interactions for Electrocatalytic Reduction of CO2 to HCOOH
Previous Article in Journal
From Classic to Cutting-Edge: A Near-Perfect Global Thresholding Approach with Machine Learning
Previous Article in Special Issue
Consensus Guided Multi-View Unsupervised Feature Selection with Hybrid Regularization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-View Cluster Structure Guided One-Class BLS-Autoencoder for Intrusion Detection

1
School of Computer and Cyber Security, Fujian Normal University, Fuzhou 350108, China
2
School of Engineering, Huaqiao University, Quanzhou 362000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(14), 8094; https://doi.org/10.3390/app15148094
Submission received: 19 May 2025 / Revised: 11 July 2025 / Accepted: 18 July 2025 / Published: 21 July 2025

Abstract

Intrusion detection systems are crucial for cybersecurity applications. Network traffic data originate from diverse terminal sources, exhibiting multi-view feature spaces, while the collection of unknown intrusion data is costly. Current one-class classification (OCC) approaches are mainly designed for single-view data. Multi-view OCC approaches usually require collecting multi-view traffic data from all sources and have difficulty detecting intrusion independently in each view. Furthermore, they commonly ignore the potential subcategories in normal traffic data. To address these limitations, this paper utilizes the Broad Learning System (BLS) technique and proposes an intrusion detection framework based on a multi-view cluster structure guided one-class BLS-autoencoder (IDF-MOCBLSAE). Specifically, a multi-view co-association matrix optimization objective function with doubly-stochastic constraints is first designed to capture the cross-view cluster structure. Then, a multi-view cluster structure guided one-class BLS-autoencoder (MOCBLSAEs) is proposed, which learns the discriminative patterns of normal traffic data by preserving the cross-view clustering structure while minimizing the intra-view sample reconstruction errors, thereby enabling the identification of unknown intrusion data. Finally, an intrusion detection framework is constructed based on multiple MOCBLSAEs to achieve both individual and ensemble intrusion detection. Through experimentation, IDF-MOCBLSAE is validated on real-world network traffic datasets for multi-view one-class classification tasks, demonstrating its superiority over state-of-the-art one-class approaches.

1. Introduction

In recent years, the rapid advancement of Internet of Things (IoT) technology has positioned network intrusion detection systems (IDS) as a critical component for ensuring IoT security. However, IDS usually deploy multiple sensors in different locations of the network and collect multi-view network traffic data by monitoring network traffic or system logs. Moreover, network intrusions are typically unknown threats, making the data collection process costly and difficult to perform comprehensively. These challenges have led to extensive research on multi-view one-class classification methods for anomaly detection, which aim to train the model based on normal traffic data from different sources, thus realizing the detection of unknown intrusions.
To address the scarcity of anomalous samples in cybersecurity scenarios, one-class classification methods (OCC) have emerged as a specialized solution. Unlike conventional binary classification frameworks that require both normal and anomalous samples for training, OCC models are trained solely from normal samples, which enables the identification of unknown intrusions during testing. Early OCC research focused on adapting traditional machine learning techniques, such as support vector machines [1], K-nearest neighbors [2], and fuzzy learning [3], to one-class scenarios. However, these methods exhibited limited detection performance due to their shallow architectures and manual feature engineering constraints. Subsequent advancements introduced deep neural networks, including convolutional neural network [4], multi-layer perceptron-based [5], and adversarial network-based [6,7,8] frameworks, which improved feature representation capabilities at the cost of increased structural complexity. Recent innovations prioritize computational efficiency for edge-deployed IDS through simplified neural architectures. Notable advancements leverage flat network models like extreme learning machine (ELM) [9,10] and the broad learning system (BLS) [11,12]. These methods eliminate deep hierarchical structures while preserving detection efficacy, achieving rapid training convergence through analytical learning paradigms.
Most existing OCC models are designed for single-view data, making them incapable of collaboratively learning multiple feature spaces from multi-source data. To address this limitation, some researchers began to explore multi-view one-class classification methods [13,14,15,16]. However, current methods still have two main limitations: (1) They primarily focus on mapping multi-view features into a shared feature space and then using a single OCC model for anomaly detection. As a result, they require collecting multi-view traffic data from all sources and have difficulty detecting intrusion independently in each view. (2) Due to various network applications, normal traffic data contain potential subcategories, and these models lack consideration of the cross-view cluster structure.
To address these problems, this paper proposes an intrusion detection framework based on a multi-view cluster structure guided one-class BLS-autoencoder (IDF-MOCBLSAE). Firstly, the ensemble clustering technique is utilized to derive multiple view-specific co-association matrices; then, a multi-view co-association matrix optimization objective function with doubly-stochastic constraints is designed to capture the global cluster structure. Secondly, a multi-view cluster structure guided one-class BLS-autoencoder (MOCBLSAE) is presented, which learns discriminative patterns of normal traffic data based on the cross-view global cluster structure and the reconstruction of intra-view local features, thereby enabling the identification of unknown anomalous data of each view. Finally, multiple MOCBLSAE models from different views are integrated into the intrusion detection framework to achieve both individual and ensemble intrusion detection. Extensive experiments on diverse OCC intrusion detection tasks are produced to demonstrate the effectiveness of the propose method. In summary, the main contributions of this paper are as follows:
1.
A multi-view co-association matrix optimization objective function with doubly-stochastic constraints is designed, which systematically unifies view-specific spectral embedding from ensemble partitions and better captures the underlying cluster structure across different views.
2.
A multi-view cluster structure guided one-class BLS-autoencoder (MOCBLSAE) is proposed, which learns discriminative patterns of normal traffic data by preserving the cross-view clustering structure while minimizing the intra-view sample reconstruction errors, thereby enabling the identification of unknown intrusion data.
3.
Based on MOCBLSAE, an intrusion detection framework is constructed and validated on real-world network traffic datasets for multi-view one-class classification tasks, demonstrating its superiority over state-of-the-art one-class methods.
The subsequent sections of this paper are structured as follows. Section 2 reviews related works on single-view and multi-view one-class classification. Section 3 introduces the proposed method in detail. Section 4 validates the effectiveness of the proposed method through a series of experiments. Section 5 summarizes this paper and discusses potential future research.

2. Related Works

In scenarios of extreme class imbalance, one-class classification (OCC) methods have been extensively investigated. These methods exclusively utilize samples of the target class during the training phase to develop discriminative models, which subsequently enable the detection of anomalous samples deviating from the target class during testing. Current research in OCC can be systematically categorized into three primary approaches: traditional machine learning-based, neural network-based, and multi-view OCC approaches.
Traditional machine learning-based OCC approaches were proposed previously which aimed to modify classic classification models to adapt to OCC scenarios. For example, Scholkopf et al. [17] proposed a classical one-class classification algorithm based on the principle of a support vector machine (SVM). Tax et al. [18] proposed the support vector data description (SVDD), which learned a decision boundary in the kernel space by minimizing the distance from the normal samples to the boundary. Pang et al. [1] improved the one-class SVM using vector quantization techniques, which enhanced the generalization across different OCC tasks. Ji et al. [19] combined one-class SVM with a long short-term memory network to improve the ability to discern anomalies in time-series data. Swarnkar et al. [20] designed a one-class Bayesian model for detecting HTTP attacks, reducing computational complexity while improving detection accuracy. Gornitz et al. [21] introduced a clustering-driven SVDD algorithm that constructed multiple spheres in the kernel space via K-means to identify anomalous data. Li et al. [3] proposed boundary-based fuzzy-SVDD by introducing a local–global center distance to identify boundary-proximate samples, which led to a more accurate classification boundary.
Neural network-based OCC approaches have garnered significant research interest in recent years, demonstrating superior detection capabilities in complex data environments. Many one-class deep neural networks have been proposed. For example, Pu et al. [6] presented a bi-directional generative adversarial network (GAN) for latent feature extraction of normal patterns. Peng et al. [7] designed a dual-iteration GAN framework with auxiliary losses to minimize the intra-class variance of normal samples. Xu et al. [5] proposed a calibrated one-class learning method which combines uncertainty modeling and synthetic anomalies to enhance contamination-robust normality boundary learning in anomaly detection. Massoli et al. [22] introduced multilayer one-class learning via hierarchical autoencoder optimization, combining layer-wise feature refinement and centroid-aligned training for enhanced anomaly detection across deep representations. Dong et al. [4] proposed a multitask deep one-class convolutional neural network combining an autoencoder for feature learning and a moving-window scanning approach for detection. Ding et al. [8] designed an end-to-end framework of deep transfer one-class classification, which combined adversarial generation of pseudo-negative samples with log-Euclidean manifold alignment of cross-domain positive samples to enhance discrimination and representation learning. Mauceri et al. [23] embedded time series into a latent space where the Euclidean distances matched the original dissimilarity measures using autoencoder and encoder-only neural networks. Furthermore, a series of one-class flat-structure neural networks were explored based on the extreme learning machine (ELM) [24] and the broad learning system (BLS) [25]. For example, Yang et al. [12] developed a stacked one-class BLS framework to separately extract normal traffic features and detect anomalies. Lin et al. [11] introduced a one-class BLS network with a maximum correntropy and boosting scheme to improve performance. Cao et al. [26] proposed a hierarchical ELM-based approach leveraging maximum mutual information to improve adversarial robustness. Yang et al. [27] presented an online sequential ELM capable of incrementally learning key features from normal data. Zhong et al. [28] enhanced the conventional BLS by incorporating a memory module to strengthen normal pattern representation.
Most existing OCC models are designed for single-view data, making them incapable of collaboratively learning multiple feature spaces from multi-source data. To address this limitation, some researchers began to explore multi-view one-class classification methods. For example, Xiao et al. [13] proposed a multi-view one-class SVM integrating privileged information learning to enforce both consensus and complementarity principles through quadratic programming. Lei et al. [14] designed a deep OCC framework with one-class crop extraction loss for multi-modal satellite imagery. Golo et al. [15] presented a one-class multi-modal learning method for detecting relevant reviews. Sohrab et al. [16] projected data from multiple modalities to a new subspace optimized for one-class classification.

3. Proposed Method

This section introduces the proposed method in detail, including the definition of symbols and the task, the process of multi-view co-association matrix optimization, and the intrusion detection framework, based on the multi-view cluster structure guided one-class BLS-autoencoder.

3.1. Definition

The training set from multiple sources is given by multi-view network traffic data X = { X 1 , , X V } , where V is the number of data sources (views). For each view data X v R n × d v ( v = 1 , , V ), the number of samples and features are n and d v , respectively. In the scenario of one-class intrusion detection, all multi-view training samples are normal traffic data (assumed to be labeled as 1). The multi-view one-class approach aims to jointly learn discriminative patterns of normal traffic data in the multi-view feature spaces, thereby enabling autonomous identification of anomalous traffic data during the application phase.

3.2. Multi-View Co-Association Matrix Optimization

Although all training samples are labeled as normal traffic data, they contain latent subcategories stemming from diverse network applications. Based on this intuition, a multi-view co-association matrix is first constructed from multi-view data, which preserves the cluster structure of samples.
Considering the efficiency and robustness of clustering structure analysis, an ensemble clustering strategy is adopted, which applies K-means in each view to obtain multiple base partitions { P 1 v , , P B v } , where B is the number of base partition in the v-th view. Each base partition is represented as a binary partition matrix P b v R n × c b [29]; the i j -th entry of P b v is defined as follows:
P b ( i j ) v = 1 , i f x i C j , 0 , o t h e r w i s e ,
where C j is the j-th cluster, and c b is the number of clusters randomly set within a certain range. By horizontally concatenating the binary partition matrices, the ensemble binary partition matrix P v is obtained for the v-th view, and the co-association matrix G v R n × n can be further computed as follows:
G v = 1 B P v ( P v ) T ,
P v = [ P 1 v P 2 v P B v ] .
It can be observed that G v effectively integrates diverse base partitions, representing the probability of sample pairs being assigned to the same cluster. This enables G v to comprehensively capture the intrinsic cluster structure relationship among samples within the v-th view.
Network traffic data collected from multiple sources reflect inter-sample relationships from different perspectives. Therefore, it is necessary to further integrate cross-view co-association matrices to obtain more robust cluster structure information.
However, simply averaging multiple view-specific co-association matrices fails to effectively integrate cluster structure information across different views. To address this problem, a multi-view co-association matrix optimization objective function with doubly-stochastic constraints is proposed as follows:
min G v = 1 V | | G G v | | F 2 , s . t . G 0 , G 1 = 1 , G T = G .
In the absence of prior knowledge, all views are treated equally to learn a multi-view co-association matrix that approximates the view-specific co-association matrices. By imposing doubly-stochastic constraints, the spectral relationships between samples are further refined, yielding a globally optimized co-association matrix that better reflects the underlying cluster structure across different views.
By setting Q = 1 V v = 1 V G v , problem (4) can be reformulated into a doubly-stochastic normalization problem under the Frobenius norm [30] as follows:
min G | | G Q | | F 2 , s . t . G 0 , G 1 = 1 , G T = G .
Then, according to the von Neumann successive projection lemma, the global optimal solution can be derived by iteratively solving the following two problems:
min G | | G Q | | F 2 , s . t . G 1 = 1 , G T = G .
min G | | G Q | | F 2 , s . t . G 0 .
The closed form solution of the problem (6) can be optimized by the Lagrange method as follows:
G = Q + ( I Q n + 1 1 T Q n 2 ) 1 1 T 1 n 1 1 T Q .
As for problem (7), it can be solved by simply projecting it onto a non-zero space as follows:
G = Q + .
In general, the initial G is G = Q = 1 V v = 1 V G v . Then, it is iteratively updated by computing Q = G and Equations (8) and (9) until convergence to obtain the multi-view co-association matrix. The pseudo code of the multi-view co-association matrix optimization is shown in Algorithm 1.
Algorithm 1 Multi-view co-association matrix optimization
  • Require:
    •  Input: Multi-view training set X = { X 1 , , X V } .
  • Ensure:
    1:
    for each view v = 1 to V do
    2:
        Apply K-means to produce B base partitions and compute the corresponding binary partition matrices { P 1 v , , P B v } according to Equation (1);
    3:
        Compute the view-specific co-association matrix G v according to Equations (2) and (3);
    4:
    end for
    5:
    Compute Q = 1 V v = 1 V G v ;
    6:
    repeat
    7:
        Update G according to Equation (8);
    8:
        Update G = G + ;
    9:
        Update Q = G ;
    10:
    untilG converges
    Output The multi-view co-association matrix G;

3.3. Intrusion Detection Framework Based on the Multi-View Cluster Structure Guided One-Class BLS-Autoencoder

In the scenario of multi-view one-class intrusion detection, network traffic data originate from diverse terminal devices, and the training set consists of only normal samples, making conventional binary classification algorithms difficult to adapt directly. To this end, this paper aims to fully leverage multi-view normal samples to train an intrusion detection model, which identifies anomalous samples independently and efficiently across different data sources while maintaining ensemble decision-making capabilities. Instead of learning the direct mapping relationship between samples and class labels, this paper proposes a multi-view cluster structure guided one-class BLS-autoencoder (MOCBLSAE), which learns discriminative patterns of normal traffic data based on the cross-view global cluster structure and the reconstruction of intra-view local features, thereby enabling the identification of unknown anomalous samples. Finally, an intrusion detection framework is constructed based on multiple MOCBLSAE models.
The framework of MOCBLSAE is shown in Figure 1. Specifically, in the v-th view, MOCBLSAE first applies orthogonal stochastic weights and biases during the generation of feature nodes and enhancement nodes, then constructs the hidden layer A v by concatenating multiple feature nodes and enhancement nodes to enhance the model’s generalization capability [31] as follows:
A v = [ F 1 v F 2 v F p v E 1 v E 2 v E q v ] ,
F i v = ϕ f ( X v W i v + C i v ) , s . t . W i v ( W i v ) T = I , ( C i v ) u ( C i v ) u T = 1 , ( C i v ) u = ( C i v ) v ,
E j v = ϕ e ( X v W j v + C j v ) , s . t . W j v ( W j v ) T = I , ( C j v ) u ( C j v ) u T = 1 , ( C j v ) u = ( C j v ) v ,
where W i v and W j v are orthogonal stochastic weight matrices, C i v and C j v are orthogonal stochastic bias matrices, ( C i v ) u and ( C j v ) u are the u-th row vectors of C i v and C j v , and ϕ f and ϕ e are activation functions.
After obtaining the hidden layer A v R n × l , MOCBLSAE incorporates global cluster structure information to capture the discriminative patterns of diverse normal samples, thereby achieving effective data reconstruction. In addition, MOCBLSAE lacks reconstruction capability for anomalous data, allowing for rapid identification through reconstruction errors. Based on this idea, the objective function of MOCBLSAE is designed as follows:
min W v | | A v W v X v | | F 2 + λ | | W v | | F 2 + α 2 i = 1 n j = 1 n g i j · | | ( ( W v ) T ( A v ) T ) i ( ( W v ) T ( A v ) T ) j | | 2 2 ,
where A v is the hidden layer, W v is the connecting matrix between the hidden layer and the output layer, g i j is the i j -th entry of the global co-association matrix G, ( ( W v ) T ( A v ) T ) i represents the column vector of ( W v ) T ( A v ) T , and λ and α are trade-off parameters. The first term of problem (13) minimizes the reconstruction error of normal samples, the second term prevents overfitting by imposing the Frobenius norm, and the third term preserves the global cluster relationships among samples.
In order to optimize the objective function, it can be reformulated into a matrix form as follows:
min W v J = | | A v W v X v | | F 2 + λ | | W v | | F 2 + α T r ( ( W v ) T ( A v ) T L G W v A v ) ,
where L G is the Laplacian matrix of G. By setting J / W v = 0 , the following equation can be obtained:
J W v = 2 ( ( A v ) T A v + λ I + α ( A v ) T L G A v ) W v 2 ( A v ) T X v = 0 .
When the training data are sufficient (i.e., n l ), A v is a full-rank matrix, and Equation (15) has a closed-form solution. When the training data are insufficient (i.e., n < l ), Equation  (15) has infinite solutions, W v = A v D ( D R n × n ) can be imposed on W v and both sides of the equation multiplied by ( ( A v ) T A v ) 1 A v to obtain a unique solution [32]. In general, Equation (15) can be efficiently solved as follows:
W v = λ I + ( A v ) T A v + α ( A v ) T L G A v 1 ( A v ) T X v , n l , ( A v ) T λ I + A v ( A v ) T + α L G A v ( A v ) T 1 X v , n < l .
After solving for W v , the reconstruction errors of all samples within the view can be obtained via Equation (17). For normal traffic data, MOCBLSAE achieves relatively small reconstruction errors by encoding and reconstructing the data while incorporating their cluster relationships. In contrast, since MOCBLSAE does not learn the reconstruction features of network intrusion data, it typically yields larger reconstruction errors. Based on this idea, the top n δ ( δ [ 0 , 1 ] ) largest training reconstruction error is selected as the threshold T v . When the reconstruction error of unknown traffic data exceeds this threshold, MOCBLSAE identifies it as anomalous. The specific formula is given below:
E r r ( x v ) = | | x v A v ( x v ) W v | | 2 2 ,
L v ( x t v ) = 1 ( normal ) , if E r r ( x v ) T v , 1 ( intrusion ) , otherwise ,
where E r r ( x v ) is the reconstruction error of x v , A v ( x v ) is the output of the hidden layer, and L v ( x t v ) is the view-specific predicted label of testing data x t v .
Finally, the intrusion detection framework based on the multi-view cluster structure guided one-class BLS-autoencoder (IDF-MOCBLSAE) is constructed as shown in Figure 2. In the training stage, IDF-MOCBLSAE captures the cross-view cluster structure by applying the ensemble clustering and the multi-view co-association matrix optimization. Then, MOCBLSAE is trained in each view to learn the discriminative patterns of normal traffic data by preserving the cross-view clustering structure while minimizing the intra-view sample reconstruction errors. After training, an ensemble decision pool consisting of multiple MOCBLSAE models is obtained. In the testing stage, for multi-view traffic data, IDF-MOCBLSAE can operate independently within individual views or perform cross-view ensemble decision-making (e.g.,, majority voting or one-vote veto strategies), enabling efficient and flexible identification of unknown network intrusion. The overall pseudo-code of the intrusion detection framework based on MOCBLSAE is demonstrated in Algorithm 2.
Algorithm 2 Intrusion detection based on multi-view cluster structure guided one-class BLS-autoencoder (IDF-MOCBLSAE)
  • Require:
    •  Input: Multi-view training set X = { X 1 , , X V } , parameters λ , α , δ , multi-view testing sample x t = { x t 1 , , x t V } .
  • Ensure:
    •  # Training stage
    1:
    Compute the multi-view co-association matrix G according to Algorithm 1;
    2:
    for each view v = 1 to V do
    3:
        Construct the hidden layer A v according to Equations (10)–(12);
    4:
        Compute the connecting matrix W v according to Equation (16);
    5:
        Compute the reconstruction error of all training samples according to Equation (17), and set the threshold T v by the top n δ ( δ [ 0 , 1 ] ) largest error;
    6:
    end for
    # Testing stage
    7:
    for each view v = 1 to V do
    8:
        Compute the reconstruction error of x t v according to Equation (17);
    9:
        Compute the view-specific predicted label L v ( x t v ) according to Equation (18);
    10:
    end for
    11:
    Obtain the prediction result of L ( x t ) based on the one-vote veto strategie.
    Output: The prediction result of x t .

4. Experiments

4.1. Datasets and Experimental Setup

To validate the effectiveness of IDF-MOCBLSAE, this section constructs 10 multi-view one-class intrusion detection tasks based on two real-world network traffic datasets (i.e., KDD [33] and UNSW [34]). Table 1 presents the basic information of each task, where the training set consists of normal data, with n t r a i n denoting the number of training samples, and the testing set comprises both normal and intrusion data, with n t e s t indicating the number of testing samples. For each task, the original feature space is randomly partitioned into two disjoint views and provides the corresponding number of features for each view.
In the experiments performed, the one-vote veto strategy is used in IDF-MOCBLSAE. All methods are executed 20 times, and their average accuracy and F1-score metrics, as shown in Equations (19) and (20) and Table 2, are reported to comprehensively evaluate the performance.
Accuracy = T P + T N T P + T N + F P + F N ,
F 1 - score = 2 P R P + R , P = T P T P + F P , R = T P T P + F N .

4.2. Comparison of OCC Methods

In this experiment, IDF-MOCBLSAE is compared with several state-of-the-art OCC methods, including the stacked one-class broad learning system (ST-OCBLS) [12], multilayer one-class extreme learning machine (ML-OCELM) [10], deep one-class classification based on SVDD (DeepSVDD) [35], outlier analysis based on deep autoencoder (DeepAE) [36], and the Hybrid Ensemble Broad Learning System (SMC-OCBLS) [11]. Among them, ST-OCBLS is a stacked autoencoder variant of OCBLS, enhancing detection efficiency and accuracy through hierarchical feature learning. ML-OCELM, another stacked AE-based extension of OCELM, leverages the integration of deep autoencoders and kernel learning frameworks to improve performance on complex data and multi-class classification tasks. Deep-SVDD, on the other hand, is a deep neural network adaptation of SVDD [18], designed to strengthen its applicability as a one-class classifier. DeepAE employs a deeper architecture to enhance learning of the features for OCC tasks. SMC-OCBLS is a one-class BLS network with a maximum correntropy and boosting scheme to improve the performance. All views are concatenated into a complete feature space and the comparative OCC methods are directly applied for intrusion detection.
The comparison results in terms of accuracy and F1-score are shown in Table 3 and Table 4, respectively, with the best-performing results highlighted in bold. It can be observed that IDF-MOCBLSAE achieves superior performance on 6 out of 10 tasks in terms of accuracy and 7 out of 10 tasks in terms of the F1-score. This performance advantage can be attributed to two key factors: (1) Existing state-of-the-art OCC methods simply concatenate multi-view features without effective collaborative learning mechanisms; (2) IDF-MOCBLSAE enhances overall detection performance by simultaneously preserving the inter-view cluster structures while minimizing intra-view reconstruction errors.

4.3. Parameter Sensitivity Analysis

This subsection systematically examines the effect of parameters α and λ , which control the trade-off importance of preserving the cross-view cluster structure and the Frobenius norm, respectively. Through controlled experimentation that involves independently modulating each parameter while maintaining others fixed, our sensitivity analysis reveals parameter-dependent performance variations as illustrated in Figure 3. A grid search is conducted for both parameters within the range [ 0.001 , 1.0 ] , with red color indicating better performance in terms of accuracy. It can be observed that IDF-MOCBLSAE demonstrates better robustness to the selection of λ . As for α , IDF-MOCBLSAE maintains better performance when α < 0.05 , and the performance decreases when α increases. The possible reason is that applying the excessive constraint of preserving the cross-view cluster structure may affect the learning process of sample reconstruction. In conclusion, the recommended parameter settings are α [ 0.001 , 0.05 ] and λ [ 0.01 , 0.5 ] .

4.4. Ablation Analysis

In this paper, the multi-view co-association matrix optimization objective function with doubly-stochastic constraints is proposed in Equation (4) to capture a higher quality cross-view cluster structure. Furthermore, MOCBLSAE is improved by preserving the cross-view clustering structure while minimizing the intra-view sample reconstruction errors. To verify the effectiveness of the components of IDF-MOCBLSAE, IDF-MOCBLSAE is compared with two methods: OCBLSAE removes the third term of cluster structure preserving in Equation (13); MOCBLSAE-G simply computes G = 1 V G v instead of solving problem (4). The experimental results in terms of accuracy and F1-score are shown in Figure 4 and Figure 5, respectively. According to the experimental results, IDF-MOCBLSAE has an overall better performance than OCBLSAE and MOCBLSAE-G, which demonstrates the effectiveness of the cross-view cluster structure guidance in MOCBLSAE and the optimization of the multi-view co-association matrix.

4.5. Time Efficiency Analysis

To evaluate the training efficiency of different OCC methods, this section conducts model training on the KDD dataset using varying amounts of normal traffic data ranging from 2500 to 20,000 samples and recording the corresponding time consumption. The experimental results are shown in Figure 6 where the red line represents the running time of IDF-MOCBLSAE. As can be observed, deep learning-based OCC approaches (DeepAE and DeepSVDD) exhibit a significant increase in time consumption as the sample size grows, while flat-structured network-based OCC approaches demonstrate superior training efficiency.
Figure 4. Ablation experiment in terms of accuracy.
Figure 4. Ablation experiment in terms of accuracy.
Applsci 15 08094 g004
Figure 5. Ablation experiment in terms of F1-score.
Figure 5. Ablation experiment in terms of F1-score.
Applsci 15 08094 g005
Figure 6. Time complexity of different OCC methods.
Figure 6. Time complexity of different OCC methods.
Applsci 15 08094 g006

5. Conclusions

With the growing prevalence of network traffic data sharing across multiple sources, cybersecurity has become a critical priority. This paper proposes an intrusion detection framework based on a multi-view cluster structure guided one-class BLS-autoencoder (IDF-MOCBLSAE). A multi-view co-association matrix optimization objective function with doubly-stochastic constraints is first designed to combine view-specific spectral embeddings from ensemble partitions and better captures the underlying cluster structure across different views. Then, a multi-view cluster structure guided one-class BLS-autoencoder (MOCBLSAE) is proposed, which learns discriminative patterns of normal traffic data by preserving the cross-view clustering structure while minimizing the intra-view sample reconstruction errors. Finally, the intrusion detection framework is constructed based on multiple MOCBLSAEs to achieve both individual and ensemble intrusion detection. In the application, IDF-MOCBLSAE is deployed on the data analysis node of IDS. During the training phase, it collects multi-view normal network traffic data from multiple sources and trains MOCBLSAE models for each view. In the detection phase, each view-specific datum can be independently predicted using the corresponding MOCBLSAE model, without collecting data from all sources. In addition, to achieve a more comprehensive decision-making process, multiple view-specific prediction results can be integrated into the ensemble prediction by majority voting or one-vote veto strategies.
To demonstrate the effectiveness of the proposed method, this paper conducts a series of experiments and analyzes the results: (1) IDF-MOCBLSAE is compared with state-of-the-art OCC methods. The experimental results show that IDF-MOCBLSAE achieves overall superior performance to the comparison methods. (2) The effects of two trade-off parameters in IDF-MOCBLSAE are analyzed, and the recommended parameter settings are provided. (3) The ablation analysis on IDF-MOCBLSAE is presented, which demonstrates the effectiveness of the cross-view cluster structure guidance in MOCBLSAE and the optimization of the multi-view co-association matrix. (4) The training efficiency of IDF-MOCBLSAE and the comparison OCC methods is evaluated. The experimental results show that IDF-MOCBLSAE can achieve better intrusion detection performance with training efficiency similar to state-of-the-art OCC methods.
The limitations of the proposed method include the following: (1) The detection performance for high-dimensional noisy traffic data still needs further improvement. (2) The training process of IDF-MOCBLSAE is centralized, lacking an online learning mechanism for data streaming. Therefore, the future research directions are summarized as follows: (1) Incorporating attention mechanisms, feature attribution methods, or visualization clustering to identify discriminative features or views. (2) Exploring asynchronous incremental learning techniques for multi-view OCC models by integrating continual learning paradigms, enhancing adaptability to evolving threats.

Author Contributions

Funding acquisition, Y.S.; Methodology, Q.Y., Y.-A.C. and Y.S.; Supervision, Y.S.; Validation, Q.Y. and Y.-A.C.; Writing—original draft, Q.Y. and Y.-A.C.; Writing—review and editing, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under No. 62306122, the Fundamental Research Funds for the Central Universities under No. ZQN-1122, the Scientific Research Funds of Huaqiao University under Grant 24BS141, and the High-level Talent Team Project of Quanzhou City under grant No. 2023CT001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in "A Detailed Analysis of the KDD CUP 99 Dataset" at https://doi.org/10.1109/cisda.2009.5356528 (accessed on 18 December 2009), reference number [33] and UNSW-NB15: A Comprehensive Dataset for Network Intrusion Detection Systems (UNSW-NB15 Network Dataset) at https://doi.org/10.1109/milcis.2015.7348942 (accessed on 10 December 2015), reference number [34].

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Pang, J.; Pu, X.; Li, C. A Hybrid Algorithm Incorporating Vector Quantization and One-Class Support Vector Machine for Industrial Anomaly Detection. IEEE Trans. Ind. Inform. 2022, 18, 8786–8796. [Google Scholar] [CrossRef]
  2. Sarmadi, H.; Karamodin, A. A novel anomaly detection method based on adaptive mahalanobis-squared distance and one-class knn rule for structural health monitoring under environmental effects. Mech. Syst. Signal Process. 2020, 140, 106495. [Google Scholar] [CrossRef]
  3. Li, D.; Xu, X.; Wang, Z.; Cao, C.; Wang, M. Boundary-based Fuzzy-SVDD for one-class classification. Int. J. Intell. Syst. 2022, 37, 2266–2292. [Google Scholar] [CrossRef]
  4. Dong, X.; Taylor, C.J.; Cootes, T.F. Defect Classification and Detection Using a Multitask Deep One-Class CNN. IEEE Trans. Autom. Sci. Eng. 2022, 19, 1719–1730. [Google Scholar] [CrossRef]
  5. Xu, H.; Wang, Y.; Jian, S.; Liao, Q.; Wang, Y.; Pang, G. Calibrated One-Class Classification for Unsupervised Time Series Anomaly Detection. IEEE Trans. Knowl. Data Eng. 2024, 36, 5723–5736. [Google Scholar] [CrossRef]
  6. Peng, H.; Zhao, J.; Li, L.; Ren, Y.; Zhao, S. One-Class Adversarial Fraud Detection Nets With Class Specific Representations. IEEE Trans. Netw. Sci. Eng. 2023, 10, 3793–3803. [Google Scholar] [CrossRef]
  7. Pu, Z.; Cabrera, D.; Bai, Y.; Li, C. A One-Class Generative Adversarial Detection Framework for Multifunctional Fault Diagnoses. IEEE Trans. Ind. Electron. 2022, 69, 8411–8419. [Google Scholar] [CrossRef]
  8. Ding, Y.; Jia, M.; Cao, Y.; Yan, X.; Zhao, X.; Lee, C.G. Unsupervised Fault Detection With Deep One-Class Classification and Manifold Distribution Alignment. IEEE Trans. Ind. Inform. 2024, 20, 1313–1323. [Google Scholar] [CrossRef]
  9. Leng, Q.; Qi, H.; Miao, J.; Zhu, W.; Su, G. One-Class Classification with Extreme Learning Machine. Math. Probl. Eng. 2015, 2015, 412957. [Google Scholar] [CrossRef]
  10. Dai, H.; Cao, J.; Wang, T.; Deng, M.; Yang, Z. Multilayer One-Class Extreme Learning Machine. Neural Netw. 2019, 115, 11–22. [Google Scholar] [CrossRef] [PubMed]
  11. Lin, M.; Yang, K.; Yu, Z.; Shi, Y.; Chen, C.L.P. Hybrid Ensemble Broad Learning System for Network Intrusion Detection. IEEE Trans. Ind. Inform. 2024, 20, 5622–5633. [Google Scholar] [CrossRef]
  12. Yang, K.; Shi, Y.; Yu, Z.; Yang, Q.; Sangaiah, A.K.; Zeng, H. Stacked One-Class Broad Learning System for Intrusion Detection in Industry 4.0. IEEE Trans. Ind. Inform. 2023, 19, 251–260. [Google Scholar] [CrossRef]
  13. Xiao, Y.; Pan, G.; Liu, B.; Zhao, L.; Kong, X.; Hao, Z. Privileged Multi-View One-Class Support Vector Machine. Neurocomputing 2024, 572, 127186. [Google Scholar] [CrossRef]
  14. Lei, L.; Wang, X.; Zhong, Y.; Zhao, H.; Hu, X.; Luo, C. DOCC: Deep one-class crop classification via positive and unlabeled learning for multi-modal satellite imagery. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102598. [Google Scholar] [CrossRef]
  15. Golo, M.P.S.; Araujo, A.F.; Rossi, R.G.; Marcacini, R.M. Detecting Relevant App Reviews for Software Evolution and Maintenance Through Multimodal One-Class Learning. Inf. Softw. Technol. 2022, 151, 106998. [Google Scholar] [CrossRef]
  16. Sohrab, F.; Raitoharju, J.; Iosifidis, A.; Gabbouj, M. Multimodal subspace support vector data description. Pattern Recognit. 2021, 110, 107648. [Google Scholar] [CrossRef]
  17. Schölkopf, B.; Platt, J.C.; Shawe-Taylor, J.; Smola, A.J.; Williamson, R.C. Estimating the support of a high-dimensional distribution. Neural Comput. 2001, 13, 1443–1471. [Google Scholar] [CrossRef] [PubMed]
  18. Tax, D.M.; Duin, R.P. Support vector data description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  19. Ji, Y.; Lee, H. Event-Based Anomaly Detection Using a One-Class SVM for a Hybrid Electric Vehicle. IEEE Trans. Veh. Technol. 2022, 71, 6032–6043. [Google Scholar] [CrossRef]
  20. Swarnkar, M.; Hubballi, N. OCPAD: One class Naive Bayes classifier for payload based anomaly detection. Expert Syst. Appl. 2016, 64, 330–339. [Google Scholar] [CrossRef]
  21. Goernitz, N.; Lima, L.A.; Mueller, K.R.; Kloft, M.; Nakajima, S. Support vector data descriptions and k-means clustering: One class. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 3994–4006. [Google Scholar] [CrossRef] [PubMed]
  22. Massoli, F.V.; Falchi, F.; Kantarci, A.; Akti, Ş.; Ekenel, H.K.; Amato, G. MOCCA: Multilayer one-class classification for anomaly detection. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 2313–2323. [Google Scholar] [CrossRef] [PubMed]
  23. Mauceri, S.; Sweeney, J.; Nicolau, M.; McDermott, J. Dissimilarity-Preserving Representation Learning for One-Class Time Series Classification. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 13951–13962. [Google Scholar] [CrossRef] [PubMed]
  24. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  25. Chen, C.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 10–24. [Google Scholar] [CrossRef] [PubMed]
  26. Cao, J.; Dai, H.; Lei, B.; Yin, C.; Zeng, H.; Kummert, A. Maximum Correntropy Criterion-Based Hierarchical One-Class Classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3748–3754. [Google Scholar] [CrossRef] [PubMed]
  27. Yang, Z.; Long, J.; Zi, Y.; Zhang, S.; Li, C. Incremental Novelty Identification From Initially One-Class Learning to Unknown Abnormality Classification. IEEE Trans. Ind. Electron. 2022, 69, 7394–7404. [Google Scholar] [CrossRef]
  28. Zhong, Z.; Yu, Z.; Fan, Z.; Chen, C.L.P.; Yang, K. Adaptive Memory Broad Learning System for Unsupervised Time Series Anomaly Detection. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 8331–8345. [Google Scholar] [CrossRef] [PubMed]
  29. Strehl, A.; Ghosh, J. Cluster ensembles: A knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 2002, 3, 583–617. [Google Scholar] [CrossRef]
  30. Zass, R.; Shashua, A. Doubly stochastic normalization for spectral clustering. In Proceedings of the Conference and Workshop on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 1569–1576. [Google Scholar] [CrossRef]
  31. Kasun, L.; Zhou, H.; Huang, G.; Vong, C. Representational learning with elms for big data. IEEE Intell. Syst. 2013, 28, 31–34. [Google Scholar]
  32. Huang, G.; Song, S.; Gupta, J.N.D.; Wu, C. Semi-Supervised and Unsupervised Extreme Learning Machines. IEEE Trans. Cybern. 2014, 44, 2405–2417. [Google Scholar] [CrossRef] [PubMed]
  33. Tavallaee, M.; Bagheri, E.; Lu, W.; Ghorbani, A. A Detailed Analysis of the KDD CUP 99 Data Set. In Proceedings of the IEEE Symposium on Computational Intelligence for Security and Defense Applications, Ottawa, ON, Canada, 8–10 July 2009; pp. 1–6. [Google Scholar] [CrossRef]
  34. Moustafa, N.; Slay, J. UNSW-NB15: A Comprehensive Data Set for Network Intrusion Detection Systems (UNSW-NB15 Network Data Set). In Proceedings of the Military Communications and Information Systems Conference, Canberra, Australia, 10–12 November 2015; pp. 1–6. [Google Scholar] [CrossRef]
  35. Ruff, L.; Vandermeulen, R.; Goernitz, N.; Deecke, L.; Siddiqui, S.A.; Binder, A.; Müller, E.; Kloft, M. Deep one-class classification. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 4393–4402. [Google Scholar]
  36. Aggarwal, C.C. Outlier Analysis; Springer: Cham, Switzerland, 2015; pp. 75–79. [Google Scholar] [CrossRef]
Figure 1. The framework of MOCBLSAE.
Figure 1. The framework of MOCBLSAE.
Applsci 15 08094 g001
Figure 2. The framework of IDF-MOCBLSAE.
Figure 2. The framework of IDF-MOCBLSAE.
Applsci 15 08094 g002
Figure 3. Effects of parameters α and λ in terms of accuracy.
Figure 3. Effects of parameters α and λ in terms of accuracy.
Applsci 15 08094 g003
Table 1. Multi-view one-class intrusion detection tasks based on two network traffic datasets.
Table 1. Multi-view one-class intrusion detection tasks based on two network traffic datasets.
TaskIntrusion Type n train n test (Normal + Intrusion)Dimensions of Each View
KDD1Warezclient20,00020,000 + 890{20, 21}
KDD2Portsweep20,00020,000 + 2931{20, 21}
KDD3Teardrop20,00020,000 + 892{20, 21}
KDD4Nmap20,00020,000 + 1493{20, 21}
KDD5Pod20,00020,000 + 201{20, 21}
UNSW1Reconnaissance18,50018,500 + 3496{21, 21}
UNSW2DoS18,50018,500 + 4089{21, 21}
UNSW3Exploits18,50018,500 + 11,132{21, 21}
UNSW4Fuzzers18,50018,500 + 6062{21, 21}
UNSW5Generic18,50018,500 + 18,871{21, 21}
Table 2. Confusion matrix.
Table 2. Confusion matrix.
Actual ClassPredicted Class
Intrusion (Positive)Normal (Negative)
Intrusion (Positive)TPFN
Normal (Negative)FPTN
Table 3. The comparison results in terms of accuracy where the bold values indicate the best performance.
Table 3. The comparison results in terms of accuracy where the bold values indicate the best performance.
TaskIDF-MOCBLSAESMC-OCBLSST-OCBLSML-OCELMDeepAEDeepSVDD
KDD10.95520.88040.80180.76240.57990.6261
KDD20.87110.94780.93260.84250.96010.8801
KDD30.95250.89640.88730.91430.36160.7525
KDD40.93810.91810.88910.76910.42340.604
KDD50.88580.80010.76820.81190.47280.7921
UNSW10.84820.84810.84150.84560.76990.7774
UNSW20.92230.94550.93030.92240.85590.8681
UNSW30.71390.86390.85310.74510.68170.7326
UNSW40.81460.81080.80230.78680.74890.7545
UNSW50.88810.98310.94040.76140.83280.7433
Table 4. The comparison results in terms of F1-score where the bold values indicate the best performance.
Table 4. The comparison results in terms of F1-score where the bold values indicate the best performance.
TaskIDF-MOCBLSAESMC-OCBLSST-OCBLSML-OCELMDeepAEDeepSVDD
KDD10.97710.91630.85310.79930.57260.9163
KDD20.93090.96230.95150.86710.97070.9623
KDD30.97570.92810.92210.93840.16910.9281
KDD40.96760.94190.92050.78830.30190.9419
KDD50.94280.86940.85150.87590.39450.8694
UNSW10.31210.22210.23060.15520.11020.2221
UNSW20.89870.83570.78890.76430.34250.8357
UNSW30.76790.78560.76770.49650.43190.5493
UNSW40.47840.42520.41010.27610.36020.3824
UNSW50.85810.98310.93750.65810.82260.6898
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Q.; Chen, Y.-A.; Shi, Y. Multi-View Cluster Structure Guided One-Class BLS-Autoencoder for Intrusion Detection. Appl. Sci. 2025, 15, 8094. https://doi.org/10.3390/app15148094

AMA Style

Yang Q, Chen Y-A, Shi Y. Multi-View Cluster Structure Guided One-Class BLS-Autoencoder for Intrusion Detection. Applied Sciences. 2025; 15(14):8094. https://doi.org/10.3390/app15148094

Chicago/Turabian Style

Yang, Qifan, Yu-Ang Chen, and Yifan Shi. 2025. "Multi-View Cluster Structure Guided One-Class BLS-Autoencoder for Intrusion Detection" Applied Sciences 15, no. 14: 8094. https://doi.org/10.3390/app15148094

APA Style

Yang, Q., Chen, Y.-A., & Shi, Y. (2025). Multi-View Cluster Structure Guided One-Class BLS-Autoencoder for Intrusion Detection. Applied Sciences, 15(14), 8094. https://doi.org/10.3390/app15148094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop