Deep Ensemble Learning for Application Traffic Classification Using Differential Model Selection Technique
Abstract
:1. Introduction
- Traffic monitoring and optimization: Network resources can be efficiently allocated by analyzing the traffic volume of specific applications.
- Security enhancement: Abnormal traffic or attack traffic can be detected to prevent or respond to network attacks.
- Bandwidth management: Bandwidth can be limited or prioritized for applications that use high bandwidth.
- Network performance analysis: The impact of specific applications on network performance can be analyzed and improved.
- Regulation and audit: Traffic records can be analyzed to meet legal or regulatory requirements.
- Effectiveness: It should provide traffic visibility and accurately classify network traffic.
- Deployability: Traffic classification models should be deployable within network assets and constraints.
- Trustworthiness: The results of traffic classification should be reliable.
- Robustness: The model should continue to function properly despite changes in the network environment.
- Adaptivity: When adjusting classification tasks according to environmental changes, the classification model should be able to adapt to these changes.
2. Related Works
2.1. Deep Learning-Based Application Traffic Classification
2.2. Application Traffic Classification Using Ensemble Techniques
2.3. Straight-Through Gumbel-Softmax
2.4. Model Selection Mechanism
- Intermediate output generation: Intermediate classifiers (ensemble agent models) receive data samples and generate intermediate outputs (agent predictions).
- Model selection: The model selector (selection net) receives data samples and generates a one-hot vector determining which model’s intermediate output to use.
- Aggregation and classification: The final output is generated by multiplying the intermediate classifier’s intermediate output by the model selector’s one-hot vector.
3. Deep Ensemble Using the Model Selection Technique
3.1. Overview of the Deep Ensemble Process
3.2. Baselines for Intermediate Classifiers
3.3. Loss Functions for Improving Error Diversity and Learning Stability
4. Experiments and Evaluation
4.1. Datasets
- Incomplete session: Elimination of TCP or TLS sessions without a hand-shake process
- Non-payload session: Elimination of TCP sessions that do not contain a payload
- Unrelated protocol sessions: Elimination of sessions from protocols considered to be unrelated to applications, such as DNS, LLMNR, MDNS, etc. [31].
4.2. Overall Comparisons with Other Methods
4.3. Analysis of Training History Based on the Application of Loss Functions
4.4. Comparative Analysis of Loss Function Weights
4.5. Analysis of Error Diversity
4.6. Comparison of Homogeneous and Heterogeneous Model Ensembles
5. Discussion
6. Conclusions
- Combining with the bagging mechanism, which is one of the existing ensemble methods, as mentioned in the Discussion.
- Improving the combination method to address cases where the proposed method, despite its superior performance compared to existing methods, fails to classify some data instances that were classified by existing models.
- Gaining a deeper understanding of the operating principles of the model selector and extending it towards Explainable Artificial Intelligence (XAI).
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Patel, S. 101+ Software Development Statistics You Should Know in 2025. MindInventory. Available online: https://www.mindinventory.com/blog/software-development-statistics/ (accessed on 15 January 2025).
- The Business Research Company. Low-Code Development Platform Market Report 2025—Low-Code Development Platform Market Trends and Size. Available online: https://www.thebusinessresearchcompany.com/report/low-code-development-platform-global-market-report (accessed on 15 January 2025).
- Gitnux.org. Internet Traffic Statistics: A Look at Data Driving Online Behavior. Available online: https://gitnux.org/internet-traffic-statistics/ (accessed on 15 January 2025).
- The Cloudflare Blog. Cloudflare 2023 Year in Review. Available online: https://blog.cloudflare.com/radar-2023-year-in-review/ (accessed on 15 January 2025).
- Wang, Z.; Zeng, Q.; Liu, Y.; Li, P. Malware traffic classification using convolutional neural network for representation learning. In Proceedings of the 2017 International Conference on Information Networking (ICOIN), Da Nang, Vietnam, 11–13 January 2017; pp. 712–717. [Google Scholar]
- Wang, W.; Zhu, M.; Wang, J.; Zeng, X.; Yang, Z. End-to-end encrypted traffic classification with one-dimensional convolution neural networks. In Proceedings of the 2017 IEEE International Conference on Intelligence and Security Informatics (ISI), Beijing, China, 22–24 July 2017; pp. 43–48. [Google Scholar]
- Yang, Y.; Kang, K.; Jiang, L.; Gao, Z.; Guo, Y.; Zhang, J.; Deng, J. HAST-IDS: Learning Hierarchical Spatial-Temporal Features Using Deep Neural Networks to Improve Intrusion Detection. IEEE Access 2017, 5, 26954–26964. [Google Scholar]
- Liu, C.; Liu, L.; Wang, W.; Wang, G.; Zhang, H. Self-attentive deep learning method for online traffic classification and its interpretability. Comput. Netw. 2021, 197, 108285. [Google Scholar]
- Lin, X.; Xiong, G.; Gou, G.; Li, Z.; Shi, J.; Yu, J. ET-BERT: A Contextualized Datagram Representation with Pre-training Transformers for Encrypted Traffic Classification. In Proceedings of the ACM Web Conference 2022 (WWW ’22), Lyon, France, 25–29 April 2022; pp. 633–642. [Google Scholar]
- Zhang, S.; Fu, D.; Liang, W.; Zhang, Z.; Yu, B.; Cai, P.; Yao, B. TrafficGPT: Viewing, processing and interacting with traffic foundation models. Transp. Policy 2024, 150, 95–105. [Google Scholar] [CrossRef]
- Wang, Q.; Qian, C.; Li, X.; Yao, Z.; Zhou, G.; Shao, H. Lens: A Foundation Model for Network Traffic. arXiv 2024, arXiv:2402.03646. [Google Scholar]
- Aceto, G.; Ciuonzo, D.; Montieri, A.; Persico, V.; Pescapé, A. AI-Powered Internet Traffic Classification: Past, Present, and Future. IEEE Commun. Mag. 2024, 62, 168–175. [Google Scholar] [CrossRef]
- Ying, X. An Overview of Overfitting and its Solutions. J. Phys. Conf. Ser. 2019, 1168, 022022. [Google Scholar] [CrossRef]
- Mienye, I.D.; Sun, Y. A Survey of Ensemble Learning: Concepts, Algorithms, Applications, and Prospects. IEEE Access 2022, 10, 99129–99149. [Google Scholar] [CrossRef]
- Zhang, X.; Liu, S.; Wang, X.; Li, Y. A fragmented neural network ensemble method and its application to image classification. Sci. Rep. 2024, 14, 2291. [Google Scholar] [CrossRef] [PubMed]
- Valenti, S.; Rossi, D.; Dainotti, A.; Pescapè, A.; Finamore, A.; Mellia, M. Reviewing Traffic Classification. In Data Traffic Monitoring and Analysis; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
- Wang, T.; Xie, X.; Wang, W.; Wang, C.; Zhao, Y.; Cui, Y. NetMamba: Efficient Network Traffic Classification via Pre-training Unidirectional Mamba. arXiv 2024, arXiv:2405.11449. [Google Scholar]
- Gabilondo, Á.; Fernández, Z.; Viola, R.; Martín, Á.; Zorrilla, M.; Angueira, P.; Montalbán, J. Traffic Classification for Network Slicing in Mobile Networks. Electronics 2022, 11, 1097. [Google Scholar] [CrossRef]
- Akbar, M.S.; Hussain, Z.; Ikram, M.; Sheng, Q.Z.; Mukhopadhyay, S.C. On challenges of sixth-generation (6G) wireless networks: A comprehensive survey of requirements, applications, and security issues. J. Netw. Comput. Appl. 2025, 233, 104040. [Google Scholar] [CrossRef]
- Possebon, I.P.; Silva, A.S.; Granville, L.Z.; Schaeffer-Filho, A.; Marnerides, A. Improved Network Traffic Classification Using Ensemble Learning. In Proceedings of the 2019 IEEE Symposium on Computers and Communications (ISCC), Barcelona, Spain, 29 June–3 July 2019; pp. 1–6. [Google Scholar]
- Shahraki, A.; Abbasi, M.; Taherkordi, A.; Kaosar, M. Internet Traffic Classification Using an Ensemble of Deep Convolutional Neural Networks. In Proceedings of the 4th FlexNets Workshop on Flexible Networks Artificial Intelligence Supported Network Flexibility and Agility (FlexNets ’21), Virtual Event, 23 August 2021; pp. 38–43. [Google Scholar]
- Aouedi, O.; Piamrat, K.; Parrein, B. Ensemble-Based Deep Learning Model for Network Traffic Classification. IEEE Trans. Netw. Serv. Manag. 2022, 19, 4124–4135. [Google Scholar] [CrossRef]
- Jang, E.; Gu, S.; Poole, B. Categorical Reparameterization with Gumbel-Softmax. arXiv 2017, arXiv:1611.01144. [Google Scholar]
- Maddison, C.J.; Mnih, A.; Teh, Y.W. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables. arXiv 2017, arXiv:1611.00712. [Google Scholar]
- Donà, J.; Gallinari, P. Differentiable Feature Selection, A Reparameterization Approach. In Machine Learning and Knowledge Discovery in Databases. Research Track; Springer: Cham, Switzerland, 2021; pp. 414–429. [Google Scholar]
- Strypsteen, T.; Bertrand, A. Conditional Gumbel-Softmax for constrained feature selection with application to node selection in wireless sensor networks. arXiv 2024, arXiv:2406.01162. [Google Scholar]
- Kim, S.; Kim, I.; Lim, S.; Baek, W.; Kim, C.; Cho, H.; Yoon, B.; Kim, T. Scalable Neural Architecture Search for 3D Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2019; Springer: Cham, Switzerland, 2019; pp. 220–228. [Google Scholar]
- Kotary, J.; Vito, V.D.; Fioretto, F. Differentiable Model Selection for Ensemble Learning. arXiv 2023, arXiv:2211.00251. [Google Scholar]
- Draper-Gil, G.; Lashkari, A.H.; Mamun, M.S.I.; Ghorbani, A.A. Characterization of Encrypted and VPN Traffic Using Time-Related Features. In Proceedings of the 2nd International Conference on Information Systems Security and Privacy (ICISSP 2016), Rome, Italy, 19–21 February 2016; pp. 407–414. [Google Scholar]
- Lashkari, A.H.; Draper-Gil, G.; Mamun, M.S.I.; Ghorbani, A.A. Characterization of Tor Traffic using Time based Features. In Proceedings of the 3rd International Conference on Information Systems Security and Privacy (ICISSP 2017), Porto, Portugal, 19–21 February 2017; pp. 253–262. [Google Scholar]
- Baek, U.-J.; Lee, M.-S.; Park, J.-T.; Choi, J.-W.; Shin, C.-Y.; Kim, M.-S. Preprocessing and Analysis of an Open Dataset in Application Traffic Classification. In Proceedings of the 2023 24st Asia-Pacific Network Operations and Management Symposium (APNOMS), Jeju Island, Republic of Korea, 13–15 September 2023; pp. 227–230. [Google Scholar]
Dataset | Publicly | #Task | #Applications | #Sessions | |
---|---|---|---|---|---|
Raw | Preprocessed | ||||
Private | N | 2 | 50 | 71,841 | 23,846 |
ISCX VPN 2016 [29] | Y | 3 | 23 | 187,336 | 10,011 |
ISCX Tor 2016 [30] | Y | 3 | 21 | 57,605 | 36,947 |
Model | Private | ISCX VPN 2016 | ISCX Tor 2016 | ||||||
---|---|---|---|---|---|---|---|---|---|
Accuracy | F1-Score | Inference Time * | Accuracy | F1-Score | Inference Time * | Accuracy | F1-Score | Inference Time * | |
2D-CNN | 61.6 | 60.8 | 113 | 64.9 | 64.6 | 47 | 96.5 | 96.3 | 175 |
1D-CNN | 59.5 | 58.3 | 113 | 66.1 | 65.8 | 47 | 96.6 | 96.3 | 175 |
HAST-1 | 92.5 | 92.5 | 1646 | 73.1 | 73.1 | 691 | 97.0 | 96.8 | 2550 |
HAST-2 | 91.9 | 91.9 | 2384 | 69.9 | 70.1 | 1000 | 96.4 | 96.1 | 3693 |
SAM | 92.0 | 92.0 | 146 | 71.9 | 71.9 | 61 | 94.2 | 93.8 | 226 |
XGB | 93.6 | 93.6 | 15 | 79.1 | 73.1 | 6 | 90.0 | 87.1 | 23 |
ET-BERT | 91.8 | 91.3 | 6663 | 70.9 | 70.9 | 2794 | 94.4 | 94.2 | 10,321 |
Hard vote | 87.0 | 87.1 | 2384 | 72.0 | 73.1 | 1000 | 97.7 | 97.5 | 3693 |
Kotary et al. | 93.1 | 93.1 | 1015 | 74.1 | 74.1 | 486 | 97.4 | 97.4 | 1423 |
Proposed | 94.3 | 94.0 | 908 | 79.2 | 79.1 | 371 | 98.0 | 97.9 | 1374 |
2D-CNN | 1D-CNN | HAST-1 | HAST-2 | SAM | |
---|---|---|---|---|---|
Pre-trained | 61.6 | 59.5 | 92.5 | 91.9 | 92 |
Proposed (Entire test dataset) | 67.1 | 65.9 | 91.5 | 91.5 | 91.3 |
Proposed (Only assigned dataset) | 93.94 | 100 | 99.3 | 77.3 | 93.2 |
#Models | Accuracy | ||||
---|---|---|---|---|---|
Homogeneous | Heterogeneous | ||||
2D-CNN | 1D-CNN | SAM | 2D-CNN 1D-CNN SAM | ALL | |
1 | 61.6 | 59.5 | 92.0 | 92.8 | 94.3 |
2 | 88.1 | 83.4 | 91.4 | ||
5 | 87.8 | 88.4 | 91.4 | ||
10 | 90.6 | 88.0 | 91.7 |
Items | Proposed | Bagging |
---|---|---|
Overlap between subsets | Implicit | Explicit |
Subset creation | Model selector | Random |
Aggregation strategy | Winner takes all Model selector | Vote, averaging |
Out-of-bag strategy | Absent | Present |
Number of subsets | Variable | Fixed |
Number of divisions per subset | Variable | Fixed |
Number of features | Fixed | Variable |
Inference time | Balanced | Inefficient |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Baek, U.-J.; Jang, Y.-S.; Kim, J.-S.; Choi, Y.-S.; Kim, M.-S. Deep Ensemble Learning for Application Traffic Classification Using Differential Model Selection Technique. Sensors 2025, 25, 2853. https://doi.org/10.3390/s25092853
Baek U-J, Jang Y-S, Kim J-S, Choi Y-S, Kim M-S. Deep Ensemble Learning for Application Traffic Classification Using Differential Model Selection Technique. Sensors. 2025; 25(9):2853. https://doi.org/10.3390/s25092853
Chicago/Turabian StyleBaek, Ui-Jun, Yoon-Seong Jang, Ju-Sung Kim, Yang-Seo Choi, and Myung-Sup Kim. 2025. "Deep Ensemble Learning for Application Traffic Classification Using Differential Model Selection Technique" Sensors 25, no. 9: 2853. https://doi.org/10.3390/s25092853
APA StyleBaek, U.-J., Jang, Y.-S., Kim, J.-S., Choi, Y.-S., & Kim, M.-S. (2025). Deep Ensemble Learning for Application Traffic Classification Using Differential Model Selection Technique. Sensors, 25(9), 2853. https://doi.org/10.3390/s25092853