Next Article in Journal
X-ray Spectroscopic Studies of a Solid-Density Germanium Plasma Created by a Free Electron Laser
Previous Article in Journal
Understanding Biotic Stress and Hormone Signalling in Cassava (Manihot esculenta): Potential for Using Hyphenated Analytical Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GPON PLOAMd Message Analysis Using Supervised Neural Networks

Department of Telecommunication, Brno University of Technology, Technicka 12, 616 00 Brno, Czech Republic
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(22), 8139; https://doi.org/10.3390/app10228139
Submission received: 6 September 2020 / Revised: 4 November 2020 / Accepted: 13 November 2020 / Published: 18 November 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
This paper discusses the possibility of analyzing the orchestration protocol used in gigabit-capable passive optical networks (GPONs). Considering the fact that a GPON is defined by the International Telecommunication Union Telecommunication sector (ITU-T) as a set of recommendations, implementation across device vendors might exhibit few differences, which complicates analysis of such protocols. Therefore, machine learning techniques are used (e.g., neural networks) to evaluate differences in GPONs among various device vendors. As a result, this paper compares three neural network models based on different types of recurrent cells and discusses their suitability for such analysis.

1. Introduction

Customer requirements for provided services increase over time, which also increases requirements for provided bandwidth [1,2]. A modern solution to meet customer needs is the gigabit-capable passive optical network (GPON). Another modern solution is passive optical network (PON) technology [3,4,5,6,7,8,9,10,11,12]. The protocol used in these networks is defined by the International Telecommunication Union Telecommunication sector (ITU-T) as a set of recommendations, which allows device vendors to modify this protocol with little adjustment [13,14]. An analysis of such protocols requires advanced techniques because conventional (e.g., pushdown automata) methods are not sufficient for undocumented messages or message usages. Therefore, this paper investigates the possibility of GPON protocol analysis using machine learning algorithms.
A passive optical network exhibits a tree-like topology, where the root is the optical line terminal (OLT) located at the provider site and the leaves are optical network units (ONUs) or optical network terminals (ONTs) located at the customer sites. The distributional network between the OLT and ONUs/ONTs consists of fiber cables and passive optical components (e.g., splitters, splices, and connectors). The OLT provides organization and management for components of this network, especially for effective bandwidth allocation, transmit synchronization and others [15,16,17,18]. Passive optical components cause the data to be sent using broadcasting for downstream communication (from the root to leaves). Each communication participant must be correctly synchronized to avoid transmission collisions, which is achieved by forcing leaves to use the activation process when they want to join the network [19]. This process uses physical layer operations, administration and maintenance downstream (PLOAMd) messages located in the header of the downstream transmission frame.
The communication medium (optical fiber) and GPON maximal symmetrical transmission speed of 2.48832 Gbps make the data capture process challenging. The downstream data are mirrored using an optical splitter into a network probe consisting of a hardware accelerated signal decoder [20]. The captured data are sent to a message database, which is the source for the learning data used in this paper.
The main contribution of this paper is a comparison of three different types of recurrent neural network layer used for the analysis of PLOAMd message sequences using supervised machine learning methods. All of these layer types are tested in the exactly the same network architecture to achieve reasonable performance comparison. All of the models share the same optimizer and objective function. The only difference is recurrent layer type. The tested models should be able to recognize patterns in communication sequences and decide whether the communication satisfies the GPON standard. Thus, the models classify input sequence into two categories: a correct class representing communication according to the GPON recommendation and a faulty class representing a violation of GPON recommendation. Exemplary GPON communication is captured in a real GPON network in a laboratory environment. This communication is considered as suitable for the GPON standard and is used as the source of learning data. A non-standard communication sequences are generated by custom corrupting function, which add/drop/modify messages in the captured sequences. The output of this paper is comparison of various recurrent layers and recommended recurrent layer for protocol analysis capable of learning patterns in message sequences.
The rest of this paper is structured as follows. Section 2 introduces related works. The learning data generation process and its final data shape are described in Section 3. Section 4 introduces the model design based on long short-term memory (LSTM) cells and evaluates its detection capabilities (accuracy). Finally, Section 5 concludes this paper.

2. Related Work

The GPON protocol is a following protocol, which replaces old technologies providing services such as internet connectivity, cable television (CATV), and others. Passive optical networks are still under research and development, and, hence, experiments and protocol upgrades are being conducted [21]. The newest version of this protocol family is next-generation passive optical network stage 2 (NG-PON2), which is capable of symmetrical 40 Gbps transfer owing to time and wavelength division multiplexing (TWDM) [22]. There is also an experimental study attempting to extend a long-range GPON with dense wavelength division multiplexing (DWDM) over 135 km [23]. The ITU-T also has a competitor in the PON standard, which is defined by the Institute of Electrical and Electronics Engineers (IEEE) as the Ethernet passive optical network (EPON), which uses standard Ethernet frames instead of frames based on asynchronous transfer mode (ATM) cells [24]. The newest research defines EPON standards with bitrates up to 100 Gbps [25,26].
Artificial intelligence (AI) and machine learning (ML) algorithms are used to solve various problems of current networks [27,28,29,30,31,32,33,34,35,36,37]. In a given location, the minimal GPON deployment cost is achieved by choosing optimal routing information, type of aggregating equipment and position of network elements. This is an NP-hard problem, and the optimal solution can be approximated using ant colony optimization (ACO) [38]. Traffic analysis using current methods (e.g., port-based and payload-based) shows various disadvantages, but the analysis using ML methods could be a potential replacement, even for encrypted data traffic [39,40]. ML may also help with parameter tuning for a GPON rule-based expert system used for the problem of root cause localization [41].
The authors of [42,43] presented an artificial intelligence modeling for future and intelligent elastic optical networks (EONs) as a review. We can also mention the article by Musumeci et al. [44], which provides a comprehensive study about machine learning for optical networks. Yi et al. [45] proposed the novel machine learning-based equalization algorithm for intensity modulation and direct detection PON (IM-DD PON) and extended the capacity of PON from 50 Gbps per λ to 100 Gbps per λ over 25 km standard single-mode fiber (SSMF). Further, Yi et al. [46] proposed using neural network (NN)-based equalizer to mitigate both the linear and nonlinear distortions in IM-DD 100 Gbps per λ PON based on 20G-class optical and electrical devices.
Neural networks based on various types of recurrent layers help to solve difficult problems (e.g., speech recognition, natural language processing, sequence analysis, etc.). Therefore, researchers investigate performance differences of these recurrent layers in certain applications. Mangal et al. [47] evaluated the performance of a bidirectional a recurrent neural network (BiRNN), a gated recurrent unit (GRU), and a LSTM used for TV script generation. The results of this paper show the best performance of all tested layers has the LSTM layer. Similarly, Apaydin et al. [48] evaluated the performance of various recurrent layers for a reservoir flow forecast, and the LSTM cells show the most accurate predictions. On the other hand, Shewalkar et al. [49] evaluated various recurrent layers for speech recognition and proposes a GRU-based network as the best. The network has slightly worse accuracy compared to the LSTM-based network, but the number of parameters in the GRU is much lower making GRU layer preferable for speech recognition.

3. Data Characteristics

The training data are captured in the GPON laboratory environment. Frames are captured using a special field-programmable gate array (FPGA) card, correctly parsed, and stored in a database. A detailed explanation of this process is outside of the scope of this paper. This section describes the GPON downstream frame header in Section 3.1, the process of data preparation in Section 3.2, the data shape in Section 3.3), and the preparation of invalid data sequences needed for supervised learning in Section 3.4.

3.1. GPON Frame Header

The frame structure sent by the OLT consists of a physical control block downstream (PCBd) containing information for specific ONUs/ONTs and a GPON transmission container (GTC) payload carrying transferred data. The frame structure is shown in Figure 1. The most interesting part is the PLOAMd field, which is used during the ONU/ONT activation process. The other parts are irrelevant in this paper. The PLOAMd field is 13 bytes long and contains the following parts:
  • ONU ID specifies the receiving ONU.
  • Message ID defines the type of message.
  • DATA are unique data for specific message.
  • Cyclic redundancy check (CRC) is used to verify the correctness of field data.

3.2. Raw Data Preparation

The first step of the data analysis process is to count occurrences of each message type in the captured communication. Surprisingly, the most used message type is NoMessage11, which creates 99.98% of captured traffic and does not contain any useful data. This type of message must be filtered out from the learning data to avoid false positive results. The number of occurrences for each message type is shown in Table 1.
The remaining data are processed using several procedures resulting in a sequence of PLOAMd messages. These procedures remove the useless parts of the frame header and expand PLOAMd data fields into a suitable form.

3.3. Data Shape

Each PLOAMd message is 13 bytes long, resulting in a learning vector of 13 elements (each field contains separate elements, and 10 bytes of the data fields are expanded). The header is shown in Figure 1. As mentioned above, the protocol analysis is inspired by natural language processing, which analyzes a sequence of words. In this scenario, words are replaced by vectors of 13 elements and shaped into a sequence of 30 vectors. The sequence length is empirically evaluated as the optimal length considering the ability to locate a possible non-defined sequence and to distinguish sequences. Final learning samples are generated by using a sliding time window over captured messages and returning sequences of correct shape (30 messages × 13 features).

3.4. Invalid Data Generation

Supervised learning algorithms require a balanced number of samples for each classification set. The model described by this paper classifies message sequences into a set that satisfies the GPON standard and a set that does not. The standard communication samples are generated using procedures described in the previous section. The non-standard sequence samples are generated using three procedures that knowingly corrupt standard sequences (the input of these procedures is the entire captured communication). The first corruption procedure changes values in random PLOAMd messages using Equation (1).
V a l u e n e w = ( V a l u e o l d + 128 ) mod 255
The second procedure removes important message types from the activation process. The activation process consists of Messages 4, 8, 10, and 18, in this exact order (a more detailed explanation of these messages can be found in GPON recommendation [50]). The procedure runs for each message on a distinct instance of the captured dataset separately and deletes all occurrences of that message. The third procedure creates faulty messages by adding zero-mean Gaussian noise with standard deviation equal to 20 into the captured communication. The last procedure simulates modification of the GPON protocol by a device vendor. This procedure duplicates the important messages of activation process mentioned earlier. Each message is duplicated in a separate instance of the captured message list.
All of the mentioned procedures are applied on separate instances of a filtered PLOAMd message sequence. The sliding window used for sequence sample generation runs through the faulty sequences with stride equal to 4. As a result, the number of generated faulty samples is almost the same as the number of correct sequence samples.

4. PLOAMd Analysis Using Various Recurrent Neural Networks and Results Discussion

Considering the shape of the data samples, the suitable neural network for protocol analysis is based on recurrent neural networks (RNNs). They pass their internal state across the analyzing sequence [51]. RNNs are also able to analyze variably long sequence samples. RNN layers used in following experiments are: SimpleRNN, LSTM, and GRU. These are the most common RNNs, which have different tools for sequence analysis, and all of them are implemented by Keras application programming interface (API). A simple RNN might be insufficient for analysis of long sequences, and, therefore, this paper verifies various types of RNNs (e.g., LSTM and GRU). LSTM enhances an RNN with an internal memory capable of holding information in longer sequences and a gating mechanism suitable for input/output data manipulation [52]. GRU is similar to an LSTM cell, but less complex and with fewer gates resulting in fewer learning parameters [53].
All experimental models classify input message sequences into two classes. Thus, all models are binary classifiers. The first class represents a message sequence that keeps the rules of GPON protocol. The second class represents an abnormal or non-standard sequence. The final learned model is used for abnormal or faulty traffic identification in a real GPON network.

4.1. Model Design

A generic model architecture is necessary for the comparison of various RNN types, especially for an evaluation in the exactly the same use case. The first proposed model is based on a single RNN layer with 32 neurons and a single dense layer with 1 neuron. The model is not learning properly and cannot distinguish correct and faulty data samples, which leads to poor model accuracy. The second proposed model evolves from the first model by doubling the number of both layers and adding more neurons into the first dense layer in the row. This model has much better learning capabilities. Therefore, the second model architecture is chosen as a generic model for further comparison of all inspected RNN layer types. The generic model architecture used for all RNN types is shown in Listing 1. These models are learned using Adam [54] optimizer and binary cross entropy loss function.
Listing 1: Generation process of GPON analyzing models.
  • from tensorflow.keras.layers import SimpleRNN, GRU, LSTM, Dense
  • from tensroflow.keras import Sequential
  • rnn_types = [SimpleRNN, GRU, LSTM]
  • rnn_models = []
  • for rnn in rnn_types:
  • rnn_models.append(Sequential([
  • rnn(32,input_shape=data[0].shape,
  • return_sequences=True, activation=’tanh’),
  • rnn(32, activation=’tanh’),
  • Dense(64, activation=’relu’),
  • Dense(1, activation=’sigmoid’)])
  • )

4.2. Learning Process

The learning data preparation is described in Section 3. The dataset is shuffled with predefined constant seed to guarantee the same data order every time. Ten-fold cross-validation is used to evaluate performance of models correctly [55]. This process splits the dataset into ten equally big groups. Each group is used as the validation dataset for a separate instance of the inspected model, while the remaining groups are used as the learning dataset. Based on the training history of all models, the average and standard deviation are evaluated for each model metric.
The design model input is a three-dimensional array because it is a list of sequences. Each model is learned using the Adam optimizer and updated every 50 sequences (i.e., the batch size is set to 50). Based on the learning history, it is estimated whether the model is capable of pattern recognition. The learning process of each model takes 200 epochs (the number of epochs is high enough to evaluate the accuracy and the loss in a under-fitted, good fit, and over-fitted model) to check the loss and accuracy evolution over learning epochs and whether the learning and validation loss increase, decrease, or remain constant.
The model with RNN layers is able to recognize learning samples very well, but it presents a problem in learning generalized patterns in the sequences, as shown by the validation loss and validation accuracy in Figure 2. The validation accuracy slowly raises and after few epoch it remains stable at accuracy of 74%. A rapid drop in the learning loss indicates that the model memorizes the input without generalized patterns, thus the validation loss raises slightly over epochs.
The model with LSTM layers exhibits sufficient accuracy in distinguishing standard and non-standard/corrupted data sequences for both the learning and validation datasets, as shown in Figure 3. The LSTM layers are able to learn patterns in the communication and decide whether they meet the GPON standard. On the other hand, this model is the most complex among the tested models and requires the most resources to learn.
The last evaluated model is based on layers with GRU cells. Similar to LSTM, the model starts to be perfectly fitted after approximately 75 learning epochs. More interestingly, the model accuracy for the validation dataset is slightly higher compared to the LSTM model, as shown in Figure 4.

4.3. Evaluation of Models

All of the models learn to distinguish correct and faulty samples of the training dataset, which is clearly visible in average loss (the model loss represents output of objective function used during learning process) and average accuracy (the model accuracy represents number of correctly classified samples compared to all samples) evolution over learning epochs by blue solid line in Figure 2, Figure 3 and Figure 4. In the figures, there is also the standard deviation for the average values represented by transparent fill in the same color. All inspected models have similar performance according to model metrics based on training dataset. Considering the standard deviation, it is difficult to choose one as the best.
The quality of learned capabilities is evaluated by classifying the validation dataset, which is disjunctive to the learning dataset. The evolution of accuracy and loss over the epochs is visible by orange dotted line in Figure 2, Figure 3 and Figure 4. The model based on SimpleRNN layer finds patterns in the data much more slowly compared to LSTM- or GRU-based models. This behavior is expected due to RNN being the simplest recurrent architecture and having fewer controlling instruments. All SimpleRNN model metrics in validation dataset indicate that the model has the worst performance of the all inspected models.
The LSTM-based model has better accuracy and the other metrics compared to SimpleRNN. This model finds and learns correct characteristic patterns. However, this model has the largest number of arguments, thus it is resource demanding and the learning process take longer to proceed.
The GRU-based model complexity is in between LSTM and SimpleRNN layers, but the validation dataset accuracy and other metrics point out that this model is the best of the inspected models. It is able to find distinguishing patterns better than SimpleRNN- and LSTM-based models. The GRU-based model has fewer gates (and learning parameters as well) and the learning process takes less time, even though it has the best performance according to validation dataset metrics. The training results and a summary of all model metrics are shown in Table 2.

5. Conclusions

This paper evaluates various types of recurrent layers used for GPON protocol analysis. The learning data are captured in a fully functional GPON using a special FPGA card. The card captures whole frames, but analysis focuses only on PLOAMd messages; therefore, useless data are removed during the pre-processing phase. Almost all filtered data consist of messages with ID 11, which must be removed from the learning dataset. Otherwise, it causes false positive results. The remaining data are formatted into sequences of thirty messages using a sliding window, creating a learning dataset. Created sequences are used to evaluate the analysis performance of SimpleRNN-, LSTM-, and GRU-based models. The model comparison is based on validation dataset accuracy to reveal the pattern learning potential of each layer type.
SimpleRNN is insufficient for such analysis due to it simply memorizing data instead of learning generalized principles. A single state vector without gating mechanisms in RNN design makes this layer not suitable for such analysis. The model with LSTM cells has almost the best accuracy for the learning and validation datasets; however, it consumes the most resources among all inspected models (the highest number of learned parameters). The GRU-based model has the best results in validation dataset classification compared to SimpleRNN. The GRU cell contains gating mechanisms as well, but the number of gates is lower compared to the LSTM cell. Even with fewer gates, GRU is able to learn complex data patterns specific to protocol analysis.
Among all inspected models, only the model based on GRU cells is suitable for network protocol analysis. The investigated data contains complex relations and the GRU-based model is capable to learn these relations using its sophisticated gating mechanism. Considering the standard deviation, performance comparison results correlate with the similar comparisons in [47,48,49]. Therefore, GRU layer is recommended for protocol analysis, even if it has more learning parameters and takes more time to learn.

Future Work Opportunities

This paper is focused strictly on three types of recurrent cells with supervised learning, but ML offers much more. There are some possibilities for further research on protocol analysis using neural networks:
  • Convolutional network: One possible enhancement is replacing the recurrent layer with a sequence of convolutional and max-pooling layers, which should reduce the complexity of the final neural network and speed up the learning and evaluating process.
  • Unsupervised learning techniques: Another possible continuation of this work is to create a machine learning model and train it using unsupervised learning techniques. The problem with the supervised algorithms used in this paper is that the model needs samples of non-standard or corrupted communication, but these samples may not exist during the model learning process. The result of this project would be an anomaly detector capable of marking suspicious or non-standard communication sequences.

Author Contributions

Conceptualization, A.T., M.H., V.O., T.H., and P.M.; methodology, A.T., M.H., and V.O.; software, A.T.; validation, A.T., M.H., and V.O.; formal analysis, T.H. and P.M.; investigation, A.T., M.H., and V.O.; resources, V.O., T.H., and P.M.; data curation, A.T.; writing—original draft preparation, A.T., M.H., V.O., T.H., and P.M.; and visualization, T.H. and P.M.; supervision, M.H. and V.O. All authors have read and agreed to the published version of the manuscript.

Funding

The research described in this paper was financed by a grant from the Ministry of the Interior of the Czech Republic, Program of Security Research, VI20192022135, PID VI3VS/746 for “Deep hardware detection of network traffic of next generation passive optical network in critical infrastructures”.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACOAnt colony optimization
AIArtificial intelligence
APIApplication programming interface
ATMAsynchronous transfer mode
BiRNNBidirectional a recurrent neural network
CATVCable television
CRCCyclic redundancy check
DWDMDense wavelength division multiplexing
EONsElastic optical networks
EPONEthernet passive optical network
FPGAField-programmable gate array
GPONGigabit-capable passive optical network
GRUGated recurrent unit
GTCGigabit-capable passive optical network transmission container
IEEEInstitute of Electrical and Electronics Engineers
IM-DD PONIntensity modulation and direct detection passive optical network
ITUInternational Telecommunication Union
ITU–TITU Telecommunication Standardization Sector
LSTMLong short-term memory
MLMachine learning
NG-PON2Next-generation passive optical network stage 2
NNNeural network
PCBdPhysical control block downstream
PLOAMdPhysical layer operations, administration and maintenance downstream
PONPassive optical network
RNNRecurrent neural network
SSMFStandard single-mode fiber
OLTOptical line termination
ONTOptical network termination
ONUOptical network unit
TWDMTime and wavelength division multiplexing

References

  1. Ford, G.S. Is faster better? Quantifying the relationship between broadband speed and economic growth. Telecommun. Policy 2018, 42, 766–777. [Google Scholar] [CrossRef]
  2. Hernandez, J.A.; Sanchez, R.; Martin, I.; Larrabeiti, D. Meeting the Traffic Requirements of Residential Users in the Next Decade with Current FTTH Standards. IEEE Commun. Mag. 2019, 57, 120–125. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Liu, B.; Ren, J.; Wu, X.; Wang, X. Flexible Probabilistic Shaping PON Based on Ladder-Type Probability Model. IEEE Access 2020, 8, 34170–34176. [Google Scholar] [CrossRef]
  4. Li, S.; Cheng, M.; Chen, Y.; Fan, C.; Deng, L.; Zhang, M.; Fu, S.; Tang, M.; Shum, P.P.; Liu, D. Enhancing the Physical Layer Security of OFDM-PONs With Hardware Fingerprint Authentication. J. Light. Technol. 2020, 38, 3238–3245. [Google Scholar] [CrossRef]
  5. Houtsma, V.E.; Veen, D.T.V. Investigation of Modulation Schemes for Flexible Line-Rate High-Speed TDM-PON. J. Light. Technol. 2020, 38, 3261–3267. [Google Scholar] [CrossRef]
  6. Mikaeil, A.; Hu, W.; Hussain, S.; Sultan, A. Traffic-Estimation-Based Low-Latency XGS-PON Mobile Front-Haul for Small-Cell C-RAN Based on an Adaptive Learning Neural Network. Appl. Sci. 2018, 8, 1097. [Google Scholar] [CrossRef] [Green Version]
  7. Harstead, E.; Bonk, R.; Walklin, S.; van Veen, D.; Houtsma, V.; Kaneda, N.; Mahadevan, A.; Borkowski, R. From 25 Gb/s to 50 Gb/s TDM PON. J. Opt. Commun. Netw. 2020, 12, D17–D26. [Google Scholar] [CrossRef]
  8. Ichi Kani, J.; Terada, J.; Hatano, T.; Kim, S.Y.; Asaka, K.; Yamada, T. Future optical access network enabled by modularization and softwarization of access and transmission functions [Invited]. J. Opt. Commun. Netw. 2020, 12, D48–D56. [Google Scholar] [CrossRef]
  9. DeSanti, C.; Du, L.; Guarin, J.; Bone, J.; Lam, C.F. Super-PON. J. Opt. Commun. Netw. 2020, 12, D66–D77. [Google Scholar] [CrossRef]
  10. Zou, J.S.; Sasu, S.A.; Lawin, M.; Dochhan, A.; Elbers, J.P.; Eiselt, M. Advanced optical access technologies for next-generation (5G) mobile networks [Invited]. J. Opt. Commun. Netw. 2020, 12, D86–D98. [Google Scholar] [CrossRef]
  11. Zhang, D.; Liu, D.; Wu, X.; Nesset, D. Progress of ITU-T higher speed passive optical network (50G-PON) standardization. J. Opt. Commun. Netw. 2020, 12, D99–D108. [Google Scholar] [CrossRef]
  12. Das, S.; Slyne, F.; Kaszubowska, A.; Ruffini, M. Virtualized EAST–WEST PON architecture supporting low-latency communication for mobile functional split based on multiaccess edge computing. J. Opt. Commun. Netw. 2020, 12, D109–D119. [Google Scholar] [CrossRef]
  13. Li, C.; Guo, W.; Wang, W.; Hu, W.; Xia, M. Bandwidth Resource Sharing on the XG-PON Transmission Convergence Layer in a Multi-operator Scenario. J. Opt. Commun. Netw. 2016, 8, 835–843. [Google Scholar] [CrossRef]
  14. Luo, Y.; Effenberger, F.; Gao, B. Transmission convergence layer framing in XG-PON1. In 2009 IEEE Sarnoff Symposium; IEEE: Princeton, NJ, USA, 2009; pp. 1–5. [Google Scholar] [CrossRef]
  15. Kyriakopoulos, C.A.; Papadimitriou, G.I. Predicting and allocating bandwidth in the optical access architecture XG-PON. Opt. Switch. Netw. 2017, 25, 91–99. [Google Scholar] [CrossRef]
  16. Uzawa, H.; Honda, K.; Nakamura, H.; Hirano, Y.; ichi Nakura, K.; Kozaki, S.; Terada, J. Dynamic bandwidth allocation scheme for network-slicing-based TDM-PON toward the beyond-5G era. J. Opt. Commun. Netw. 2020, 12, A135–A143. [Google Scholar] [CrossRef]
  17. Skubic, B.; Chen, J.; Ahmed, J.; Chen, B.; Wosinska, L.; Mukherjee, B. Dynamic bandwidth allocation for long-reach PON. IEEE Commun. Mag. 2010, 48, 100–108. [Google Scholar] [CrossRef]
  18. Skubic, B.; Chen, J.; Ahmed, J.; Wosinska, L.; Mukherjee, B. A comparison of dynamic bandwidth allocation for EPON, GPON, and next-generation TDM PON. IEEE Commun. Mag. 2009, 47, S40–S48. [Google Scholar] [CrossRef]
  19. Horvath, T.; Munster, P.; Bao, N.H. Lasers in Passive Optical Networks and the Activation Process of an End Unit. Electronics 2020, 9, 1114. [Google Scholar] [CrossRef]
  20. Oujezsky, V.; Horvath, T.; Jurcik, M.; Skorpil, V.; Holik, M.; Kvas, M. Fpga Network Card And System For Gpon Frames Analysis At Optical Layer. In 2019 42nd International Conference on Telecommunications and Signal Processing (TSP); IEEE: Budapest, Hungary, 2019; pp. 19–23. [Google Scholar] [CrossRef]
  21. Satyanarayana, K.; Abhinov, B. Recent trends in future proof fiber access passive networks: GPON and WDM PON. In Proceedings of the 2014 International Conference on Recent Trends in Information Technology, Chennai, India, 10–12 April 2014; pp. 1–5. [Google Scholar]
  22. Nesset, D. NG-PON2 Technology and Standards. J. Light. Technol. 2015, 33, 1136–1143. [Google Scholar] [CrossRef]
  23. Cale, I.; Salihovic, A.; Ivekovic, M. Gigabit Passive Optical Network—GPON. In Proceedings of the 2007 29th International Conference on Information Technology Interfaces, Cavtat/Dubrovnik, Croatia, 25–28 June 2007; pp. 679–684. [Google Scholar]
  24. Kramer, G.; Pesavento, G. Ethernet passive optical network (EPON): Building a next-generation optical access network. IEEE Commun. Mag. 2002, 40, 66–73. [Google Scholar] [CrossRef] [Green Version]
  25. Houtsma, V.; van Veen, D.; Harstead, E. Recent Progress on Standardization of Next-Generation 25, 50, and 100G EPON. J. Light. Technol. 2017, 35, 1228–1234. [Google Scholar] [CrossRef]
  26. Knittle, C. IEEE 100 Gb/s EPON. In Proceedings of the Optical Fiber Communication Conference, Anaheim, CA, USA, 20–24 March 2016; Optical Society of America: Anaheim, CA, USA; p. Th1I.6. [Google Scholar] [CrossRef]
  27. Fadlullah, Z.M.; Tang, F.; Mao, B.; Kato, N.; Akashi, O.; Inoue, T.; Mizutani, K. State-of-the-Art Deep Learning. IEEE Commun. Surv. Tutorials 2017, 19, 2432–2455. [Google Scholar] [CrossRef]
  28. Liu, N.; Wang, H. Ensemble Based Extreme Learning Machine. IEEE Signal Process. Lett. 2010, 17, 754–757. [Google Scholar] [CrossRef]
  29. Srinivasan, S.M.; Truong-Huu, T.; Gurusamy, M. Machine Learning-Based Link Fault Identification and Localization in Complex Networks. IEEE Internet Things J. 2019, 6, 6556–6566. [Google Scholar] [CrossRef] [Green Version]
  30. Singh, S.K.; Jukan, A. Machine-Learning-Based Prediction for Resource (Re)allocation in Optical Data Center Networks. J. Opt. Commun. Netw. 2018, 10, D12–D28. [Google Scholar] [CrossRef]
  31. Morais, R.M.; Pedro, J. Machine Learning Models for Estimating Quality of Transmission in DWDM Networks. J. Opt. Commun. Netw. 2018, 10, D84–D99. [Google Scholar] [CrossRef]
  32. Côté, D. Using Machine Learning in Communication Networks [Invited]. J. Opt. Commun. Netw. 2018, 10, D100–D109. [Google Scholar] [CrossRef]
  33. Li, B.; Lu, W.; Liu, S.; Zhu, Z. Deep-Learning-Assisted Network Orchestration for On-Demand and Cost-Effective vNF Service Chaining in Inter-DC Elastic Optical Networks. J. Opt. Commun. Netw. 2018, 10, D29–D41. [Google Scholar] [CrossRef]
  34. Zhao, Y.; Li, Y.; Zhang, X.; Geng, G.; Zhang, W.; Sun, Y. A Survey of Networking Applications Applying the Software Defined Networking Concept Based on Machine Learning. IEEE Access 2019, 7, 95397–95417. [Google Scholar] [CrossRef]
  35. Kibria, M.G.; Nguyen, K.; Villardi, G.P.; Zhao, O.; Ishizu, K.; Kojima, F. Big Data Analytics, Machine Learning, and Artificial Intelligence in Next-Generation Wireless Networks. IEEE Access 2018, 6, 32328–32338. [Google Scholar] [CrossRef]
  36. Wang, X.; Li, X.; Leung, V.C.M. Artificial Intelligence-Based Techniques for Emerging Heterogeneous Network. IEEE Access 2015, 3, 1379–1391. [Google Scholar] [CrossRef]
  37. Rafique, D.; Velasco, L. Machine Learning for Network Automation. J. Opt. Commun. Netw. 2018, 10, D126–D143. [Google Scholar] [CrossRef] [Green Version]
  38. Chu, A.; Poon, K.F.; Ouali, A. Using Ant Colony Optimization to design GPON-FTTH networks with aggregating equipment. In Proceedings of the 2013 IEEE Symposium on Computational Intelligence for Communication Systems and Networks (CIComms), Singapore, 16–19 April 2013; pp. 10–17. [Google Scholar] [CrossRef]
  39. Liu, Y.; Li, W.; Li, Y. Network Traffic Classification Using K-means Clustering. In Proceedings of the Second International Multi-Symposiums on Computer and Computational Sciences (IMSCCS 2007), Iowa City, IA, USA, 13–15 August 2007; pp. 360–365. [Google Scholar]
  40. Arndt, D.J.; Zincir-Heywood, A.N. A Comparison of three machine learning techniques for encrypted network traffic analysis. In Proceedings of the 2011 IEEE Symposium on Computational Intelligence for Security and Defense Applications (CISDA), Paris, France, 11–15 April 2011; pp. 107–114. [Google Scholar]
  41. Gosselin, S.; Courant, J.; Tembo, S.R.; Vaton, S. Application of probabilistic modeling and machine learning to the diagnosis of FTTH GPON networks. In Proceedings of the 2017 International Conference on Optical Network Design and Modeling (ONDM), Budapest, Hungary, 15–18 May 2017; pp. 1–3. [Google Scholar]
  42. Liu, X.; Lun, H.; Fu, M.; Fan, Y.; Yi, L.; Hu, W.; Zhuge, Q. AI-Based Modeling and Monitoring Techniques for Future Intelligent Elastic Optical Networks. Appl. Sci. 2020, 10, 363. [Google Scholar] [CrossRef] [Green Version]
  43. Mata, J.; de Miguel, I.; Durán, R.J.; Merayo, N.; Singh, S.K.; Jukan, A.; Chamania, M. Artificial intelligence (AI) methods in optical networks. Opt. Switch. Netw. 2018, 28, 43–57. [Google Scholar] [CrossRef]
  44. Musumeci, F.; Rottondi, C.; Nag, A.; Macaluso, I.; Zibar, D.; Ruffini, M.; Tornatore, M. An Overview on Application of Machine Learning Techniques in Optical Networks. IEEE Commun. Surv. Tutorials 2019, 21, 1383–1408. [Google Scholar] [CrossRef] [Green Version]
  45. Yi, L.; Li, P.; Liao, T.; Hu, W. 100 Gb/s/λ IM-DD PON Using 20G-Class Optical Devices by Machine Learning Based Equalization. In 2018 European Conference on Optical Communication (ECOC); IEEE: Rome, Italy, 2018; pp. 1–3. [Google Scholar] [CrossRef]
  46. Yi, L.; Liao, T.; Huang, L.; Xue, L.; Li, P.; Hu, W. Machine Learning for 100 Gb/s/λ Passive Optical Network. J. Light. Technol. 2019, 37, 1621–1630. [Google Scholar] [CrossRef]
  47. Mangal, S.; Joshi, P.; Modak, R. LSTM vs. GRU vs. Bidirectional RNN for script generation. arXiv 2019, arXiv:1908.04332. [Google Scholar]
  48. Apaydin, H.; Feizi, H.; Sattari, M.T.; Colak, M.S.; Shamshirband, S.; Chau, K.W. Comparative Analysis of Recurrent Neural Network Architectures for Reservoir Inflow Forecasting. Water 2020, 12, 1500. [Google Scholar] [CrossRef]
  49. Shewalkar, A.; Nyavanandi, D.; Ludwig, S.A. Performance Evaluation of Deep Neural Networks Applied to Speech Recognition: RNN, LSTM and GRU. J. Artif. Intell. Soft Comput. Res. 2019, 9, 235–245. [Google Scholar] [CrossRef] [Green Version]
  50. International Telecommunication Union. Gigabit-Capable Passive Optical Networks (G-PON): Transmission Convergence Layer Specification; International Telecommunication Union: Geneva, Switzerland, 2014. [Google Scholar]
  51. Boden, M. A guide to recurrent neural networks and backpropagation. Dallas Proj. 2002, 1–10. Available online: https://www.researchgate.net/publication/2903062_A_Guide_to_Recurrent_Neural_Networks_and_Backpropagation (accessed on 17 November 2020).
  52. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  53. Dey, R.; Salem, F.M. Gate-variants of Gated Recurrent Unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600. [Google Scholar]
  54. Zhang, Z. Improved Adam Optimizer for Deep Neural Networks. In Proceedings of the 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS), Banff, AB, Canada, 4–6 June 2018; pp. 1–2. [Google Scholar]
  55. Blum, A.; Kalai, A.; Langford, J. Beating the hold-out: Bounds for k-fold and progressive cross-validation. In Proceedings of the Twelfth Annual Conference on Computational Learning Theory, Santa Cruz, CA, USA, 6–9 July 1999; pp. 203–208. [Google Scholar]
Figure 1. GPON downstream frame structure.
Figure 1. GPON downstream frame structure.
Applsci 10 08139 g001
Figure 2. Average loss and accuracy of the models with an RNN over learning epochs.
Figure 2. Average loss and accuracy of the models with an RNN over learning epochs.
Applsci 10 08139 g002
Figure 3. Average loss and accuracy of the models with LSTM over learning epochs.
Figure 3. Average loss and accuracy of the models with LSTM over learning epochs.
Applsci 10 08139 g003
Figure 4. Average loss and accuracy of the models with GRU over learning epochs.
Figure 4. Average loss and accuracy of the models with GRU over learning epochs.
Applsci 10 08139 g004
Table 1. Filtered PLOAMd message counts from captured data traffic.
Table 1. Filtered PLOAMd message counts from captured data traffic.
Message ID13456810111418202124
Count8288114945428686545,9349310363541
Table 2. Recurrent models mean metrics with standard deviation.
Table 2. Recurrent models mean metrics with standard deviation.
SimpleRNNLSTMGRU
Learningloss 0.04897 ± 0.00470 0.11183 ± 0.05380 0.06980 ± 0.01116
accuracy 0.97088 ± 0.00318 0.95212 ± 0.01856 0.96655 ± 0.00392
recall 0.97005 ± 0.00491 0.93242 ± 0.02067 0.95103 ± 0.00751
precision 0.97176 ± 0.00380 0.97085 ± 0.01821 0.98162 ± 0.00295
auc 0.99846 ± 0.00026 0.99060 ± 0.00957 0.99685 ± 0.00106
Validationloss 1.11480 ± 0.41351 0.47466 ± 0.12479 0.43730 ± 0.12738
accuracy 0.74893 ± 0.07493 0.85820 ± 0.02028 0.87654 ± 0.02105
recall 0.70229 ± 0.09837 0.80995 ± 0.03961 0.84339 ± 0.04841
precision 0.76943 ± 0.08938 0.89800 ± 0.03948 0.90251 ± 0.01986
auc 0.80492 ± 0.08378 0.92984 ± 0.01851 0.93931 ± 0.01630
Trainable parameters572916,38513,025
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tomasov, A.; Holik, M.; Oujezsky, V.; Horvath, T.; Munster, P. GPON PLOAMd Message Analysis Using Supervised Neural Networks. Appl. Sci. 2020, 10, 8139. https://doi.org/10.3390/app10228139

AMA Style

Tomasov A, Holik M, Oujezsky V, Horvath T, Munster P. GPON PLOAMd Message Analysis Using Supervised Neural Networks. Applied Sciences. 2020; 10(22):8139. https://doi.org/10.3390/app10228139

Chicago/Turabian Style

Tomasov, Adrian, Martin Holik, Vaclav Oujezsky, Tomas Horvath, and Petr Munster. 2020. "GPON PLOAMd Message Analysis Using Supervised Neural Networks" Applied Sciences 10, no. 22: 8139. https://doi.org/10.3390/app10228139

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop