Next Article in Journal
Power Load Probabilistic Prediction Based on Multi-Value Quantile Regression and Timing Fusion Ensemble Learning Model
Previous Article in Journal
Disorder-Assisted Adiabaticity in Correlated Many-Particle Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DELP-Net: A Differentiable Entropy Layer Pyramid Network for End-to-End Low-Rate DoS Detection

1
College of Artificial Intelligence, Jiaxing University, Jiaxing 314001, China
2
College of Computer Science and Technology, Zhejiang University, Hangzhou 310058, China
*
Author to whom correspondence should be addressed.
Entropy 2026, 28(3), 328; https://doi.org/10.3390/e28030328
Submission received: 7 February 2026 / Revised: 12 March 2026 / Accepted: 14 March 2026 / Published: 15 March 2026
(This article belongs to the Section Multidisciplinary Applications)

Abstract

Low-rate Denial-of-Service (LDoS) attacks exploit periodic traffic pulses to trigger congestion while maintaining a low average rate, making them highly stealthy and difficult to distinguish from legitimate bursty traffic using threshold-based or simple statistical detectors. To address this challenge, this paper proposes DELP-Net, an end-to-end Differentiable Entropy Layer Pyramid Network for window-level online LDoS detection directly from raw traffic. DELP-Net combines a multi-scale one-dimensional convolutional pyramid with a differentiable Rényi-entropy-driven attention mechanism to capture distributional regularity and weak repetitive patterns characteristic of LDoS traffic. In addition, an entropy-conditioned temporal convolutional network is employed to model cross-window periodic dependencies in a lightweight manner, together with an entropy-regularized hybrid loss to enhance robustness under complex background traffic. Experiments on the low-rate DoS dataset show that DELP-Net achieves an average F1 score of 0.9877 across six LDoS attack types, with a detection rate of 98.69% and a false-positive rate of 1.15%, demonstrating its effectiveness and suitability for practical online intrusion detection deployments.

1. Introduction

Low-rate Denial-of-Service (LDoS) attacks represent a class of highly stealthy yet destructive network attacks. Unlike conventional flooding-based Distributed Denial-of-Service (DDoS) attacks, LDoS attacks typically inject periodic traffic pulses at carefully selected moments to trigger congestion or resource bottlenecks [1,2]. By exploiting protocol dynamics and queue behaviors, these attacks can continuously suppress the throughput of legitimate services and even cause service unavailability while maintaining a very low average traffic rate. As a result, detection mechanisms that rely on traffic amplitude thresholds or simple statistical indicators often fail to respond effectively. Moreover, real-world network traffic exhibits inherent burstiness and diversity [3]. Legitimate phenomena such as flash crowds, application-level request bursts, link jitter, and retransmissions further increase detection difficulty, causing traditional statistical detectors to suffer from high false-positive rates and poor generalization.
Information entropy provides a powerful theoretical tool for network security analysis. Entropy characterizes the uncertainty and distributional complexity of random variables and has been widely used to quantify variations in source addresses, ports, packet lengths, inter-arrival times, and protocol field patterns [4]. In the context of LDoS attacks, traffic during pulse phases typically exhibits stronger regularity and repetitiveness [5]. Such aggregation effects induce observable changes in entropy-related measures. Compared with Shannon entropy [6], Rényi entropy [7] introduces a tunable order parameter α to adjust sensitivity to different regions of a distribution. When α > 1 , Rényi entropy becomes more sensitive to high-probability events, which makes it particularly suitable for amplifying distributional concentration caused by repetitive structures. This property aligns well with the statistical characteristics of LDoS traffic under low-rate conditions.
In recent years, deep learning has been extensively applied to intrusion detection and traffic classification [8], with representative models including convolutional neural networks (CNNs) [9], recurrent neural networks (RNNs) [10], temporal convolutional networks (TCNs) [11], and attention-based architectures [12]. One line of research relies on tools such as CICFlowMeter to extract flow-level statistical features before training classifiers [13]. Although these approaches achieve good performance on benchmark datasets, they suffer from feature engineering bottlenecks and complex deployment pipelines. Another line of work attempts end-to-end modeling from raw bytes or packet sequences. However, many existing methods treat network traffic as generic sequential signals and commonly employ amplitude- or mean-driven attention mechanisms, such as Squeeze-and-Excitation (SE) [14] or Convolutional Block Attention Module (CBAM) [15]. These attention mechanisms tend to emphasize channels with high energy responses, whereas the key cues of LDoS attacks are often not high-magnitude signals, but rather subtle distributional regularity and entropy anomalies. Consequently, how to elevate entropy from an external statistical descriptor to an internal, learnable operator within deep networks, while maintaining a lightweight and deployment-friendly design, remains a challenging and insufficiently explored problem.
To address these challenges, this paper proposes DELP-Net, an end-to-end differentiable entropy pyramid network for LDoS detection. The proposed network adopts time-window slicing as the basic data organization strategy, transforming raw network traffic into fixed-shape window tensors that are directly fed into a deep model to produce window-level predictions. DELP-Net employs a multi-scale one-dimensional convolutional pyramid to extract representations with diverse receptive fields, covering byte-level local patterns, protocol field structures, and intra-window packet sequence characteristics. On top of this, we design a Differentiable Rényi Entropy Attention Module (DREAM), which embeds differentiable Rényi entropy computation into intermediate network layers and performs entropy-driven channel recalibration. This design enables the model to adaptively focus on entropy-abnormal channels and informative segments. To further capture periodic pulse dependencies across windows, an entropy-conditioned TCN is introduced for temporal modeling. Dilated convolutions are used to obtain large receptive fields with low inference overhead. Finally, an entropy-regularized hybrid loss is incorporated to enhance the separability between attack and benign samples in the entropy domain, thereby improving robustness and reducing false positives under complex background traffic.
The main contributions of this paper are summarized as follows.
1.
We propose DELP-Net, an end-to-end differentiable entropy pyramid network for low-rate Denial-of-Service (LDoS) detection. The proposed approach operates directly on raw traffic streams segmented into time windows and upgrades information entropy from an external statistical descriptor to an intrinsic, learnable operator within the network, establishing a unified information-theoretic deep learning framework for detecting low-rate and weak-signal attacks.
2.
We design a Differentiable Rényi Entropy Attention Module (DREAM), which embeds Rényi entropy computation into the forward and backward propagation of the network and leverages entropy-driven channel recalibration. By characterizing distributional concentration and repetitive patterns, DREAM enhances the model’s sensitivity to weak and regular LDoS behaviors, outperforming conventional amplitude- or mean-based attention mechanisms.
3.
We introduce a multi-scale entropy pyramid learning framework that jointly models feature representations and entropy statistics across multiple receptive-field scales. Through scale-wise entropy-aware fusion, the proposed framework effectively captures both short burst behaviors and long-period structures of LDoS attacks, improving robustness to diverse low-rate attack patterns.
4.
We develop an entropy-conditioned temporal convolutional network (TCN) for cross-window temporal modeling, in which entropy-guided gating mechanisms are used to modulate dilated convolutional responses. This design enables the model to focus on entropy anomalies and abrupt entropy variations over time, facilitating effective modeling of periodic pulses and long-range dependencies in LDoS traffic.
The remainder of this paper is organized as follows. Section 2 reviews related work. Section 3 presents the overall methodology of DELP-Net. Section 4 describes the experimental setup and main results. Section 5 further validates the effectiveness of the proposed method through ablation studies, parameter sensitivity analysis, robustness analysis, and feature visualization. Finally, Section 6 concludes the paper.

2. Related Work

LDoS attacks are characterized by low average traffic rates, strong periodicity, and high similarity to legitimate bursty traffic. These properties distinguish LDoS detection from traditional flooding-based DoS detection in terms of modeling objectives, feature selection, and deployment constraints [16]. Existing studies on LDoS detection can be broadly categorized into three groups: statistical and signal processing-based methods, traditional machine learning-based methods, and deep learning-based methods.

2.1. Statistical and Signal Processing-Based LDoS Detection

Early studies on LDoS detection primarily adopt perspectives from signal processing and statistical analysis, treating network traffic as time series or stochastic processes and exploiting periodic pulse characteristics for detection. Representative approaches include power spectral density analysis in the frequency domain [17], short-time Fourier transform [18], and wavelet transform [19]. These methods typically identify anomalies by detecting low-frequency energy aggregation or periodic peaks, and are effective to some extent in scenarios where attack periodicity remains stable.
In parallel, information entropy has been applied to anomaly detection tasks by measuring uncertainty variations in attribute distributions such as source addresses, ports, packet lengths, and inter-arrival times. Considering the low-rate and high-regularity nature of LDoS attacks, some studies further introduce generalized entropy measures, including Rényi entropy and Tsallis entropy, to enhance sensitivity to distributional concentration and repetitive structures [20,21,22].
Despite their low computational cost and ease of implementation, statistical and signal processing-based methods face notable limitations in real-world network environments. First, these approaches often rely on manually designed features and fixed thresholds. Background traffic burstiness and jitter can significantly perturb statistical measures, making false-positive rates difficult to control. Second, periodic characteristics may be intentionally weakened through attack parameter randomization, such as varying pulse periods, duty cycles, or injecting jitter, which degrades the reliability of frequency-domain and threshold-based detectors. Third, statistical methods usually operate on low-dimensional attribute distributions and fail to exploit fine-grained structural information in protocol fields and payload fragments, resulting in limited expressive power against complex attack variants. These limitations indicate that relying solely on external statistical indicators is insufficient for robust LDoS detection under complex traffic conditions.

2.2. Traditional Machine Learning for LDoS and DoS Detection

To overcome the rigidity of fixed thresholds and single statistical metrics, traditional machine learning-based approaches typically adopt a two-stage pipeline [23]. In the first stage, handcrafted statistical features are extracted from PCAP files or flow records. In the second stage, classifiers are trained to perform detection. Common feature extraction tools generate flow-level or window-level feature vectors that include packet counts, byte counts, flow durations, inter-arrival time statistics, directional statistics, flag counts, and packet length quantiles [24,25,26].
In terms of classifiers, support vector machines [27], random forests [28], gradient boosting trees [29], autoencoders [30] and ensemble learning [31] methods have achieved relatively high accuracy on intrusion detection benchmark datasets. These models also provide a degree of nonlinear discrimination capability and feature importance analysis.
However, the major bottleneck of this category lies in feature engineering and deployment complexity. First, handcrafted features heavily depend on expert knowledge and often fail to cover diverse LDoS implementations. Attackers can adjust pulse patterns or perturb protocol fields to make statistical features resemble benign traffic, leading to degraded detection performance. Second, feature extraction usually requires complex preprocessing pipelines, introducing additional latency and maintenance overhead, and posing consistency risks between offline training and online deployment. Third, fixed feature spaces constrain model adaptability, causing generalization performance to deteriorate when network environments, application behaviors, or sampling conditions change. These issues limit the applicability of traditional machine learning frameworks for end-to-end, low-latency, and robust online LDoS detection.

2.3. Deep Learning Based Intrusion Detection

Deep learning provides end-to-end representation learning capabilities for intrusion detection. In existing research, CNNs are commonly used to capture local patterns with efficient inference, RNNs are employed to model temporal dependencies, and TCNs leverage dilated convolutions to achieve large receptive fields while preserving parallel computation [32,33,34]. Transformer and self-attention architectures are further introduced to learn global dependencies [35,36,37].
Regarding data representation, many studies still rely on handcrafted statistical features as model inputs, while others attempt to map traffic into image-like representations or directly model raw byte sequences in an end-to-end manner. To enhance discriminative ability, attention mechanisms such as SE and CBAM are frequently incorporated for channel recalibration, improving feature selection and information aggregation [38].
Although deep models achieve notable performance gains on several benchmark datasets, structural limitations remain for LDoS detection. First, many approaches still depend on engineered feature inputs, and end-to-end learning directly from raw network traffic remains relatively underexplored, limiting the utilization of fine-grained protocol fingerprints and payload structures. Second, mainstream attention modules usually rely on mean, amplitude, or correlation statistics, which tend to emphasize high-energy responses. In contrast, the critical cues of LDoS attacks often manifest as distributional regularity and entropy anomalies rather than prominent amplitude changes. Third, some methods adopt heavy sequence modeling backbones, leading to high inference overhead that is incompatible with latency and resource constraints in high-throughput online environments.
These observations suggest the need for an end-to-end framework that integrates information entropy as an internal, learnable operator and cooperates with lightweight temporal modeling, enabling stable perception of weak LDoS signals under complex background traffic while remaining suitable for practical deployment.

2.4. Recent Deep Learning Advances in IDS

In recent years, more advanced deep learning architectures have been explored for network intrusion detection. Transformer-based models have attracted increasing attention due to their ability to capture long-range dependencies and complex feature interactions in network traffic. By leveraging self-attention mechanisms, these models can automatically learn contextual relationships among traffic features and improve detection accuracy for sophisticated attacks. For example, Long et al. [39] proposed a transformer-based intrusion detection approach for cloud environments, demonstrating the effectiveness of attention mechanisms in modeling network traffic patterns. Xi et al. [40] further introduced a multi-scale transformer architecture to capture traffic features at different temporal scales, which improves detection performance on complex intrusion patterns. Recent studies have also explored self-supervised transformer-based representation learning to enhance generalization capability when labeled traffic data is limited [41]. These studies highlight the strong representation ability of transformer architectures for network security tasks. However, transformer-based IDS models often involve higher computational cost and require larger training datasets.
Another emerging research direction focuses on graph-based cybersecurity models. Since network communications naturally form relational structures among hosts, flows, and services, several studies model network traffic as graphs and apply graph neural networks to learn structural patterns of attacks. For instance, Lo et al. [42] proposed E-GraphSAGE, a graph neural network-based intrusion detection method that captures both flow features and network topology. Sun et al. [43] presented a graph neural network intrusion detection framework that integrates dynamic runtime telemetry with attack graph information. Graph-based approaches are also widely used for modeling communication relationships and multi-stage attack behaviors in complex networks [44]. Although these models can effectively capture relational dependencies among network entities, they typically require explicit graph construction and introduce additional preprocessing overhead.
Compared with these approaches, the method proposed in this work focuses on lightweight traffic representation and efficient feature extraction from flow-based traffic samples. The proposed architecture combines entropy-aware feature modeling with temporal convolutional learning to capture subtle traffic dynamics associated with low-rate DoS behaviors while maintaining relatively low computational complexity for practical deployment.

3. Methodology

3.1. Input Representation

This study focuses on the end-to-end detection of low-rate Denial-of-Service attacks. Given raw network traffic data, the input can be obtained from PCAP files captured at mirror ports or from equivalent online traffic streams. The model takes time-window-sliced raw data as input and directly outputs window-level attack labels, thereby realizing an end-to-end mapping from raw traffic to labels.
Let the time-ordered packet sequence obtained from packet capture be denoted as P = { p i } i = 1 N p , where each packet p i is associated with a timestamp t i and a byte sequence b i { 0 , 1 , , 255 } i . A fixed-length sliding time window is applied to segment P . Let the window length be T w and the stride be T s . The n-th time window is defined as W n = [ ( n 1 ) T s , ( n 1 ) T s + T w ) . The set of packets within this window is given by
P n = { p i P t i W n } .
To satisfy the fixed-size input requirement of neural networks, at most K packets are selected within each time window, and each packet is truncated or zero-padded to a fixed byte length L. Let TrimPad ( · , L ) denote the truncation and padding operator. The input tensor corresponding to window W n is constructed as X n R K × L :
X n [ k , : ] = 1 255 · TrimPad ( b ( n , k ) , L ) , k = 1 , 2 , , K ,
where b ( n , k ) denotes the byte sequence of the k-th packet selected from P n in chronological order. If the number of packets in the window is smaller than K, the remaining rows are padded with zeros. This representation preserves protocol header fields and partial payload information at the byte level while enabling low-latency streaming construction in practical deployments.
Window-level labels are used as supervision. Let T a ( W n ) denote the duration of attack traffic overlapping with window W n , and define the attack ratio as r n = T a ( W n ) / T w . The window label y n { 0 , 1 } is determined by
y n = 1 , r n τ 0 , r n < τ ,
where τ is a predefined label-alignment threshold. For offline experiments on public datasets, T a ( W n ) can be computed from original annotations or attack start and end times. For online environments, y n can be generated from security alerts or injected attack scripts. With the sliding-window formulation, the model produces a prediction every stride T s , supporting continuous detection and early warning.
Algorithm 1 summarizes the complete procedure of sliding-window segmentation, fixed-shape tensor construction, and window-level label alignment.
Algorithm 1: Window-Based Tensor Construction and Label Alignment
Entropy 28 00328 i001

3.2. LDoS Pulse Characteristics from an Entropy Perspective

The key characteristic of LDoS attacks lies in their low average traffic rate combined with periodic pulses that trigger congestion or resource exhaustion at critical moments, thereby significantly degrading the throughput or availability of legitimate services. This property makes detection methods based solely on traffic magnitude or simple statistical thresholds ineffective, and also limits the ability of conventional mean-based attention mechanisms to highlight attack signals.
The LDoS process can be abstracted as a periodic square-wave pulse function [45]. Let the attack period be T, the pulse duration be L p , and the peak rate be R peak . The attack traffic A ( t ) can be expressed as
A ( t ) = R peak , k T t < k T + L p 0 , k T + L p t < ( k + 1 ) T ,
where k Z 0 denotes the period index. The corresponding average rate is
R ¯ = R peak · L p T .
In typical LDoS configurations, L p T , resulting in an average rate R ¯ that is significantly lower than the detection thresholds used for flooding attacks. The observed mixed traffic X ( t ) can be modeled as the superposition of background traffic B ( t ) and attack traffic A ( t ) :
X ( t ) = B ( t ) + A ( t ) .
Since B ( t ) itself exhibits burstiness and strong noise, amplitude-based detection in the time domain cannot reliably distinguish the presence of A ( t ) .
An information-theoretic perspective provides a more suitable basis for LDoS detection. During attack pulses, LDoS traffic often exhibits increased regularity and repetition, such as periodic convergence of packet inter-arrival times, repeated packet-length and flag patterns, and recurrent protocol fields or payload fragments. Such regularity alters the concentration of feature distributions, producing more pronounced differences in the entropy domain. Compared with Shannon entropy, Rényi entropy introduces a parameter α to control sensitivity to different regions of the distribution. When α > 1 , Rényi entropy places more emphasis on high-probability events, making it well suited for capturing distributional concentration caused by repetitive structures. For a discrete probability distribution P = { p i } i = 1 M , Rényi entropy is defined as
H α ( P ) = 1 1 α log i = 1 M p i α , α > 0 , α 1 .
As α 1 , H α converges to Shannon entropy. In practice, α = 2 is commonly adopted to obtain the collision entropy form, which is more sensitive to probability mass concentration and thus effective at amplifying repetitive structures during LDoS pulses.
Based on this motivation, the core idea of this work is to introduce differentiable Rényi entropy computation into the hidden feature space of deep networks and use entropy as a learnable attention-driving signal. This design enables the model to adaptively focus on entropy-anomalous feature channels and temporal segments under low-amplitude conditions. With window-based slicing, entropy-domain differences can be stably observed at both window and cross-window levels. Combined with dilated convolutions in TCNs, the model can capture periodic structures and long-range dependencies, enabling robust detection and early identification of LDoS attacks.

3.3. Overall Architecture of DELP-Net

DELP-Net is proposed as an end-to-end learning framework for LDoS detection, and its overall architecture is illustrated in Figure 1. The model takes windowed raw byte tensors as input and sequentially applies a multi-scale 1D-CNN pyramid, a differentiable Rényi entropy attention module (DREAM), an entropy-conditioned temporal convolutional network (TCN), and a classification head to produce window-level attack predictions. The architecture emphasizes lightweight design and online inference efficiency, relying primarily on convolutional operations and avoiding heavy self-attention backbones.
Let the input of the n-th time window be X n R K × L . To explicitly model cross-window periodic dependencies, a window segment of length M is defined as
X n = { X n M + 1 , X n M + 2 , , X n } .
DELP-Net learns a mapping from X n to the predicted label y ^ n . The overall forward computation can be summarized as
y ^ n = f θ ( X n ) = Cls TCN ω ( Fuse ( DREAM ψ ( PyrCNN ϕ ( X n ) ) ) ) ,
where PyrCNN ϕ denotes multi-scale 1D-CNN pyramid feature extraction, DREAM ψ denotes differentiable Rényi entropy attention, Fuse represents multi-scale entropy-aware fusion, TCN ω denotes entropy-conditioned temporal modeling, and  Cls is the classification head. The parameter set θ = { ϕ , ψ , ω } contains all learnable parameters.
In practice, DELP-Net adopts a streaming window-level inference strategy. Each time a new window X n becomes available, the model uses X n to generate y ^ n , enabling continuous detection and early warning. Since LDoS attacks typically exhibit periodic pulse behavior, the segment length M together with the receptive field of the TCN determines the model’s ability to capture such periodic structures.

3.4. Multi-Scale 1D-CNN Pyramid for Window Feature Extraction

The multi-scale 1D-CNN pyramid module is designed to extract hierarchical representations from raw byte sequences within each time window, covering multiple receptive-field scales to accommodate the multi-scale nature of network traffic patterns. This module adopts convolution as the core operator, resulting in low computational complexity and suitability for online deployment. To preserve the temporal order of the input structure, the window tensor X n is serialized according to packet order to form a one-dimensional sequence,
s n = vec ( X n ) R S , S = K · L ,
where vec ( · ) denotes row-wise vectorization. This representation preserves protocol field ordering and payload fragment continuity at the byte level while maintaining the relative positional relationships among packets through sequential concatenation.
The pyramid architecture adopts parallel multi-branch convolutions, inspired by the Inception design [46]. Let the branch set be R = { r 1 , r 2 , r 3 } , corresponding to different kernel sizes, with a typical configuration of r 1 = 3 , r 2 = 5 , and  r 3 = 11 . Each branch consists of a Conv1D layer followed by batch normalization, ReLU activation, and downsampling pooling, producing branch-level feature maps
F n ( r ) = Pool ReLU ( BN ( Conv 1 D ( s n ; W r ) ) ) ,
where W r denotes the convolution parameters with kernel size r, and  Pool represents average pooling. The outputs from different scales are concatenated along the channel dimension to form the fused feature map F n , which serves as the input to the subsequent DREAM entropy-attention module:
F n = Concat F n ( r 1 ) , F n ( r 2 ) , F n ( r 3 ) .
The design of multi-scale convolution kernels has clear security semantics. Small-scale kernels are effective at capturing local byte patterns, which often correspond to localized combinations of protocol fields. Medium-scale kernels are more suitable for modeling structured fragments of protocol headers. Large-scale kernels cover longer contiguous byte ranges, facilitating the capture of repeated payload fragments and periodic traces formed by cross-packet concatenation within a window. Compared with single-scale convolution, the pyramid structure enables simultaneous extraction of fine-grained and coarse-grained traffic fingerprints, providing more discriminative candidate channels for subsequent entropy-based attention.
To control model size and maintain a lightweight design, several strategies are adopted. First, the number of pyramid stages is kept small, and stable performance gains can typically be achieved with two to three stacked stages. Second, channel widths across branches are balanced, and redundant channels are reduced to mitigate overfitting. Third, grouped convolution or depthwise separable convolution can be optionally employed to replace standard convolution, further reducing parameter count and FLOPs, which facilitates model compression and practical deployment.

3.5. Differentiable Rényi Entropy Attention Mechanism (DREAM)

The DREAM module introduces a differentiable Rényi entropy operator into the feature space and maps entropy values to attention weights for channel-wise recalibration of pyramid features. The structure of DREAM is illustrated in Figure 2. Unlike attention mechanisms based on mean or activation magnitude statistics, DREAM emphasizes distributional concentration, making it particularly suitable for characterizing the regularity and repetition exhibited by LDoS traffic under low-amplitude conditions.
Let the window-level feature map output by the multi-scale 1D-CNN pyramid be denoted as F n R T × C , where T is the sequence length and C is the number of channels. The activation value at position t and channel c is given by F n [ t , c ] . Here, the feature map F n represents the intermediate representation extracted by the multi-scale convolutional pyramid from the raw traffic window X n . Each channel corresponds to a specific response pattern learned from byte-level traffic structures, while the sequence dimension preserves the temporal ordering of packets within the window. Therefore, the feature tensor F n jointly encodes protocol-field structures, packet-level byte patterns, and intra-window temporal relationships.
To compute differentiable entropy, a pseudo-probability distribution is first constructed from continuous features. Specifically, softmax normalization is applied along the sequence dimension, mapping each channel’s energy distribution into a probability vector:
p t , c = exp ( F n [ t , c ] ) j = 1 T exp ( F n [ j , c ] ) , t = 1 , 2 , , T .
This yields the channel-wise distribution P c = { p t , c } t = 1 T , satisfying t = 1 T p t , c = 1 . This construction relies only on exponential, summation, and division operations, ensuring differentiability and computational efficiency suitable for lightweight online detection.
Rényi entropy is then computed on P c to obtain the channel-level entropy vector E n R C , with the c-th component defined as
E n [ c ] = 1 1 α log t = 1 T p t , c α , α > 0 , α 1 .
In this formulation, p t , c denotes the normalized activation probability at temporal position t for channel c, obtained through the softmax operation in Equation (13). The parameter α is the order of Rényi entropy, which controls the sensitivity of the entropy measure to concentration in the feature distribution. When α > 1 , the entropy becomes more sensitive to dominant activations, which helps highlight repetitive structures induced by LDoS traffic pulses.
In the main model, α = 2 is adopted to emphasize sensitivity to high-probability concentration, thereby amplifying low-entropy phenomena induced by repetitive structures. Since both Equations (13) and (14) are composed entirely of differentiable operations such as exponential, summation, logarithm, and division, the entropy computation can be naturally incorporated into gradient-based optimization.
The entropy vector is mapped to channel weights w n ( 0 , 1 ) C through a lightweight two-layer feedforward network:
w n = σ W 2 · ReLU ( W 1 · E n ) ,
where W 1 and W 2 are learnable parameters and σ ( · ) denotes the sigmoid function. Channel-wise recalibration is then applied to the feature map:
F n [ t , c ] = w n [ c ] · F n [ t , c ] .
This operation establishes a direct interaction between entropy statistics and feature representations. Channels with entropy patterns indicative of repetitive or concentrated activations receive higher attention weights, while channels dominated by irregular or noisy responses are suppressed. As a result, the network adaptively emphasizes feature channels that exhibit entropy anomalies consistent with LDoS traffic characteristics.
To improve training stability and preserve the original representation, a residual formulation is adopted:
F ˜ n = F n + F n .
Through this process, DREAM dynamically suppresses irregular, high-entropy noise channels while enhancing the discriminative contribution of entropy-anomalous channels, enabling the model to more effectively capture weak LDoS signatures.

3.6. Entropy Pyramid Fusion

Multi-scale convolutional pyramids produce complementary features at different receptive-field scales, and LDoS periodic pulses may manifest distinct entropy anomalies across these scales. To enhance coverage of multi-scale regularities, an entropy pyramid representation is constructed and fused into a compact window-level embedding, which serves as the input to subsequent temporal modeling. An overview of this process is shown in Figure 3.
Let the pyramid consist of S scale outputs indexed by s = 1 , 2 , , S , where the feature map at scale s is denoted as F n ( s ) R T s × C s . DREAM is independently applied at each scale to obtain enhanced features F ˜ n ( s ) and entropy vectors E n ( s ) . To derive a fixed-dimensional representation at the window level, sequence-wise aggregation is applied to the enhanced features using global average pooling, yielding a scale-level vector g n ( s ) R C s :
g n ( s ) [ c ] = 1 T s t = 1 T s F ˜ n ( s ) [ t , c ] , c = 1 , 2 , , C s .
To explicitly inject entropy information, the scale-level entropy vector is normalized and concatenated with the corresponding feature vector, forming an entropy-enhanced representation:
z n ( s ) = Concat g n ( s ) , Norm E n ( s ) ,
where Norm ( · ) denotes linear normalization or standardization to ensure consistency of entropy magnitudes across scales. The representations from all scales are concatenated to form the window-level entropy pyramid vector:
z n = Concat z n ( 1 ) , z n ( 2 ) , , z n ( S ) .
To control dimensionality and maintain a lightweight design, a linear projection is applied to compress z n into a fixed-dimensional embedding h n R d :
h n = W f z n + b f ,
where W f and b f are learnable parameters. The resulting vector h n serves as the window-level fused representation for subsequent TCN-based temporal modeling.
Considering the strong cross-window regularity of LDoS attacks, entropy variation trends also provide valuable discriminative information. Entropy difference features are further evaluated:
Δ E n ( s ) = E n ( s ) E n 1 ( s ) .
This feature measures entropy change between adjacent windows and can be concatenated with E n ( s ) during fusion to enhance sensitivity to pulse boundaries. Since Δ E n ( s ) involves only difference operations, it introduces negligible additional inference overhead.

3.7. Temporal Modeling with Entropy-Conditioned TCN

The key behavioral characteristic of LDoS attacks is manifested in cross-window periodic pulses and long-range temporal dependencies. Relying solely on within-window features is insufficient to stably capture such rhythmic structures. To address this issue, a lightweight Temporal Convolutional Network (TCN) is employed to model sequences of window-level representations. By using dilated convolutions, TCN expands the receptive field without a significant increase in parameter count, making it suitable for periodic pattern modeling in online detection scenarios.
Let the window-level representation obtained from entropy pyramid fusion be denoted as h n R d . A sequence of M consecutive windows is constructed as the input to the temporal model:
H n = { h n M + 1 , h n M + 2 , , h n } .
The TCN consists of L t stacked layers of one-dimensional dilated convolutions. For the -th layer, the dilation rate is denoted as δ and the kernel size is k t . For the i-th position in the sequence, the dilated convolution can be written as
u i ( ) = j = 0 k t 1 W j ( ) · u i j δ ( 1 ) + b ( ) ,
where u i ( 0 ) = h n M + i denotes the input feature vector at position i, and  W j ( ) and b ( ) are learnable parameters. Exponentially increasing dilation rates δ = 2 1 are adopted to rapidly enlarge the receptive field, allowing the model to cover multiple LDoS attack periods. Residual connections, together with normalization and ReLU activation, are applied to improve training stability and mitigate gradient vanishing.
To explicitly inject entropy information into the temporal modeling process, an entropy-conditioned mechanism is introduced to modulate the channel responses of the TCN. Let the entropy condition vector corresponding to window n be denoted as e n R d e . This vector can be constructed by concatenating entropy statistics from the entropy pyramid stage or extracted as a normalized entropy subvector during fusion. A sequence E n = { e n M + 1 , , e n } is formed and linearly mapped to generate gating coefficients:
g i = σ ( W g e n M + i + b g ) ,
where g i ( 0 , 1 ) d g denotes the gating vector, and  W g and b g are learnable parameters. For an intermediate TCN representation u ^ i ( ) , channel-wise modulation is applied as
u ˜ i ( ) = g i u ^ i ( ) ,
where ⊙ denotes the Hadamard product. This mechanism enables the model to place greater emphasis on windows exhibiting entropy anomalies or abrupt entropy changes, thereby enhancing sensitivity to LDoS pulses under low signal-to-noise conditions. Compared with incorporating self-attention backbones, entropy-conditioned TCN introduces only a small number of additional linear parameters, maintaining a lightweight and deployment-friendly design.
After L t TCN layers, the output sequence { u i ( L t ) } i = 1 M is obtained. A tail-aggregation strategy is adopted to derive the temporal representation for the current window n:
r n = u M ( L t ) .
This representation is fed into the classification head to produce the final prediction. Tail aggregation is consistent with the sliding-window inference strategy and facilitates real-time alert generation in online systems.

3.8. Classification with Entropy-Regularized Hybrid Loss

Based on the temporal representation r n , a lightweight classification head is employed to output the window-level prediction probability. The classifier consists of fully connected layers with ReLU activation and dropout to reduce overfitting. Let the classifier parameters be denoted as Θ c . For the binary classification case, the output is given by
y ^ n = Softmax ( W c r n + b c ) ,
where y ^ n R 2 , and  W c and b c are learnable parameters. The formulation can be directly extended to multi-class scenarios by increasing the output dimension to the number of classes C y .
An entropy-regularized hybrid loss is adopted to enhance class separability in the entropy domain and improve robustness under complex background traffic. The base classification loss is defined as the cross-entropy loss:
L CE = 1 N n = 1 N c = 1 C y 1 [ y n = c ] log ( y ^ n , c ) ,
where N denotes the number of training samples, C y is the number of classes, 1 [ · ] is the indicator function, and  y ^ n , c is the predicted probability for class c.
The entropy regularization term is applied to the channel entropy vectors produced by DREAM to directly constrain the distributional properties of hidden representations. Let E n denote the set of entropy vectors obtained from selected scales. For simplicity, these vectors are aggregated into a scalar entropy statistic:
E ¯ n = 1 | E n | E E n 1 dim ( E ) c = 1 dim ( E ) E [ c ] .
A class-conditional entropy constraint is adopted to encourage stronger concentration for attack samples while avoiding overly restrictive assumptions on benign traffic. The entropy regularization loss is defined as
L ER = 1 N n = 1 N y n · E ¯ n ,
where y n { 0 , 1 } denotes the binary label, with  y n = 1 indicating an attack window. Minimizing L ER encourages lower average entropy for attack samples, reinforcing the concentration of repetitive structures in the feature space. This constraint aligns with the sensitivity of Rényi entropy to high-probability concentration and contributes to improved detection performance for LDoS traffic.
The overall training objective is defined as
L Total = L CE + λ L ER ,
where λ is a weighting coefficient. The model parameters are optimized by minimizing L Total using stochastic gradient-based optimizers. To avoid representation collapse caused by excessive entropy regularization, a relatively small λ is adopted and tuned on the validation set. Dropout is also applied in the DREAM and TCN modules to further improve generalization. The impact of entropy regularization on the false-positive rate and robustness is analyzed in the Section 4.
Algorithm 2 summarizes the end-to-end training process of DELP-Net, including multi-scale feature extraction, differentiable Rényi entropy attention, entropy pyramid fusion, entropy-conditioned TCN temporal modeling, and optimization with the entropy-regularized hybrid loss.
Algorithm 2: End-To-End Training Process of DELP-Net
Entropy 28 00328 i002

4. Experiments

4.1. Datasets and Sample Construction

4.1.1. Datasets

To closely reflect practical intrusion detection scenarios, this study constructs the evaluation benchmark by combining publicly available datasets with traffic collected from a self-built experimental platform.
Three categories of data sources are used in this work. The first is the CICIDS2017 dataset [47], which contains a wide range of benign traffic and representative DoS-related attack traffic, together with relatively complete flow-level and packet-level annotations. The second is the CICIDS2018 dataset, which is larger in scale and exhibits stronger diversity in background traffic, enabling evaluation of model generalization under complex noise conditions [48]. The third is an extended low-rate DoS dataset collected on a self-built traffic generation and capture platform. This dataset incorporates more realistic characteristics such as link jitter and service-level traffic fluctuations, and better reflects practical deployment environments [49].
All data sources include benign traffic as the background class. The low-rate DoS attack types evaluated in this study mainly include Slowloris, Slow POST, Slow Read, Pwnloris, Torshammer, and Httpbog. These six attack categories are selected because they represent typical low-rate DoS behaviors and are consistently observable across the public and self-collected data sources used in this work.
For the two public datasets, traffic samples are extracted from the original traffic records, while benign samples are collected from normal background traffic sessions provided in the datasets. CICIDS2017 contains protocol-diverse traffic including TCP, UDP, and ICMP, whereas CICIDS2018 mainly contains TCP and UDP traffic under more complex background conditions. After filtering and sample construction, the subset used in this study contains approximately 62,000 samples from CICIDS2017 and 71,000 samples from CICIDS2018. For CICIDS2017, about 31,500 samples are benign and 30,500 samples are attack samples. For CICIDS2018, about 36,200 samples are benign and 34,800 samples are attack samples.
The self-built dataset is collected in a controlled experimental environment consisting of a traffic generation host, a target web service, and a monitoring node used for packet capture. Benign traffic is generated through common service access behaviors such as web browsing, file transfers, and streaming-like HTTP sessions, so the collected background traffic contains realistic application-level fluctuations. Low-rate DoS traffic is then injected by launching Slowloris, Slow POST, Slow Read, Pwnloris, Torshammer, and Httpbog attacks against the target service using publicly available tools. During the collection process, network delay perturbation and traffic jitter are intentionally introduced at the link and application levels to simulate unstable service access conditions that may appear in practice.
The self-collected dataset mainly consists of TCP traffic and contains approximately 28,500 samples, including about 14,800 benign samples and 13,700 attack samples. The total traffic collection duration is approximately 4 h, and the raw packet traces are captured continuously during both benign-only and attack-injection periods. Labels are assigned according to the known attack injection schedule and then verified through timestamp alignment between attack launch logs and captured traffic records. In this way, each traffic sample can be associated with a clear benign or malicious label, which improves the reliability of the annotation process.
Table 1 summarizes the main statistics of the datasets used in our experiments, including the number of samples, benign and attack distribution, protocol types, and training/testing splits.
To ensure fair comparison across different data sources, all datasets are processed using the same preprocessing and sample construction pipeline. Traffic flows are segmented into fixed-length traffic samples and normalized before being fed into the model. For each dataset, the training and testing sets are divided using a 70/30 split. The datasets are split chronologically at the flow level before window segmentation in order to avoid potential data leakage between training and testing samples.

4.1.2. Sample Construction

To match the window-level end-to-end input representation described in Section 3, the original PCAP files are parsed into time-ordered packet sequences and first aggregated into flows based on the five-tuple [50]. Flow aggregation is mainly used for label alignment and to prevent window concatenation across different sessions. Each flow is then segmented into fixed-length sliding time windows, and each window corresponds to one sample.
For the raw byte input within each window, at most K packets are selected in chronological order. Each packet is truncated or zero-padded to L bytes, forming a fixed-shape tensor X R K × L . This construction preserves protocol header fields and partial payload fragments at the byte level while explicitly aligning with the periodic pulse behavior of LDoS attacks through time-window slicing.

4.2. Evaluation Metrics

This study adopts A c c u r a c y , D e t e c t i o n R a t e , F a l s e P o s i t i v e R a t e , P r e c i s i o n , and F 1 score as evaluation metrics. To be consistent with common reporting practices in the intrusion detection literature, all metrics except for F 1 are reported in percentage form. Let T P , T N , F P , and F N denote true positives, true negatives, false positives, and false negatives, respectively. The metrics are defined as follows:
A c c u r a c y ( A C C ) = T P + T N T P + T N + F P + F N D e t e c t i o n R a t e ( D R ) = T P T P + F N F a l s e P o s i t i v e R a t e ( F P R ) = F P F P + T N P r e c i s i o n ( P r e ) = T P T P + F P F 1 = 2 · P r e · D R P r e + D R

4.3. Window-Based Preprocessing and Data Splits

A sliding time window is used as the basic sampling unit. The window length is set to T w and the stride is set to T s . Within each window, at most K packets are selected in chronological order, and each packet is truncated or zero-padded to L bytes, yielding a fixed-shape input tensor X R K × L . If the number of packets in a window is smaller than K, the remaining rows are padded with zeros. If a window contains far fewer than K packets and remains largely idle, it may lack sufficient discriminative information. To control the proportion of such low-information windows, we discard windows whose packet count is lower than K min during data construction, where K min is selected on the validation set.
For temporal modeling, adjacent windows within the same flow are grouped into an ordered segment of length M as the input to the TCN. If a flow is too short to form M consecutive windows, it is discarded to ensure consistency for sequence modeling. Window labels are generated from original annotations and attack intervals: if the overlap ratio between a window and an attack interval exceeds the threshold τ , the window is labeled as an attack window; otherwise, it is labeled as benign. The preprocessing parameters used are summarized in Table 2.

4.4. Baselines

We consider the following classic baselines, covering statistical detection, traditional machine learning, and deep learning models. All baselines are trained and evaluated under the same window-based preprocessing configuration as DELP-Net.
  • Entropy-Threshold: A window-level entropy threshold detector. For each time window, empirical distributions are constructed on discretized traffic attributes. Rényi entropy statistics are computed and compared against thresholds estimated from benign windows in the training set and tuned on the validation set to control false alarms.
  • RF: A Random Forest classifier trained on handcrafted window-level statistical features. The feature vector includes byte counts, mean and variance of inter-arrival times, simple directional statistics within the window, TCP flag counts, and packet-length quantiles. Hyperparameters are selected on the validation set.
  • XGBoost: A gradient-boosted decision tree classifier trained on the same handcrafted window-level feature set as RF. Model hyperparameters are tuned on the validation set to ensure a fair comparison.
  • 1D-CNN: An end-to-end convolutional baseline operating on raw bytes. Each window is represented as a fixed-shape K × L byte tensor (with truncation/padding) and then serialized as a 1D sequence. Stacked single-scale Conv1D layers extract window representations, followed by a fully connected classification head.
  • LSTM: A deep sequence baseline for raw-window modeling. The K × L byte tensor is treated as an ordered sequence, and an LSTM encoder is used to model temporal dependencies within the window. The final hidden state is fed into a fully connected classifier.

4.5. Implementation Details

All experiments are performed on a workstation featuring an Intel Xeon(R) Platinum 8352V processor operating at 2.1 GHz, 64 GB of system memory, and an NVIDIA RTX 4090D graphics card with 24 GB of video memory. The operating system is Ubuntu 22.04 LTS. All models are implemented using PyTorch 2.3.0, and GPU acceleration is enabled via CUDA 12.1 and cuDNN 8.
During training, the Adam optimizer is used with an initial learning rate of 1 × 10 3 , and the learning rate is decayed when the validation performance no longer improves. To enhance generalization, dropout is applied to convolutional layers and fully connected layers, with the dropout rate selected on the validation set. In the DREAM module, the Rényi order is fixed to α = 2 , and the entropy-regularization weight λ is tuned on the validation set. Early stopping is adopted for all models: if the validation F1 score does not improve for a number of consecutive epochs, training is terminated and the best-performing checkpoint is restored. The key hyperparameters used in the final configuration are summarized in Table 3.

4.6. Detection Performance

This subsection reports the main detection results of DELP-Net for LDoS attack subclasses. To highlight the model’s adaptability to different LDoS patterns, we evaluate window-level detection performance for six attack types: Slowloris, Slow POST, Slow Read, Pwnloris, Torshammer, and Httpbog. The per-attack results of the proposed method are presented in Table 4. DELP-Net consistently achieves high detection rates and high precision across all subclasses, with an average F1 score of 0.9877. These results indicate that the proposed method effectively captures different weak regularity patterns of LDoS traffic.
Table 5 reports window-based comparison results against representative classical baselines, including an entropy-threshold detector, traditional machine learning classifiers, and commonly used deep sequence models. As can be observed, the entropy-threshold baseline suffers from a relatively high false-positive rate, indicating that fixed-threshold statistics are easily perturbed by bursty and diverse benign traffic. Traditional machine learning methods (RF and XGBoost) further show limited discriminative capability under the raw-traffic window setting, resulting in both lower DR and higher FPR. End-to-end deep baselines (1D-CNN and LSTM) improve overall performance but still exhibit non-negligible false alarms, suggesting that modeling raw bytes alone without explicitly leveraging distributional regularity is insufficient for robust low-rate DoS detection.
In contrast, DELP-Net achieves the best trade-off across all key metrics, reaching a DR of 98.69% and an F1 score of 0.9877 while reducing the false-positive rate to 1.15%. The consistent gains in both DR (98.69%) and Precision (98.85%) indicate that the proposed entropy-driven representation learning and temporal modeling effectively capture the weak yet regular patterns of LDoS traffic, thereby improving detection robustness under complex benign background dynamics.

5. Discussion

5.1. Comparison with Recent State-of-the-Art Methods

To further assess the effectiveness of the proposed DELP-Net, we compare it with several recent state-of-the-art detection methods reported in the literature. These methods span different detection paradigms, including feature-engineered classifiers, signal- or image-based representations, hybrid learning frameworks, and deep neural network models.
As shown in Table 6, DELP-Net achieves the best overall performance among the compared methods, particularly in terms of F1 score and detection rate. Compared with earlier approaches that rely on handcrafted features or fixed representations, the proposed model benefits from end-to-end learning directly on raw traffic windows, which helps preserve fine-grained protocol and temporal information.
Recent deep learning-based methods, such as MSCBL-AND and ERT-EDR, demonstrate competitive results by leveraging multi-scale representations or ensemble decision strategies. In contrast, DELP-Net explicitly embeds differentiable Rényi entropy into the network architecture, enabling the model to capture distributional regularity and repetition patterns that are characteristic of low-rate DoS attacks but are often overlooked by amplitude- or energy-driven mechanisms.
Moreover, unlike methods that depend on complex preprocessing pipelines or high-dimensional handcrafted features, DELP-Net adopts a lightweight convolutional backbone and supports window-level online inference. This design allows the model to achieve superior detection performance while remaining suitable for practical deployment scenarios with strict latency and throughput constraints.

5.2. Ablation Studies

We evaluate the contribution of each major component in DELP-Net under the same window-based preprocessing and split strategy as in Section 4. We report ACC, DR, FPR, Precision, and F1 to reflect the trade-off between detection performance and false alarms.

5.2.1. Ablation Settings

To quantify the effect of each component, we construct the following variants by removing or replacing specific modules while keeping all other settings unchanged.
  • w/o DREAM: Remove DREAM and disable entropy-driven channel recalibration, and pyramid features are forwarded directly to subsequent modules.
  • DREAM to SE: Replace DREAM with the Squeeze-and-Excitation (SE) channel attention [14].
  • DREAM to CBAM: Replace DREAM with CBAM using mean-based channel and spatial attention [57].
  • w/o Entropy Pyramid Fusion: Keep multi-scale convolution and DREAM, but remove multi-scale entropy fusion and use only the final-scale feature as the window representation.
  • w/o TCN: Remove the temporal modeling module, and window embeddings are fed directly into the classifier.
  • TCN w/o Entropy Conditioning: Keep TCN but disable entropy-conditioned gating, and temporal modeling uses window embeddings only.
  • DELP-Net (Full): The full model.

5.2.2. Ablation Results

Table 7 summarizes the ablation results. Removing DREAM causes a clear degradation, with F1 dropping to 0.9439 and FPR increasing to 5.48%, indicating that entropy-driven channel recalibration is critical for suppressing false alarms in low-rate settings. Replacing DREAM with SE improves performance (F1 = 0.9585), while CBAM yields further gains (F1 = 0.9656), showing that conventional attention helps but remains less effective than entropy-driven attention for capturing LDoS regularity. Disabling entropy pyramid fusion reduces F1 to 0.9670, suggesting that aggregating entropy-aware representations across multiple receptive-field scales improves coverage of diverse LDoS patterns. Removing the TCN decreases F1 to 0.9761, confirming the importance of modeling cross-window temporal dependencies. Moreover, disabling entropy-conditioned gating within TCN further reduces F1 to 0.9820, highlighting the complementary role of entropy information in guiding temporal decision-making.
Overall, DELP-Net achieves the best performance, validating the effectiveness of DREAM, entropy pyramid fusion, and entropy-conditioned temporal modeling.

5.3. Sensitivity Analysis of Rényi Parameter α

This subsection investigates the sensitivity of DELP-Net to the Rényi order α used in the DREAM module. Since α controls how strongly the entropy measure emphasizes high-probability events, it directly affects how DREAM responds to distributional concentration and repetitive patterns, which are key characteristics of low-rate DoS traffic. Intuitively, a larger α assigns more weight to dominant activations, whereas a smaller α increases the contribution of low-probability components, potentially weakening the contrast between benign burstiness and attack-induced regularity.

5.3.1. Experimental Setting

We keep the network architecture, window-based preprocessing, training schedule, and data splits unchanged, and vary only the Rényi order in DREAM. Specifically, we evaluate α { 0.5 , 1 , 2 , 3 , 4 , 5 } . All settings share the same hyperparameters and early-stopping strategy to ensure a fair comparison.

5.3.2. Results and Discussion

Figure 4 summarizes the performance trends under different α values. The best overall performance is achieved at α = 2 . When α deviates from 2, all key metrics degrade to varying degrees, with a consistent drop in F1. For example, using Shannon entropy ( α = 1 ) yields slightly lower performance, while smaller α ( 0.5 ) further reduces F1 to 0.9708. For larger α values, the performance also decreases, suggesting that overly emphasizing a small set of dominant activations may reduce robustness under background fluctuations.
From a mathematical perspective, α controls how strongly the Rényi entropy emphasizes different regions of the feature distribution. When α < 1 , relatively greater importance is assigned to low-probability components, making the entropy more sensitive to weak background fluctuations and minor irregularities. As a result, the contrast between benign burstiness and attack-induced concentration may be weakened, which can reduce detection sensitivity. In contrast, when α > 1 , the entropy places more emphasis on dominant probability mass, thereby making the model more sensitive to concentrated and repetitive activation patterns that are characteristic of LDoS attacks. However, if α becomes excessively large, the entropy may over-focus on only a few dominant activations, which can reduce robustness to normal traffic variation and lead to slight performance degradation.
Overall, α = 2 provides the most favorable balance between detection capability and false-alarm suppression. This observation is consistent with the collision-entropy-like behavior at α = 2 , which effectively highlights concentration caused by repetitive attack patterns while avoiding excessive sensitivity that may arise with very large α . Therefore, we adopt α = 2 as the default configuration for DREAM in all main experiments.

5.4. Robustness to Attack Parameter Variations

Low-rate DoS attacks can be easily adapted by adversaries through manipulating key pulse parameters, such as the pulse period T, pulse duration L p , and peak rate R peak , and by injecting timing jitter to reduce regularity. Therefore, beyond in-distribution evaluation, it is important to quantify how stable DELP-Net remains under controlled parameter perturbations that alter the temporal dynamics while preserving the low-rate nature of the attack.

5.4.1. Experimental Setting

We construct perturbed test sets by modifying one or more pulse parameters in the LDoS process, following the pulse abstraction in Equation (4). Specifically, we consider four settings: (i) None, the default evaluation configuration; (ii) Low (in-distribution), where parameters remain close to the training distribution with a small amount of noise; (iii) Medium, where T and L p are varied to change the duty cycle and periodicity while keeping the overall behavior comparable; and (iv) High, where T, L p , and R peak are jointly perturbed and additional timing jitter is introduced to weaken the periodic signature [58,59]. All other components are kept unchanged. We report DR, Precision, and F1 under the same window-based protocol as the main experiments.

5.4.2. Results and Discussion

Figure 5 summarizes the performance trends under different perturbation levels. DELP-Net maintains consistently strong performance as perturbation intensity increases. Starting from the unperturbed setting, the model exhibits only mild and monotonic degradation under Low and Medium perturbations, indicating that modest changes in T and L p do not substantially affect the learned detection cues. Even under the High perturbation setting, where multiple parameters are varied and jitter is injected, DELP-Net retains a high detection rate and low false-positive rate.
These results suggest that the proposed entropy-driven representation learning is not overly dependent on a single fixed pulse configuration. Instead, DELP-Net captures more stable distributional regularity and cross-window temporal patterns that persist across a range of LDoS parameterizations. The limited decrease in F1 under stronger perturbations further indicates that the model does not simply overfit to narrow periodic signatures, but remains robust under adversarial timing variations.

5.5. Feature Space Visualization

Although the detection task can be formulated as binary classification, we visualize the learned embeddings under a multi-class setting to examine whether the model can distinguish fine-grained LDoS variants with highly similar low-rate and bursty behaviors. We employ t-distributed stochastic neighbor embedding (t-SNE) to project the high-dimensional classifier input features into a two-dimensional space [60]. Figure 6 compares the resulting feature distributions of the 1D-CNN baseline and the proposed DELP-Net.
In the 1D-CNN embedding, benign windows and multiple LDoS subclasses exhibit substantial overlap, and closely related attacks are poorly separated, indicating that raw-byte convolution is strongly affected by background burstiness and struggles to capture subtle distributional regularities. In contrast, DELP-Net produces more compact intra-class clusters and clearer margins between benign traffic and all attack subclasses while significantly reducing overlap among different low-rate DoS variants. This suggests that DREAM and entropy-regularized training promote an entropy-aware representation space that suppresses high-entropy background noise and emphasizes repetitive low-entropy patterns, resulting in improved separability and more stable decision boundaries.

5.6. Computational Cost and Deployment Feasibility

To further evaluate the practicality of DELP-Net for real-time intrusion detection, we analyze its computational cost and deployment efficiency. In particular, we report the number of trainable parameters, approximate computational complexity (FLOPs), inference latency per traffic sample, and memory usage during inference.
All experiments are conducted using the same implementation environment described in Section 4.5. During latency measurement, the model is executed in inference mode with gradient computation disabled. The reported latency corresponds to the average runtime per sample after a warm-up stage.
Table 8 reports the overall model complexity. DELP-Net maintains a moderate number of parameters while avoiding the higher computational overhead typically introduced by deeper sequence models. The additional computation mainly originates from the entropy-aware feature recalibration and temporal modeling modules, which significantly improve detection performance while keeping the overall complexity manageable.
As shown in Table 9, DELP-Net requires only a small inference delay for each traffic sample and is able to process 610 samples per second. Considering that each detection sample corresponds to one flow of traffic in our dataset construction, the measured inference latency is significantly lower than the data acquisition interval. This indicates that DELP-Net can operate in real time without introducing processing bottlenecks.
In addition, the memory consumption remains moderate, which supports the practical deployment of DELP-Net in online IDS environments on conventional GPU-enabled monitoring servers.

5.7. Detection Granularity

The proposed method evaluates detection performance at the traffic-sample level, where each sample corresponds to a fixed-length traffic window extracted from network flows. This window-based representation is commonly adopted in intrusion detection studies because it allows continuous monitoring of network behavior while maintaining stable input dimensions for deep learning models.
Although the reported metrics are computed at the sample level, the detection results can be naturally aggregated at the flow or attack-event level in practical deployment. In an online IDS environment, multiple consecutive windows belonging to the same flow or service session are analyzed sequentially. Once a certain number of windows are classified as malicious, the corresponding flow or attack event can be identified. Therefore, the window-level detection used in this work provides a fine-grained and early indication of potential attacks.
In addition, the inference latency analysis reported in Section 5.6 shows that the proposed model can process traffic samples in real time. Since each sample corresponds to a short observation interval, the detection delay for an attack event is limited to only a few windows in practice. This property allows the system to respond to low-rate DoS behaviors before the attack traffic accumulates to a disruptive level.
Regarding traffic load variations, the datasets used in our experiments contain diverse background traffic patterns and protocol distributions, including both public benchmark datasets and traffic collected from a self-built experimental environment with injected jitter and service fluctuations. These characteristics provide a certain degree of robustness evaluation under varying traffic conditions.

6. Conclusions

Low-rate Denial-of-Service (LDoS) attacks induce congestion or resource exhaustion at critical moments through periodic traffic pulses. Due to their low average traffic rate and the similarity of their behavior to legitimate bursty traffic, traditional detection methods based on amplitude thresholds or simple statistical features find it difficult to achieve both high detection rates and low false-positive rates under complex background traffic. To address this challenge, this paper proposes an end-to-end differentiable entropy pyramid network, DELP-Net, for window-level low-rate DoS detection directly from raw network traffic, from a joint information-theoretic and deep representation learning perspective. The core idea of the proposed method is to transform information entropy from an external statistical feature into an internal learnable operator and to embed it deeply into feature representation and temporal modeling. Specifically, a differentiable Rényi entropy attention module, DREAM, is designed to explicitly model distributional concentration and repetitive structures in the feature space, enabling the model to effectively perceive entropy anomalies caused by LDoS attacks under low-amplitude conditions. Meanwhile, a multi-scale entropy pyramid learning framework is constructed to jointly model convolutional features and entropy statistics across different receptive-field scales, covering the multi-scale characteristics of LDoS attacks in both local burst behaviors and long-period structures. On this basis, an entropy-conditioned temporal convolutional network is introduced for cross-window temporal modeling, guiding the model to focus on time segments corresponding to entropy anomalies or abrupt entropy variations, so as to stably capture the periodic pulse dependencies of LDoS attacks. In addition, an entropy-regularized hybrid loss function is introduced to further enhance the separability of attack and benign samples in the entropy domain, improving robustness and false-positive control under complex background traffic. Experimental results on public datasets show that DELP-Net achieves stable and significant performance advantages across six LDoS attack types, with an average F1 score of 0.9877 while maintaining a low false-positive rate. Ablation studies verify the effectiveness of DREAM, multi-scale entropy pyramid fusion, and entropy-conditioned temporal modeling in improving overall performance. Moreover, the model mainly relies on convolutional operators, providing good computational efficiency and online inference feasibility, and meeting the basic requirements of real-world network environments for real-time performance and stability.
Future work will focus on two main directions. First, the robustness of DELP-Net against adversarial LDoS attacks will be systematically investigated, with particular attention to adversarially crafted traffic patterns, dynamic attack parameter variations, and evasion strategies. In this context, online adaptive and continual learning mechanisms will also be explored to enhance model robustness while mitigating the risk of catastrophic forgetting. Second, more efficient deployment optimization strategies will be studied to further reduce inference latency and improve real-time processing capability in high-throughput network environments.

Author Contributions

Conceptualization, J.W. and C.X.; methodology, C.X.; software, J.Y.; validation, J.W.; writing—original draft preparation, C.X.; writing—review and editing, J.W., C.X. and J.Y.; funding acquisition, C.X. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Natural Science Foundation of China (No. 62302197) and the Jiaxing City Science and Technology Project (No. 2024AY40010).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The CICIDS2017 dataset used in this research can be retrieved via the following URL: https://www.unb.ca/cic/datasets/ids-2017.html (accessed on 6 February 2026). The CICIDS2018 dataset used in this research can be retrieved via the following URL: https://www.unb.ca/cic/datasets/ids-2018.html (accessed on 6 February 2026).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tang, D.; Dai, R.; Yan, Y.; Li, K.; Liang, W.; Qin, Z. When SDN Meets Low-rate Threats: A Survey of Attacks and Countermeasures in Programmable Networks. ACM Comput. Surv. 2024, 57, 1–32. [Google Scholar] [CrossRef]
  2. Setitra, M.A.; Fan, M.; Benkhaddra, I.; Bensalem, Z.E.A. DoS/DDoS attacks in Software Defined Networks: Current situation, challenges and future directions. Comput. Commun. 2024, 222, 77–96. [Google Scholar] [CrossRef]
  3. Benmoussa, A.; Kerrache, C.A.; Lagraa, N.; Mastorakis, S.; Lakas, A.; Tahari, A.E.K. Interest flooding attacks in named data networking: Survey of existing solutions, open issues, requirements, and future directions. ACM Comput. Surv. 2022, 55, 1–37. [Google Scholar] [CrossRef]
  4. Wang, B. Improving multivariate time-series anomaly detection in industrial sensor networks using entropy-based feature aggregation. Entropy 2026, 28, 14. [Google Scholar] [CrossRef]
  5. Wu, Z.; Zhang, L.; Yue, M. Low-rate DoS attacks detection based on network multifractal. IEEE Trans. Dependable Secur. Comput. 2015, 13, 559–567. [Google Scholar] [CrossRef]
  6. Lin, J. Divergence measures based on the Shannon entropy. IEEE Trans. Inf. Theory 2002, 37, 145–151. [Google Scholar] [CrossRef]
  7. Dong, X. The gravity dual of Rényi entropy. Nat. Commun. 2016, 7, 12472. [Google Scholar] [CrossRef]
  8. Liu, Z.; Yin, X.; Liu, D. A Comprehensive Survey on Low-rate DDoS Attacks Detection Based on Deep Learning. In Proceedings of the 2024 International Conference on Computers, Information Processing and Advanced Education (CIPAE), Ottawa, ON, Canada, 26–28 August 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 259–263. [Google Scholar]
  9. Ren, K.; Yuan, S.; Zhang, C.; Shi, Y.; Huang, Z. CANET: A hierarchical CNN-attention model for network intrusion detection. Comput. Commun. 2023, 205, 170–181. [Google Scholar] [CrossRef]
  10. Gamal, M.; Elhamahmy, M.; Taha, S.; Elmahdy, H. Improving intrusion detection using LSTM-RNN to protect drones’ networks. Egypt. Inform. J. 2024, 27, 100501. [Google Scholar] [CrossRef]
  11. Rani, Y.A.; Reddy, E.S. Deep intrusion net: An efficient framework for network intrusion detection using hybrid deep TCN and GRU with integral features. Wirel. Netw. 2025, 31, 1255–1278. [Google Scholar] [CrossRef]
  12. Laghrissi, F.; Douzi, S.; Douzi, K.; Hssina, B. IDS-attention: An efficient algorithm for intrusion detection systems using attention mechanism. J. Big Data 2021, 8, 149. [Google Scholar] [CrossRef]
  13. Lashkari, A.H.; Gil, G.D.; Mamun, M.S.I.; Ghorbani, A.A. Characterization of Tor Traffic using Time based Features. In Proceedings of the 3rd International Conference on Information Systems Security and Privacy—Volume 1: ICISSP; SciTePress: Setúbal, Portugal, 2017; pp. 253–262. [Google Scholar] [CrossRef]
  14. Jin, X.; Xie, Y.; Wei, X.S.; Zhao, B.R.; Chen, Z.M.; Tan, X. Delving deep into spatial pooling for squeeze-and-excitation networks. Pattern Recognit. 2022, 121, 108159. [Google Scholar] [CrossRef]
  15. Hosseini, S.; Ebrahimi, A.; Mosavi, M.; Shahhoseini, H. A novel hybrid CNN-CBAM-GRU method for intrusion detection in modern networks. Results Eng. 2025, 28, 107103. [Google Scholar] [CrossRef]
  16. Bhuyan, M.H.; Bhattacharyya, D.; Kalita, J.K. An empirical evaluation of information metrics for low-rate and high-rate DDoS attack detection. Pattern Recognit. Lett. 2015, 51, 1–7. [Google Scholar] [CrossRef]
  17. Chen, Z.; Yeo, C.K.; Lee, B.S.; Lau, C.T. Power spectrum entropy based detection and mitigation of low-rate DoS attacks. Comput. Netw. 2018, 136, 80–94. [Google Scholar] [CrossRef]
  18. Fu, Y.; Duan, X.; Wang, K.; Li, B. Low-rate Denial of Service attack detection method based on time-frequency characteristics. J. Cloud Comput. 2022, 11, 31. [Google Scholar] [CrossRef]
  19. Yue, M.; Liu, L.; Wu, Z.; Wang, M. Identifying LDoS attack traffic based on wavelet energy spectrum and combined neural network. Int. J. Commun. Syst. 2018, 31, e3449. [Google Scholar] [CrossRef]
  20. Yu, H.; Yang, W.; Cui, B.; Sui, R.; Wu, X. Renyi entropy-driven network traffic anomaly detection with dynamic threshold. Cybersecurity 2024, 7, 64. [Google Scholar] [CrossRef]
  21. Lima, C.F.L.; Assis, F.M.; de Souza, C.P. A comparative study of use of Shannon, Rényi and Tsallis entropy for attribute selecting in network intrusion detection. In Proceedings of the 2011 IEEE International Workshop on Measurements and Networking Proceedings (M&N), Anacapri, Italy, 10–11 October 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 77–82. [Google Scholar]
  22. Tang, D.; Dai, R.; Zuo, C.; Chen, J.; Li, K.; Qin, Z. A Low-Rate DoS Attack Mitigation Scheme Based on Port and Traffic State in SDN. IEEE Trans. Comput. 2025, 74, 1758–1770. [Google Scholar] [CrossRef]
  23. Tulczyjew, L.; Biruk, I.; Bilgic, M.; Abondo, C.; Weill, N. PCAPVision: PCAP-Based High-Velocity and Large-Volume Network Failure Detection. In Proceedings of the NAIC ’24: 2024 SIGCOMM Workshop on Networks for AI Computing, Sydney, NSW, Australia, 4–8 August 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 26–33. [Google Scholar] [CrossRef]
  24. Liu, L.; Wang, H.; Wu, Z.; Yue, M. The detection method of low-rate DoS attack based on multi-feature fusion. Digit. Commun. Netw. 2020, 6, 504–513. [Google Scholar] [CrossRef]
  25. Tang, D.; Tang, L.; Dai, R.; Chen, J.; Li, X.; Rodrigues, J.J. MF-Adaboost: LDoS attack detection based on multi-features and improved Adaboost. Future Gener. Comput. Syst. 2020, 106, 347–359. [Google Scholar] [CrossRef]
  26. Li, S.; Cao, Y.; Liu, S.; Lai, Y.; Zhu, Y.; Ahmad, N. HDA-IDS: A Hybrid DoS Attacks Intrusion Detection System for IoT by using semi-supervised CL-GAN. Expert Syst. Appl. 2024, 238, 122198. [Google Scholar] [CrossRef]
  27. Azimjonov, J.; Kim, T. Designing accurate lightweight intrusion detection systems for IoT networks using fine-tuned linear SVM and feature selectors. Comput. Secur. 2024, 137, 103598. [Google Scholar] [CrossRef]
  28. Abdelaziz, M.T.; Radwan, A.; Mamdouh, H.; Saad, A.S.; Abuzaid, A.S.; AbdElhakeem, A.A.; Zakzouk, S.; Moussa, K.; Darweesh, M.S. Enhancing network threat detection with random forest-based NIDS and permutation feature importance. J. Netw. Syst. Manag. 2025, 33, 2. [Google Scholar] [CrossRef]
  29. Chalichalamala, S.; Govindan, N.; Kasarapu, R. An extreme gradient boost based classification and regression tree for network intrusion detection in IoT. Bull. Electr. Eng. Inform. 2024, 13, 1741–1751. [Google Scholar] [CrossRef]
  30. Wang, Z.; Yu, X. CSCVAE-NID: A Conditionally Symmetric Two-Stage CVAE Framework with Cost-Sensitive Learning for Imbalanced Network Intrusion Detection. Entropy 2025, 27, 1086. [Google Scholar] [CrossRef] [PubMed]
  31. Mohy-Eddine, M.; Guezzaz, A.; Benkirane, S.; Azrour, M.; Farhaoui, Y. An ensemble learning based intrusion detection model for industrial IoT security. Big Data Min. Anal. 2023, 6, 273–287. [Google Scholar] [CrossRef]
  32. Salmi, S.; Oughdir, L. Performance evaluation of deep learning techniques for DoS attacks detection in wireless sensor network. J. Big Data 2023, 10, 17. [Google Scholar] [CrossRef]
  33. Akgun, D.; Hizal, S.; Cavusoglu, U. A new DDoS attacks intrusion detection model based on deep learning for cybersecurity. Comput. Secur. 2022, 118, 102748. [Google Scholar] [CrossRef]
  34. Abdel Wahab, O. Intrusion Detection in the IoT Under Data and Concept Drifts: Online Deep Learning Approach. IEEE Internet Things J. 2022, 9, 19706–19716. [Google Scholar] [CrossRef]
  35. Manocchio, L.D.; Layeghy, S.; Lo, W.W.; Kulatilleke, G.K.; Sarhan, M.; Portmann, M. Flowtransformer: A transformer framework for flow-based network intrusion detection systems. Expert Syst. Appl. 2024, 241, 122564. [Google Scholar] [CrossRef]
  36. Kheddar, H. Transformers and large language models for efficient intrusion detection systems: A comprehensive survey. Inf. Fusion 2025, 124, 103347. [Google Scholar] [CrossRef]
  37. Qathrady, M.A.; Ullah, S.; Alshehri, M.S.; Ahmad, J.; Almakdi, S.; Alqhtani, S.M.; Khan, M.A.; Ghaleb, B. SACNN-IDS: A self-attention convolutional neural network for intrusion detection in industrial internet of things. CAAI Trans. Intell. Technol. 2024, 9, 1398–1411. [Google Scholar] [CrossRef]
  38. Mehedi, S.T.; Anwar, A.; Rahman, Z.; Ahmed, K.; Islam, R. Dependable intrusion detection system for IoT: A deep transfer learning based approach. IEEE Trans. Ind. Inform. 2022, 19, 1006–1017. [Google Scholar] [CrossRef]
  39. Long, Z.; Yan, H.; Shen, G.; Zhang, X.; He, H.; Cheng, L. A Transformer-based network intrusion detection approach for cloud security. J. Cloud Comput. 2024, 13, 5. [Google Scholar] [CrossRef]
  40. Xi, C.; Wang, H.; Wang, X. A novel multi-scale network intrusion detection model with transformer. Sci. Rep. 2024, 14, 23239. [Google Scholar] [CrossRef]
  41. Koukoulis, I.; Syrigos, I.; Korakis, T. Self-supervised transformer-based contrastive learning for intrusion detection systems. arXiv 2025, arXiv:2505.08816. [Google Scholar]
  42. Lo, W.W.; Layeghy, S.; Sarhan, M.; Gallagher, M.; Portmann, M. E-GraphSAGE: A Graph Neural Network based Intrusion Detection System for IoT. In Proceedings of the NOMS 2022–2022 IEEE/IFIP Network Operations and Management Symposium, Budapest, Hungary, 25–29 April 2022; pp. 1–9. [Google Scholar]
  43. Sun, Z.; Teixeira, A.M.; Toor, S. GNN-IDS: Graph Neural Network based Intrusion Detection System. In Proceedings of the ARES ’24: 19th International Conference on Availability, Reliability and Security, Vienna, Austria, 30 July–2 August 2024; Association for Computing Machinery: New York, NY, USA, 2024. [Google Scholar] [CrossRef]
  44. Friji, H.; Olivereau, A.; Sarkiss, M. Efficient Network Representation for GNN-Based Intrusion Detection. In Proceedings of the Applied Cryptography and Network Security, Kyoto, Japan, 19–22 June 2023; Springer: Cham, Switzerland, 2023; pp. 532–554. [Google Scholar]
  45. Xiang, Y.; Li, K.; Zhou, W. Low-rate DDoS attacks detection and traceback by using new information metrics. IEEE Trans. Inf. Forensics Secur. 2011, 6, 426–437. [Google Scholar] [CrossRef]
  46. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  47. Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Toward generating a new intrusion detection dataset and intrusion traffic characterization. ICISSP 2018, 1, 108–116. [Google Scholar]
  48. Liu, L.; Engelen, G.; Lynar, T.; Essam, D.; Joosen, W. Error Prevalence in NIDS datasets: A Case Study on CIC-IDS-2017 and CSE-CIC-IDS-2018. In Proceedings of the 2022 IEEE Conference on Communications and Network Security (CNS), Austin, TX, USA, 3–5 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 254–262. [Google Scholar]
  49. Xu, C.; Shen, J.; Du, X. Low-rate DoS attack detection method based on hybrid deep neural networks. J. Inf. Secur. Appl. 2021, 60, 102879. [Google Scholar] [CrossRef]
  50. Xu, C.; Shen, J.; Du, X. A Method of Few-Shot Network Intrusion Detection Based on Meta-Learning Framework. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3540–3552. [Google Scholar] [CrossRef]
  51. Ilango, H.S.; Ma, M.; Su, R. A feedforward–convolutional neural network to detect low-rate DoS in IoT. Eng. Appl. Artif. Intell. 2022, 114, 105059. [Google Scholar] [CrossRef]
  52. Asgharzadeh, H.; Ghaffari, A.; Masdari, M.; Gharehchopogh, F.S. Anomaly-based intrusion detection system in the Internet of Things using a convolutional neural network and multi-objective enhanced Capuchin Search Algorithm. J. Parallel Distrib. Comput. 2023, 175, 1–21. [Google Scholar] [CrossRef]
  53. Tang, D.; Wang, S.; Liu, B.; Jin, W.; Zhang, J. GASF-IPP: Detection and mitigation of LDoS attack in SDN. IEEE Trans. Serv. Comput. 2023, 16, 3373–3384. [Google Scholar] [CrossRef]
  54. Yin, X.; Fang, W.; Liu, Z.; Liu, D. A novel multi-scale CNN and Bi-LSTM arbitration dense network model for low-rate DDoS attack detection. Sci. Rep. 2024, 14, 5111. [Google Scholar] [CrossRef] [PubMed]
  55. Liu, B.; Tang, D.; Chen, J.; Liang, W.; Liu, Y.; Yang, Q. ERT-EDR: Online defense framework for TCP-targeted LDoS attacks in SDN. Expert Syst. Appl. 2024, 254, 124356. [Google Scholar] [CrossRef]
  56. Lei, F.; Jiang, X.; Jin, G.; Yu, D. HawkEye: An end-host method to detect the Low-rate Denial-of-Service attack of cross-traffic over bottleneck links. Comput. Netw. 2025, 257, 110951. [Google Scholar] [CrossRef]
  57. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  58. Wüsteney, L.; Hellmanns, D.; Schramm, M.; Osswald, L.; Hummen, R.; Menth, M.; Heer, T. Analyzing and modeling the latency and jitter behavior of mixed industrial TSN and DetNet networks. In Proceedings of the the 18th International Conference on Emerging Networking EXperiments and Technologies, Roma, Italy, 6–9 December 2022; pp. 91–109. [Google Scholar]
  59. Bechtel, L.; Muller, S.; Menth, M.; Heer, T. GeNESIS: Generator for Network Evaluation Scenarios of Industrial Systems. In Proceedings of the 2024 IEEE 29th International Conference on Emerging Technologies and Factory Automation (ETFA), Padova, Italy, 10–13 September 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–4. [Google Scholar]
  60. Maaten, L.v.d.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Overall architecture of DELP-Net.
Figure 1. Overall architecture of DELP-Net.
Entropy 28 00328 g001
Figure 2. Architecture of the DREAM module.
Figure 2. Architecture of the DREAM module.
Entropy 28 00328 g002
Figure 3. Entropy pyramid fusion across multi-scale branches with entropy-enhanced representations for window-level embedding (S = 3).
Figure 3. Entropy pyramid fusion across multi-scale branches with entropy-enhanced representations for window-level embedding (S = 3).
Entropy 28 00328 g003
Figure 4. Sensitivity analysis of the Rényi order α in DREAM.
Figure 4. Sensitivity analysis of the Rényi order α in DREAM.
Entropy 28 00328 g004
Figure 5. Robustness of DELP-Net under LDoS attack parameter variations. The perturbation levels correspond to modifying T, L p , and R peak .
Figure 5. Robustness of DELP-Net under LDoS attack parameter variations. The perturbation levels correspond to modifying T, L p , and R peak .
Entropy 28 00328 g005
Figure 6. t-SNE visualization of classifier input embeddings for different models.
Figure 6. t-SNE visualization of classifier input embeddings for different models.
Entropy 28 00328 g006
Table 1. Statistics of the datasets used in the experiments.
Table 1. Statistics of the datasets used in the experiments.
DatasetTotal SamplesBenignAttackProtocolsTrain/Test
CICIDS201762,00031,50030,500TCP/UDP70/30
CICIDS201871,00036,20034,800TCP/UDP70/30
Self-built28,50014,80013,700TCP70/30
Table 2. Window-Based Preprocessing Configuration.
Table 2. Window-Based Preprocessing Configuration.
ItemValue
Window length T w 1.0 s
Window stride T s 0.5 s
Max packets per window K16
Bytes per packet L256
Sequence length for TCN M16
Minimum packets per window K min 8
Label overlap threshold τ 0.2
Table 3. Key Hyperparameters for DELP-Net.
Table 3. Key Hyperparameters for DELP-Net.
ItemValue
OptimizerAdam
Initial learning rate 1 × 10 3
Batch size64
Dropout rate 0.3
Rényi parameter α 2
Entropy regularization weight λ 0.05
TCN layers L t 4
TCN kernel size k t 3
TCN dilation rates { 1 , 2 , 4 , 8 }
Early stopping patience10
Table 4. Per-Attack Detection Results of DELP-Net.
Table 4. Per-Attack Detection Results of DELP-Net.
TypeACC (%)DR (%)FPR (%)Precision (%)F1
Slowloris98.7298.961.5298.490.9872
Slow POST98.7698.601.0898.920.9876
Slow Read98.8298.560.9299.080.9882
Pwnloris98.5498.401.3298.680.9854
Torshammer98.9098.881.0898.920.9890
Httpbog98.8898.761.0099.000.9888
(Average)98.7798.691.1598.850.9877
Table 5. Overall comparison with baselines under window-based evaluation.
Table 5. Overall comparison with baselines under window-based evaluation.
MethodACC (%)DR (%)FPR (%)Precision (%)F1
Entropy-Threshold92.6392.767.5092.520.9264
RF91.9891.957.9992.010.9198
XGBoost91.6491.578.2991.700.9164
1D-CNN94.7094.665.2594.740.9470
LSTM94.8394.925.2694.750.9483
DELP-Net98.7798.691.1598.850.9877
Table 6. Comparison with recent state-of-the-art methods.
Table 6. Comparison with recent state-of-the-art methods.
Method (Year)ACC (%)DR (%)Precision (%)F1
FFCNN (2022) [51]96.3097.500.9690
CNN-BMECapSA-RF (2023) [52]95.9194.6994.950.9482
GASF-IPP (2023) [53]93.5793.2295.390.9429
MSCBL-AND (2024) [54]96.7496.7496.770.9675
ERT-EDR (2024) [55]95.2597.910.9557
HawkEye (2025) [56]91.7896.460.9307
DELP-Net (Proposed)98.7798.6998.850.9877
Table 7. Ablation study results.
Table 7. Ablation study results.
VariantACC
(%)
DR
(%)
FPR
(%)
Precision
(%)
F1
w/o DREAM94.4094.285.4894.510.9439
DREAM to SE95.8595.834.1395.870.9585
DREAM to CBAM96.5696.613.4896.520.9656
w/o Entropy Pyramid Fusion96.7096.663.2696.740.9670
w/o TCN97.6097.872.6697.350.9761
TCN w/o Entropy Conditioning98.2098.121.7198.280.9820
DELP-Net (Full)98.7798.691.1598.850.9877
Table 8. Model complexity comparison.
Table 8. Model complexity comparison.
ModelParameters (M)FLOPs (G)
1D-CNN0.940.41
LSTM1.570.68
DELP-Net1.160.47
Table 9. Inference efficiency and resource usage.
Table 9. Inference efficiency and resource usage.
ModelLatency/Sample (ms)Samples/sGPU Memory (MB)
1D-CNN1.21826486
LSTM2.87348664
DELP-Net1.64610528
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Xu, C.; Yang, J. DELP-Net: A Differentiable Entropy Layer Pyramid Network for End-to-End Low-Rate DoS Detection. Entropy 2026, 28, 328. https://doi.org/10.3390/e28030328

AMA Style

Wang J, Xu C, Yang J. DELP-Net: A Differentiable Entropy Layer Pyramid Network for End-to-End Low-Rate DoS Detection. Entropy. 2026; 28(3):328. https://doi.org/10.3390/e28030328

Chicago/Turabian Style

Wang, Jinyi, Congyuan Xu, and Jun Yang. 2026. "DELP-Net: A Differentiable Entropy Layer Pyramid Network for End-to-End Low-Rate DoS Detection" Entropy 28, no. 3: 328. https://doi.org/10.3390/e28030328

APA Style

Wang, J., Xu, C., & Yang, J. (2026). DELP-Net: A Differentiable Entropy Layer Pyramid Network for End-to-End Low-Rate DoS Detection. Entropy, 28(3), 328. https://doi.org/10.3390/e28030328

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop