The Yellow Active Queue Management Algorithm in ICN Routers Based on the Monitoring of Bandwidth Competition

: Deploying the active queue management (AQM) algorithm on a router is an effective way to avoid packet loss caused by congestion. In an information-centric network (ICN), routers not only play a role of packets forwarding but are also content service providers. Congestion in ICN routers can be further summarized as the competition between the external forwarding trafﬁc and the internal cache response trafﬁc for limited bandwidth resources. This indicates that the traditional AQM needs to be redesigned to adapt to ICN. In this paper, we ﬁrst demonstrated mathematically that allocating more bandwidth for the upstream forwarding ﬂow could improve the quality of service (QoS) of the whole network. Secondly, we propose a novel AQM algorithm, YELLOW, which predicts the bandwidth competition event and adjusts the input rate of request and the marking probability adaptively. Afterwards, we model YELLOW through the totally asymmetric simple exclusion process (TASEP) and deduce the approximate solution of the existence condition for each stationary phase. Finally, we evaluated the performance of YELLOW by NS-3 simulator, and veriﬁed the accuracy of modeling results by Monte Carlo. The simulation results showed that the queue of YELLOW could converge to the expected value, and the signiﬁcant gains of the router with low packet loss rate, robustness and high throughput.


Introduction
According the statistics from Cisco, global IP traffic has quadrupled over the past 5 years. From 2017 to 2022, IP traffic will continue to grow at an average annual compound growth rate of 26% [1]. The existing end-to-end communication mode of TCP/IP has been unable to meet the needs of massive content distribution. The traditional TCP/IP network is facing a series of problems such as scalability, mobility, security. Therefore, a promising future network architecture named information-centric networking (ICN) has attracted researchers' attention, such as Data-Oriented (and Beyond) Network Architecture [2], Content-Centric Network (CCN) [3], Network of Information (NetInf) [4], Publish-Subscribe Internet Routing Paradigm (PSIRP) [5]. Rather than gradually patching on the existing network like Content Delivery Network (CDN) [6] or IPv6 [7], ICN attempts to clean-state update the traditional network. In ICN, ubiquitous cache nodes reduce the user's download latency of content and the drain of network bandwidth, which enables efficient content distribution. Adding a new content store (CS) module with complete ICN protocol stack in each router makes the router become a content provider.
In the early research, the high cost Static Random-Access Memory (SRAM) limited the cache capacity of router to multi hundreds of megabytes (MB) [8]. Nowadays, researchers have carried out extensive studies on how to design content router with cache capacity of multi terabytes (TB) [9][10][11]. For example, in our previous study, in order to make

•
We analyzed how additional end attributes in ICN router contributes to congestion through probability theory. Afterwards, we proposed a novel active queue management algorithm YELLOW based on bandwidth competition prediction and limitation for ICN content routers. YELLOW provides a new idea for the design of ICN congestion control algorithm and cache service module in router.

•
In order to analyze the correlation between the input rate of request packets and congestion, we modeled the packets' dynamics of bandwidth competition in the YELLOW algorithm through the classic totally asymmetric simple exclusion process (TASEP) model, and deduced the approximate solution of the existence condition for each stationary phase.

•
We performed a large number of comparative experiments with several widespread used AMQ algorithms under the ICN environment simulating by NS-3. Experimental results demonstrated the queue of YELLOW can be stable at the expected value, and the significant gains of a router with low packet loss rate, robustness and high throughput. Moreover, we conducted extensive experiments to verify the accuracy of our analysis model for the YELLOW algorithm.
The remainder of this paper is organized as follows. Section 2 provides related research of our work. Section 3 presents the cause and solution analysis of congestion in ICN and the design of YELLOW. The TASEP model analysis of the traffic behavior pattern in YELLOW is presented in Section 4. Our experimental results are shown in Section 5. Finally, we conclude the paper and describe our future work in Section 6.

Active Queue Management
Since the change of internal state of the network can not be detected accurately in the end system, it is more effective and timely to detect and adjust congestion in a router than the window-based TCP congestion control strategy. The simplest drop-tail algorithm is widely used at the earliest, but the long-term full queue will lead to TCP global synchronization, queue deadlock and other network performance problems. Floyd et al. [23] proposed the random early detection (RED) algorithm in 1993 to overcome the shortcomings of drop-tail, which is the earliest standard AQM scheme recommended by the Internet Engineering Task Force (IETF). RED estimates congestion and calculates the marking probability according to the queue length. In the past 20 years, many AQM algorithms have been proposed, which can be divided into two types: one is the optimization algorithm based on RED, the other is designed based on control theory. The performance of the RED algorithm is closely related to the fluctuation of network traffic and the different settings of parameter configuration. Meanwhile, the lack of fairness of RED is neatly illustrated by the fact that the burst flow will quickly occupy the queue resources, which leads to the abnormal transmission of other flow. Therefore, many improved algorithms are proposed, especially GRED [24], ARED [25], SRED [26] and BLUE [27]. SRED algorithm is so complex that it consumes abundant router memory resources, resulting in low scalability. ARED and GRED make the algorithm adaptive and keep the queue length convergent by dynamically adjusting the packet loss probability, although the problem of parameter sensitivity has not been solved perfectly. Compared with the above algorithms, BLUE achieves the lowest packet loss rate with the lowest memory overhead while including a scheduling algorithm based on the Bloom filter to ensure the fairness of each flow.
Misra [28] et al. first proposed a feedback control model based on TCP to optimize the AQM algorithm while they gave the exact solution. The adaptive virtual queue algorithm (AVQ) [29] achieves a proper balance between high utilization and low delay through using linear differential equations to adjust the virtual capacity of the router. However, due to the lack of explicit control to queue length, the performance of AVQ is not ideal. Ünal [30] et al. introduced the proportional integral (PI) controller to maintain the stability of the queue, but it has the defects of slow response speed and overdependence on buffer size. With the improvement of artificial neural network and fuzzy logic control theory, some Electronics 2021, 10, 806 4 of 23 researchers tried to apply intelligent control theory to design an AQM algorithm based on a proportion integral differential (PID) controller. The self-learning ability of a neuron can adjust the three parameters of PID online, so that these algorithms [31,32] can obtain stable performance under the fluctuating network environment.
Not only can control theory be used to model the congestion caused by limited bandwidth and buffer resource of content routers, the well-assessed queuing theory is a common method to analyze and improve the performance of resource-constrained networks. For example, to eliminate the coupling effect between multiple flows in the router on the packet forwarding rate, [33] deployed multiple decoupled queues on a single node and calculated the ideal transmission rate of packets through the M/D/1 queuing model. Besides, [34,35] utilized queuing theory model called Jackson networks (M/M/k open queueing networks) [36] to analyze the dependence between service response time and node capacity constraints, and deduces the optimal cache allocation to meet the required cost/resource tradeoff.
Nowadays, there are few research works that apply the AQM algorithm on ICN. In an ICN-5G network scenario, to satisfy the strict delay constraint from the perspective of transmission control, Shahin et al. proposed a centralized AQM algorithm including load balancing and rerouting [37]. Moreover, Wang et al. [38]. proposed an AQM algorithm in which congestion is detected by monitoring the queuing time of packets. Although this method could achieve high link utilization, the simulation experiment was limited bandwidth to 100 Mbps. We believe that the high complexity of this algorithm may make it difficult for ICN routers to maintain line speed forwarding in actual deployment.

Congestion Control in Information-Centric Network (ICN)
Although ubiquitous cache nodes in ICN reduce redundant transmission traffic on the network and achieve efficient content distribution, congestion may still occur. ICN has yet to determine its approach to congestion control, including whether window-based or rate-based congestion control would be the best. Since the window-based congestion control has been studied and deployed widely in the IP network, the majority of research on ICN congestion control focuses on it as well.
ICP [18] and ICTP [19] detected retransmission timeout (RTO) as congestion signal and adjust congestion window through additive increase multiplicative decrease (AIMD). Since the segments of one content chunk may be scattered around the network, the response delay from different sources varies greatly. Unfortunately, both these two protocols do not consider the influence of multi-source content on RTO value. CCTCP [20] attempted to solve the above problem by maintaining an independent RTO timer for each content source, but the huge cost makes it difficult to extend with a large scale. However, the error of RTO estimation cannot be avoided when the RTT detection is not accurate or the network link is unreliable. M. Amadeo [39] proposed an improved protocol to reduce the error of RTO estimation in the wireless environment of Named Data Networking (NDN) through adjusting the sending gap of interest packets according to the arrival rate of data packets. The "pull" request mode in ICN turn the transmission protocol from the sender-driven to the receiver-driven, which offloads the state management and transmission control to the user. The cache module in the router does not need to save the state of each service flow, which reduces the memory overhead and improves its robustness and performance. Moreover, routers can directly participate in congestion control by shaping the forwarding rate of interest packets, and report congestion information to the requester through NDN acknowledge character (NACK), especially HOBHIS [21] that is a hop by hop congestion control algorithm. [22] considered that the shaping rate should refer to the predicted value of content popularity, which is essentially to reallocate the router bandwidth resources in term of the popularity. However, how to determine and maintain the content popularity for a long time is still an outstanding problem. Karami et al. [40] considered using neural network to predict which link may be congested in the whole network. However, how to design a reasonable neural network model with low training complexity is an urgent problem for this kind of study. The design Electronics 2021, 10, 806 5 of 23 purpose of a loss-based transport protocol is one of the main factor of queue overflow. Therefore, Liu [41] proposed and implemented a delay-based transmission protocol in a wireless environment, which calculates the transmission rate and the maximum transmission window of the request according to the real-time predictive network bandwidth. Experimental results showed that this method improves the flow fairness and robustness, and can effectively reduce the congestion in the network.

The Design of Active Queue Management (AQM) Algorithm for ICN Router
It is still an unsolved problem whether the traditional active queue management algorithm in a TCP network (such as RED, BLUE) is suitable for ICN or not. Firstly, we analyze how the end system added in ICN routers contribute to congestion. Secondly, we present the main ideas on how to relieve congestion caused by this problem. Eventually, we propose a lightweight active queue management algorithm YELLOW for ICN.

Problem Description
Above all, we attempt to explain the cause of congestion from the perspective of the traffic direction into the router. Figure 1 is a simple architecture diagram, which depicts an ICN content router with two ports. The character i represents the request packets. Suppose a request packet arrives at port 0 at a certain time, the ICN router will check whether the requested content has been cached. If not, the packet will be forwarded from port 1 to the next node according to the flow tables. When the requested data have been stored in the router, chunk will be fetched from cache memory and then put on the ICN protocol stack for transmission. The character d represents the data packets that arrive at port 0, which will be stored and forwarded by the router at the same time. Without the support of specific network strategy, the higher the cache-hit ratio in router, the higher the query per second (QPS) will be, which means that the throughput of internal service traffic will try its best to encroach on the router bandwidth. The ICN router at the edge of the network may often encounter bursting flow. For example, users watch a new popular movie at the same time. Although ICN can reduce the transmission traffic in the network, it is difficult to avoid congestion and packet loss in the router caused by the bandwidth competition of flow in different directions.  Specifically, for 1 with bandwidth , bandwidth competition always occurs between external traffic forwarded from 0 and internal traffic generated from CS in response to request traffic. We believe that the main factor of congestion and packet loss in ICN router cannot be explained by buffer overflow simply, but excessive bandwidth competition at a more practical level. As shown in Figure 2, with the increase of the request arrival rate, the throughput of response traffic increases while the available bandwidth of forwarding traffic decreases. At the right side of the figure (gray area), the bandwidth suffers conflict (i.e. + > ).
hput / andwidth port capacity conflict area Specifically, for port 1 with bandwidth B p , bandwidth competition always occurs between external traffic forwarded from port 0 and internal traffic generated from CS in response to request traffic. We believe that the main factor of congestion and packet loss in ICN router cannot be explained by buffer overflow simply, but excessive bandwidth competition at a more practical level. As shown in Figure 2, with the increase of the request arrival rate, the throughput of response traffic τ c increases while the available bandwidth of forwarding traffic τ f decreases. At the right side of the figure (gray area), the bandwidth suffers conflict (i.e. τ f + τ c > B p ). between external traffic forwarded from 0 and internal traffic generated from CS in response to request traffic. We believe that the main factor of congestion and packet loss in ICN router cannot be explained by buffer overflow simply, but excessive bandwidth competition at a more practical level. As shown in Figure 2, with the increase of the request arrival rate, the throughput of response traffic increases while the available bandwidth of forwarding traffic decreases. At the right side of the figure (gray area), the bandwidth suffers conflict (i.e. + > ).

The Priority of Flow in Bandwidth Allocation
With the increase of router cache capacity, cache hit rate and task concurrency will be improved greatly [11]. Therefore, the real-time bandwidth requirement of cache service traffic cannot be ignored and the conflict may occur at any time. Considering that some classic AQM that drop packets entering the buffer indiscriminately (such as RED

The Priority of Flow in Bandwidth Allocation
With the increase of router cache capacity, cache hit rate and task concurrency will be improved greatly [11]. Therefore, the real-time bandwidth requirement of cache service traffic cannot be ignored and the conflict may occur at any time. Considering that some classic AQM that drop packets entering the buffer indiscriminately (such as RED and BLUE), we figure that designing an AQM algorithm that can operate differently to the traffics of two directions can improve the QoS of the ICN network more effectively. Now the question is, which direction traffic should have higher priority in bandwidth allocation and how to take care of it carefully when τ f + τ c > B p . Figure 3 constructs a simple linear network being used to answer the above question from the perspective of probability theory. We allocate port bandwidth by weighting, i.e., λ 1 τ f + λ 2 τ c = B p , where λ 1 , λ 2 ≤ 1. and BLUE), we figure that designing an AQM algorithm that can operate differently to the traffics of two directions can improve the QoS of the ICN network more effectively. Now the question is, which direction traffic should have higher priority in bandwidth allocation and how to take care of it carefully when + > . Figure 3 constructs a simple linear network being used to answer the above question from the perspective of probability theory. We allocate port bandwidth by weighting, i.e., 1 + 2 = , where 1 , 2 ≤ 1.  In Figure 3, 1 , 2 , 3 are three adjacent ICN content routers with CS (dark red square). 1 is the closest to the content source server and the most upstream of the content transmission path. R3 is the closest to the user, so it is located downstream. The data chunk requested by User1 is located in router 1 while User2 request data chunk from the content source server. Many studies indicate that the large size of chunk can improve the link utilization, such as 4K in CCN [3] or even up to few MBs [13]. Owing to the limitation of the maximum transmission unit (MTU), these large chunks will be segmented into a certain number of packets and be transmitted on the network then. As shown in Figure 3, chunk and are segmented into 4 packets. The router will collect these packets and then assemble them into a complete chunk for caching in CS. It is supposed that routers adopting traditional AQM will drop the packets indiscriminately with probability 1 when congestion occurs. In addition, we analyze two ideal cases of the bandwidth allocation with different priorities under the condition of 1 + 2 = . One is 1 = 1. In other word, the router tries its best to allocate sufficient bandwidth for external traffic while the cache service traffic is dropped with probability 2 . The other is 2 = 1, which is the opposite of the above. The router tries its best to In Figure 3, R 1 , R 2 , R 3 are three adjacent ICN content routers with CS (dark red square). R 1 is the closest to the content source server and the most upstream of the content transmission path. R3 is the closest to the user, so it is located downstream. The data chunk m requested by User1 is located in router R 1 while User2 request data chunk k from the content source server. Many studies indicate that the large size of chunk can improve the link utilization, such as 4K in CCN [3] or even up to few MBs [13]. Owing to the limitation of the maximum transmission unit (MTU), these large chunks will be segmented into a certain number of packets and be transmitted on the network then. As shown in Figure 3, chunk k and m are segmented into 4 packets. The router will collect these packets and then assemble them into a complete chunk for caching in CS. It is supposed that routers adopting traditional AQM will drop the packets indiscriminately with probability p 1 when congestion occurs. In addition, we analyze two ideal cases of the bandwidth allocation with different priorities under the condition of λ 1 τ f + λ 2 τ c = B p . One is λ 1 = 1. In other word, the router tries its best to allocate sufficient bandwidth for external traffic while the cache service traffic is dropped with probability p 2 . The other is λ 2 = 1, which is the opposite of the above. The router tries its best to ensure the caching service bandwidth while the external forwarding traffic is dropped with probability p 2 . Noted that we can easily deduce that p 2 > p 1 . Therefore, a single segment of chunk dropped in a router can be regarded as a random independent event and follows the binomial distribution with probability p 1 /p 2 . We supposed that once there is a packet loss in any node, the chunk that loss packet belongs to will be unable to cache in all the subsequent nodes on the transmission path. The reason is that in a realistic environment, RTT and the queuing delay are enough to release the assembling space allocated from the router memory for this chunk before all the retransmission packets arrive. Tables 1 and 2 summarize the probability and expectation of the number of cache nodes and retransmission for chunk m and k, respectively. As the expectation analyses summarized in Tables 1 and 2 show, we can draw a conclusion that the forwarding traffic from upstream deserves more bandwidth, which can bring better benefit to QoS of cache and transmission efficiency when congestion occurs. Note that the linear topology is an important part of complex networks, so this conclusion can be easily extended to any network node. Table 1. The probability and expectation of the number of cache nodes for chunk m and k.

Strategy Chunk Cache in 3 Nodes Cache in 2 Nodes Cache in 1 Node Expectation
Forwarding without distinction Forwarding internal traffic from CS with high priority (λ 2 = 1) Table 2. The probability and expectation of retransmission for chunk m and k.

Strategy Chunk Retransmission Hop Expectation
Forwarding without distinction

The Design of YELLOW Algorithm
In light of the above analysis, we consider that designing an active queue management algorithm for ICN should assign bandwidth with higher priority to the traffic from upstream nodes. Therefore, we propose a novel AQM algorithm called YELLOW, which employs dual queue and real-time requests arrive to monitor congestion caused by bandwidth competition between forwarding and caching. Algorithm 1 shows the details of YELLOW, which differs from lots of known AQM algorithms that only use the instantaneous occupancy of a single queue to identify congestion [23][24][25][26][27]. Without increasing the memory overhead, YELLOW sets two queues Q f and Q c in which the forwarding traffic from external and the cache response traffic from internal content stores are queued. Yellow maintains two probabilities p f and p c to drop (mark) packets before they enter the queue. The instantaneous queue length is still a significant monitoring parameter. YELLOW will always firstly increase p c when the packet loss or instantaneous queue length exceed the threshold (L 1 or L 2 ), while the change of p f depends on the more restrictions. In ICN, every data packet segmented from one chunk needs to be pulled to user by one request packet [20]. Therefore, the throughput τ c required by the cache service can be well predicted by the arrival rate of request γ (γ ≤ τ c ). If γ ≥ B − τ f at a certain time, it means that the packet loss rate will rise sharply because of the rapid queue overflow. YELLOW will limit the rate of request packets entering the router by token bucket algorithm, which allows YELLOW to respond quickly to excessive bandwidth competition. Note that by contrast with some AQM algorithms who shape the rate of the request forwarding outward, YELLOW adjusts the rate of the request entering CS. If the explicit congestion notification (ECN) mechanism is allowed, router will set the ECN field in packet header to 1 to inform the requester. On the contrary, if the queue length is lower than the threshold or the link is idle, YELLOW will reduce the marking probability. δ 1 , δ 2 , δ 3 are the values with which the marking probability increases or decreases when an abnormal condition is detected. Moreover, they are set to provide the link ability to effectively adapt to macroscopic changes in load across the link. The value of δ 1 should be several orders of magnitude greater than that of δ 2 , δ 3 , which also reflects our biased treatment on cache response traffic and p f should be set to 0 at the beginning. freeze_time is the minimum interval of each update to the marking probability. The value of freeze_time should be set with reference to the effective RTT of the cache service flow in router, so as to allow the changes to be adequately fed back to the requester at the time interval. Over typical links, using freeze_time values between 10 and 200 millisecond (ms) will allow the marking probability to range from 0 to 1 on the order of 5 to 15 s.
Limiting the rate of the hit request to B − τ f 6.
set ECN_FLAG in response packet 8.

Modeling
In this chapter, we model the forwarding behavior of packets of the YELLOW algorithm through TASEP [42]. We first introduce TASEP model briefly, and then deduce the existence conditions for each stable phase of YELLOW.

Totally Asymmetric Simple Exclusion Process (TASEP)
TASEP is the most studied model, which is often used to analyze the dynamics and phase transition process of moving objects. The rules of particle motion in the TASEP model are simple, which makes the mathematical analysis easy to deduce. TASEP and its related models can be used to effectively analyze many common statistical physical phenomena, such as simulating road traffic flow and transport process of biological protein molecules. In some relatively simple scenes where the TASEP model is applied, the exact solution can be obtained easily by the Bethe ansatz method [43].
To facilitate understanding, we introduce TASEP through modeling traffic flow on the road. Figure 4 depicts a scene in which many vehicles are driving on a road that can be regarded as a one-dimensional queue with L lattices. As a particle occupying the lattice of the queue, vehicle can only move forward one step when the next lattice in the forward direction is empty. (The model turns to an asymmetric simple exclusion process (ASEP) when a particle in the queue may move forward or backward with different probability at the same time [44].) The vehicle enters the first lattice of the queue with probability α if there is no occupation by any particle. Similarly, the vehicles at the end of the queue jump out of the system with probability β at each time step. Each lattice in the system is either empty or occupied by vehicles.

Totally Asymmetric Simple Exclusion Process (TASEP)
TASEP is the most studied model, which is often used to analyze the dynamics and phase transition process of moving objects. The rules of particle motion in the TASEP model are simple, which makes the mathematical analysis easy to deduce. TASEP and its related models can be used to effectively analyze many common statistical physical phenomena, such as simulating road traffic flow and transport process of biological protein molecules. In some relatively simple scenes where the TASEP model is applied, the exact solution can be obtained easily by the Bethe ansatz method [43].
To facilitate understanding, we introduce TASEP through modeling traffic flow on the road. Figure 4 depicts a scene in which many vehicles are driving on a road that can be regarded as a one-dimensional queue with L lattices. As a particle occupying the lattice of the queue, vehicle can only move forward one step when the next lattice in the forward direction is empty. (The model turns to an asymmetric simple exclusion process (ASEP) when a particle in the queue may move forward or backward with different probability at the same time [44].) The vehicle enters the first lattice of the queue with probability α if there is no occupation by any particle. Similarly, the vehicles at the end of the queue jump out of the system with probability β at each time step. Each lattice in the system is either empty or occupied by vehicles. System may be in the following three different stationary phases when α, β take different values.

Road L
1. α < β and α < 0.5. The system is in low-density (LD) phase, which can be regarded as an unblocked road. In LD, the volume of flow J and the particle density ρ of the model are determined by the entering probability α. Namely, where the subscripts 1, L and bulk denote the head, tail and the middle lattice of TASEP queue, respectively.
2. α > β and β < 0.5. The system is in high-density (HD) phase, which can be seen as a congested road. In HD, the volume of flow J and the particle density ρ of the model are determined by the ejecting probability β. Namely, Moreover, when α = β < 0.5, the system is in a coexistence phase between LD and HD, which is divided by a shock. From this example, we can easily point out that α and β can be understood as the normalized relative value between the entrance (exit) rate of the particle and the capacity of intersection in TASEP queue.

Modeling the Dynamics of YELLOW Algorithm through TASEP
Similarly, we can use TASEP to model the forwarding behavior of packet of YEL-LOW algorithm in ICN router. Queuing packets in buffer can be regarded as identical particles that move along the lattice. The system consists of three queues with a junction System may be in the following three different stationary phases when α, β take different values.

1.
α < β and α < 0.5. The system is in low-density (LD) phase, which can be regarded as an unblocked road. In LD, the volume of flow J and the particle density ρ of the model are determined by the entering probability α. Namely, where the subscripts 1, L and bulk denote the head, tail and the middle lattice of TASEP queue, respectively.

2.
α > β and β < 0.5. The system is in high-density (HD) phase, which can be seen as a congested road. In HD, the volume of flow J and the particle density ρ of the model are determined by the ejecting probability β. Namely, J = β/(1 − β), α > 0.5 and β > 0.5. The system is in maximum-current (MC) phase in which the maximum number of vehicles running on the road. And J = 0.25, ρ 1 = 1 − 1 4 α, ρ L = β 4 , ρ bulk = 0.5. Moreover, when α = β < 0.5, the system is in a coexistence phase between LD and HD, which is divided by a shock. From this example, we can easily point out that α and β can be understood as the normalized relative value between the entrance (exit) rate of the particle and the capacity of intersection in TASEP queue.

Modeling the Dynamics of YELLOW Algorithm through TASEP
Similarly, we can use TASEP to model the forwarding behavior of packet of YELLOW algorithm in ICN router. Queuing packets in buffer can be regarded as identical particles that move along the lattice. The system consists of three queues with a junction positioned in the middle of the system as shown in Figure 5. Each queue contains L sites. queue 1 and queue 2 merge together at site 2L + 1. The external forwarding traffic/internal cache service traffic enters queue 1/queue 2 for buffering, and then converges into the port's forwarding queue queue 3 for actual forwarding. Packets enter the queue 1/queue 2 with the rate α 1 /α 2 (such as arriving following Poisson distribution). The packets can leave the queue 1/queue 2/system with the rate β 1 /β 2 /β 3 . Obviously, in some cases, queue 3 will enter the HD phase because of too much packets entering from queue 1 and queue 2, which can be used to simulate congestion on the router due to bandwidth conflict. The dynamics of this system modeling is similar to the traffic flow in crossing where two branch roads merge into one primary road. In particular, the primary road is often blocked as well in real life.
Electronics 2021, 10, x FOR PEER REVIEW 11 of 24 positioned in the middle of the system as shown in Figure 5. Each queue contains L sites. queue 1 and queue 2 merge together at site 2L + 1. The external forwarding traffic/internal cache service traffic enters queue 1/queue 2 for buffering, and then converges into the port's forwarding queue queue 3 for actual forwarding. Packets enter the queue 1/queue 2 with the rate 1 / 2 (such as arriving following Poisson distribution). The packets can leave the queue 1/queue 2/system with the rate 1 / 2 / 3 . Obviously, in some cases, queue 3 will enter the HD phase because of too much packets entering from queue 1 and queue 2, which can be used to simulate congestion on the router due to bandwidth conflict. The dynamics of this system modeling is similar to the traffic flow in crossing where two branch roads merge into one primary road. In particular, the primary road is often blocked as well in real life.  Since only three stationary phases can be found in each queue, there are possible 3 3 = 27 stationary phases in our system adopting YELLOW algorithm theoretically. However, the number of possible stationary phases decreases dramatically due to many constraints. For example: 1. Since 3 can be regarded as the normalization of forwarding capacity of router port, following inequalities are universal: 1 < 3 , 2 < 3 , 3 > 0.5. Therefore, queue 3 can only present maximal-current phase or low-density phase. This conclusion can be easily understood, because the purpose of YELLOW algorithm is to avoid keeping the congestion state (HD phase) in the router forwarding queue for a period. 2. According to the principle of flow conservation, the total current passing through the system is equal to: where 1 is the volume of flow of queue 1 and 2 is that of queue 2. It suggests that queue 3 cannot have the low-density phase when either queue 1 or queue 2 is in maximal-current phase or High-density phase. Similarly, queue 1 and queue 2 can not be in maximal-current phase at the same time when queue 3 is suffering bandwidth competition (namely at HD phase). Thus, we infer that there are only six stationary phases in the system: , where X/Y/Z in the expression (X,Y,Z) describes the phase in the queue 1/2/3 respectively. Afterwards, we attempt to deduce the approximate solution of existence conditions for these six stationary phases through mean-field theory [45] that neglects the occupation correlations of the sites before and after the junction. Mean-field approximation decomposes many-body problem into multiple one-body problem, which is a common method to study the coupling of multiple TASEP queues. Jonathan C et al. [46] have proved that the power spectrum characteristics of three queues with one junction is the same as that of a single TASEP queue. Hence, our system can be analyzed form the perspective of simple coupling of three separate TASEP queues, as the known conclusions introduced in Section 4.1 can be easily utilized such as phase diagram, particle currents and density Since only three stationary phases can be found in each queue, there are possible 3 3 = 27 stationary phases in our system adopting YELLOW algorithm theoretically. However, the number of possible stationary phases decreases dramatically due to many constraints. For example:

1.
Since β 3 can be regarded as the normalization of forwarding capacity of router port, following inequalities are universal: α 1 < β 3 , α 2 < β 3 , β 3 >0.5. Therefore, queue 3 can only present maximal-current phase or low-density phase. This conclusion can be easily understood, because the purpose of YELLOW algorithm is to avoid keeping the congestion state (HD phase) in the router forwarding queue for a period.

2.
According to the principle of flow conservation, the total current passing through the system is equal to: where J 1 is the volume of flow of queue 1 and J 2 is that of queue 2. It suggests that queue 3 cannot have the low-density phase when either queue 1 or queue 2 is in maximal-current phase or High-density phase. Similarly, queue 1 and queue 2 can not be in maximal-current phase at the same time when queue 3 is suffering bandwidth competition (namely at HD phase). Thus, we infer that there are only six stationary phases , where X/Y/Z in the expression (X,Y,Z) describes the phase in the queue 1/2/3 respectively. Afterwards, we attempt to deduce the approximate solution of existence conditions for these six stationary phases through mean-field theory [45] that neglects the occupation correlations of the sites before and after the junction. Mean-field approximation decomposes many-body problem into multiple one-body problem, which is a common method to study the coupling of multiple TASEP queues. Jonathan C et al. [46] have proved that the power spectrum characteristics of three queues with one junction is the same as that of a single TASEP queue. Hence, our system can be analyzed form the perspective of simple coupling of three separate TASEP queues, as the known conclusions introduced in Section 4.1 can be easily utilized such as phase diagram, particle currents and density profiles. In order to facilitate the analysis, we introduce the intermediate variable α e f f as the equivalent entry rate of queue 3.
Firstly, we analyze the simplest stationary phase (LD, LD, LD), which exists when the following inequality is fulfilled: , α e f f < β 3 (2) According to Equation (1), we can deduce that Thus, the possible conditions for the existence of phase (LD, LD, LD) is as follows, For phase (LD, LD, MC), it should satisfy, Similarly, the possible conditions for the existence of phase (LD, LD, MC) is as follows, So the only solution of the equation is, The (LD, HD, MC) phase is defined by the following expression: Taking into account Equation (1), we obtain: At this time, the current on queue 2 (J 2 ) should be more than that of queue 1 (J 1 ). We could obtain: Hence, we can deduce that α 1 < 2− √ 2 4 from (10). The known conclusion of mean-filed theory for the current [] can be utilized in in our system as follows, where . . . denotes a statistical average. τ i = 1 (or τ i = 0) indicates that the state of the ith lattice is occupied (or free). Thus the current of each queue can be written as: Therefore, the α e f f can be deduced from (11) and (12) as follows: From (11)-(13), we find the system is in the (LD, HD, MC) phase when: Similarly, for phase (HD, LD, MC), it should satisfy, We obtain the following equations according to (1): At this time, the current on queue 2 (J 2 ) should be less than that of queue 1 (J 1 ). We could obtain: Hence, the Equation (15) can be rewritten as: The (LD, MC, MC) is defined by the following expression: According to the law of conservation of flow in (1), we can obtain the following equation According to the law of conservation of flow in (1), we can obtain the following equation J 3 1 4 = J 1 1 4 + J 2 . Hence, J 2 ≈ 0.

Performance Evaluation
In this chapter, firstly, we evaluate the performance of algorithm YELLOW in a small linear topology network. Afterwards, we use Monte Carlo simulation method to verify the accuracy of the approximate solution for the existence of conditions of each phase in YELLOW deduced by the TASEP model in the previous section.

The Performance of the YELLOW Algorithm
To evaluate the performance of YELLOW, we carry out packet-level simulations using the NS-3 simulator [47], which is based on a discrete event. NS-3 abstracts the topologies, channels, network protocols, and some commonly network element devices used in real networks, and encapsulates them into common C++ class objects. We deploy ICN packets upon the IP layer in the form of overlay, so that routing can be easily implemented through the default IP routing rather than complex routing-by-naming scheme like CCN [3]. We adopt the lookup-by-name mechanism [11,48] to help user find the named data chunk (NDC). Specifically, a hash table is set to record the mapping of the name of NDCs and their global addresses.
A robust and receiver-driven transport protocol, content-centric TCP (CCTCP) [20], is implemented to guarantee the transmission quality, which supports the mechanism of timeout request retransmission based on RTT and retransmission time-out (RTO) estimates. Besides, CCTCP limits the user's send window of NDC request in the AIMD model like TCP to avoid congestion. We deploy the algorithm YELLOW on the bottleneck link in a small linear topology network as shown in Figure 6. For a thorough performance comparison, we also implement two other classical AQM algorithms which are RED and BLUE. The ICN client application is implemented and installed in the nodes(N 1 , N 2 , N 3 ) on the left side of the graph, which sends 10 requests for new NDC each seconds. The ICN server application is implemented and installed in the nodes(N 4 , N 5 , N 6 ) on the right side and the routers (A, B, C), which are responsible for responding the requests form client. As the most basic component in each end system, the ICN protocol stack ought to assemble received ICN packets to a chunk completely and sequentially.
Electronics 2021, 10, x FOR PEER REVIEW 14 of 24 scheme like CCN [3]. We adopt the lookup-by-name mechanism [11,48] to help user find the named data chunk (NDC). Specifically, a hash table is set to record the mapping of the name of NDCs and their global addresses. A robust and receiver-driven transport protocol, content-centric TCP (CCTCP) [20], is implemented to guarantee the transmission quality, which supports the mechanism of timeout request retransmission based on RTT and retransmission time-out (RTO) estimates. Besides, CCTCP limits the user's send window of NDC request in the AIMD model like TCP to avoid congestion. We deploy the algorithm YELLOW on the bottleneck link in a small linear topology network as shown in Figure 6. For a thorough performance comparison, we also implement two other classical AQM algorithms which are RED and BLUE. The ICN client application is implemented and installed in the nodes( 1 , 2 , 3 ) on the left side of the graph, which sends 10 requests for new NDC each seconds. The ICN server application is implemented and installed in the nodes( 4 , 5 , 6 ) on the right side and the routers (A, B, C), which are responsible for responding the requests form client. As the most basic component in each end system, the ICN protocol stack ought to assemble received ICN packets to a chunk completely and sequentially. Chunk size is set to 256 KB to allow enough time for the AQM algorithms to make the client-end to adjust the request rate when congestion occurs. The workload extracted from catalog of N = 1 × 10 5 content objects follows the Zipf distribution (α = 0.8). The size of the cache store in router is 100 (chunks). The buffer of the queue and assembling memory of cache module accommodates 200 packets which, for 1500-byte packets, corresponds to a queuing delay of 0.35 s. In order to ensure the same memory cost of YEL-LOW algorithm and comparison algorithms, the length of caching queue and forwarding queue are set to 100. Tables 3 and 4 shows the configurations used for RED and BLUE respectively, which is set according to the experience of previous studies [23,27].  Chunk size is set to 256 KB to allow enough time for the AQM algorithms to make the client-end to adjust the request rate when congestion occurs. The workload extracted from catalog of N = 1 × 10 5 content objects follows the Zipf distribution (α = 0.8). The size of the cache store in router is 100 (chunks). The buffer of the queue and assembling memory of cache module accommodates 200 packets which, for 1500-byte packets, corresponds to a queuing delay of 0.35 s. In order to ensure the same memory cost of YELLOW algorithm and comparison algorithms, the length of caching queue and forwarding queue are set to 100. Tables 3 and 4 shows the configurations used for RED and BLUE respectively, which is set according to the experience of previous studies [23,27]. In the following experiments, we find that even if YELLOW is set with different configurations, as long as it meets the parameter setting restrictions introduced in Section 3.3, the performance of YELLOW is not significantly different. For convenience of description, we set some configuration parameters to be the same as BLUE's, where freeze time = 50 ms, δ 1 = 0.02, δ 2 = 0.0025, δ 3 = 0.001, L 1 = 80 and L 2 = 50, mainly because the algorithm YELLOW is designed with reference to BLUE. The simulation program runs for 160 s. The first 80 s are used to warm up the cache space of routers, and the remaining time is used to gather statistics. Specifically, the request rate is doubled at 20 s to simulate the more serious network congestion. At 60 s, the request rate is reduced by one third to simulate the idle network. The sampling frequency of statistical data is 20 times per second, and then the average value is picked. To study the performance of algorithms under different ICN cache strategies, we also build two network scenarios adopting leave copy every (LCE) and leave copy down (LCD) [49], respectively. LCE caches chunks in every node along the transmission path. LCD improves the cache hit rate through storing the chunks only in the next hop of the previous cache node, which puts the popular contents close to users gradually. In order to fully understand the difference between YELLOW, RED and BLUE, we first observe the change of the queue length and marking probability under different loads and network scenarios as shown in Figures 7-9.  In the following experiments, we find that even if YELLOW is set with different configurations, as long as it meets the parameter setting restrictions introduced in Section 3.3, the performance of YELLOW is not significantly different. For convenience of description, we set some configuration parameters to be the same as BLUE's, where freeze time = 50 ms, 1 = 0.02, 2 = 0.0025, 3 = 0.001, 1 = 80 and 2 = 50, mainly because the algorithm YELLOW is designed with reference to BLUE. The simulation program runs for 160 s. The first 80 s are used to warm up the cache space of routers, and the remaining time is used to gather statistics. Specifically, the request rate is doubled at 20 s to simulate the more serious network congestion. At 60 s, the request rate is reduced by one third to simulate the idle network. The sampling frequency of statistical data is 20 times per second, and then the average value is picked. To study the performance of algorithms under different ICN cache strategies, we also build two network scenarios adopting leave copy every (LCE) and leave copy down (LCD) [49], respectively. LCE caches chunks in every node along the transmission path. LCD improves the cache hit rate through storing the chunks only in the next hop of the previous cache node, which puts the popular contents close to users gradually. In order to fully understand the difference between YELLOW, RED and BLUE, we first observe the change of the queue length and marking probability under different loads and network scenarios as shown in   The stability of queue length is an important index to judge the performance of an AQM algorithm. We observe that whether the network congestion is not obvious (<20 s) or the sudden traffic change (at 20 s or 60 s), the queue length amplitude of RED is the largest, which means that the serious global synchronization is happening in the net-  The stability of queue length is an important index to judge the performance of an AQM algorithm. We observe that whether the network congestion is not obvious (<20 s) or the sudden traffic change (at 20 s or 60 s), the queue length amplitude of RED is the largest, which means that the serious global synchronization is happening in the net-  The stability of queue length is an important index to judge the performance of an AQM algorithm. We observe that whether the network congestion is not obvious (<20 s) or the sudden traffic change (at 20 s or 60 s), the queue length amplitude of RED is the The stability of queue length is an important index to judge the performance of an AQM algorithm. We observe that whether the network congestion is not obvious (<20 s) or the sudden traffic change (at 20 s or 60 s), the queue length amplitude of RED is the largest, which means that the serious global synchronization is happening in the network. In the network scenarios with LCE, it is difficult for RED to keep the queue length between [min th , max th ] even when the traffic load is light (after 60 s). Blue identifies whether congestion occurs through events of packet loss or link idle. BLUE performs better than RED because its marking probability is adjusted dynamically by increment. Although the memory utilization of the router has been improved, the queue length amplitude of BLUE is still large than YELLOW. Figure 9c shows that even if the network is busy, the oscillation amplitude of both forwarding queue (up) and caching service queue (down) in YELLOW can be maintained at a small level. Since the monitoring mechanism of bandwidth competition make a good prediction of the burst flow from the router, YELLOW also performs well in dealing with the burst traffic.
Marking probability is another important index to judge the algorithm performance, which allows us to explore the effectiveness of AQM in preventing global synchronization and in removing biases against burst sources. For RED, marking probability p is equal to p b /1 − count × p b , where p b = max p × Q avg − min th /(max th − min th ) . We can find that p b is linear function of Q avg . The large shock of the average queue length will result in the marked probability changing significantly in a very short time (see Figure 7c,d), which makes RED cannot relieve the synchronization phenomenon between content sources. By contrast, the marking probability of BLUE is relatively stable, which marks packets evenly even when the traffic is constantly fluctuating. The marking probability of YELLOW cache queue fluctuates while the marking probability of the forwarding queue can also keep a uniform value for a period of time. This is because YELLOW is self-adaptive, so it will adjust the send window of each cache service flow when router detects the change of forwarding throughput.
Moreover, the ICN protocol stack in router enables YELLOW to ensure the fairness of each burst response traffic easily. In general, YELLOW can perform better than BLUE/RED and obtain better robustness while eliminating the negative effects of global synchronization. Meanwhile, we find that different caching strategies also affect the performance of AQM algorithm. Compared with LCE, the LCD strategy improves cache-hit ratio and reduces the transmission traffic on the link greatly [49]. Therefore, no matter which AQM algorithm is adopted in the network with LCD, the fluctuation amplitude of the queue length is lower than that of LCE. Figures 7d, 8c and 9c prove that the mark probability of AQM changes more smoothly in the network with LCD.
In order to better compare the performance of the three algorithms, Tables 5 and 6 list the bottleneck link utilization, throughput, download latency, average packet loss rate and the mean square error of queue length of the three algorithms under different network loads. We find that algorithm YELLOW can achieve the highest bottleneck link utilization and router throughput. Since low packet loss rate will increase the queuing delay, YELLOW with the lowest packet loss rate has the highest content download delay, followed by BLUE and RED. In terms of packet loss rate, YELLOW is also the best, Comparing the mean square error of queue length of three algorithms, we can obtain the same conclusion as the previous analysis about queue stability.  Table 6. Measurement data for heavy load network in time interval [20,40].

Verifying the Accuracy of the Model of YELLOW through Monte Carlo Simulation
We build a simulation system that is same as the model shown in Figure 5, which contains three queues (queue 1, queue 2 and queue 3) by the Monte Carlo method. queue 1 simulates buffer situation of forwarding traffic that comes from the external network while queue 2 simulates the situation of the cache traffic generated internally. The site scope of queue 1/queue 2/queue 3 with capacity L ranges from 1/L + 1/2L + 1 to L/2L/3L. In each Monte Carlo step (MCS), the system employing the random update mechanism updates 3L times according to the following rules:
If i = 1/(L + 1), it means that the first site of queue 1/queue 2 is selected. If there is no particle in this site, a new particle will be injected to system with probability α 1 /α 2 .

3.
If i = 3L, it means that the output site of queue 3 is selected. A particle locating in this site will leaves the system with probability β 3 when the site is occupied 4.
If the selected site is occupied by particle and the adjacent site on the right is empty, the particle will move to the adjacent site. 5.
In addition to the above cases, we need to add some rules of YELLOW algorithm. 6.
When the number of particles whether in queue 1 or queue 2 exceeds the threshold ( L 2 in this experiment), α 2 will be decreased by a fixed value δ 1 firstly. α 1 will be reduced by δ 2 only if the queue occupancy of queue 1 exceeds 50%. Noted that the value of δ 1 is setting more than one order of magnitude larger than δ 2 . 7.

8.
System increase the value of α 1 and α 2 by δ 3 when the stationary phase (LD, LD, LD) exists for a period of time.
To simplify the parameter of the existence condition of the stationary phase, we consider indicating α 2 by nα 1 (0 < n ≤ 1). Therefore, Equations (4), (6) and (18) can be rewritten by (LD, LD, LD): , β 3 < 1 2 (22) (LD, LD, MC): (HD, LD, MC): , β 3 > 1 2 (24) Figure 10 presents the phase diagram of model system on α 1 − β 3 axis when n is fixed at 0.5. Points and lines correspond to the results of simulation and analysis respectively. We find that there are symbiotic stationary phases in some phases. in (HD, LD, MC). The symbiotic phase exists because the system is simulating the YEL-LOW algorithm to relieve congestion so that probability α 1 , α 2 will change in a small range over a period of time. However, due to the phase transition occurring in the time domain, the shock can not be seen in this coordinate system. The reason why there is no stationary coexistence phase in (LD, MC, MC) is that the marking probability of queue 2 changes dramatically, which makes jumping phases only exist in a very short time. By contrast, the marking probability of queue 1 changes slowly and uniformly making it easy to stabilize for jumping phases (see experiment Figure 9c).  ( ,  ,  ),(LD, LD, MC) in (HD, LD, MC). The symbiotic phase exists because the system is simulating the YELLOW algorithm to relieve congestion so that probability 1 , 2 will change in a small range over a period of time. However, due to the phase transition occurring in the time domain, the shock can not be seen in this coordinate system. The reason why there is no stationary coexistence phase in ( , , ) is that the marking probability of queue 2 changes dramatically, which makes jumping phases only exist in a very short time. By contrast, the marking probability of queue 1 changes slowly and uniformly making it easy to stabilize for jumping phases (see experiment Figure 9c).
Besides, as shown in Figure 11, with the decrease (increase) of n, the areas occupied by ( , ,  )   . Figure 11. System phase diagram with variable n. Figure 12 presents the particle density distribution of six stationary phase running under specific parameters. The coordinate-axis X is the sequential site in the queue 1/2/3. Points and lines correspond to the results of simulation and analysis respectively. The simulation results are in good agreement with the analytical results. For the LD phase in Figure 11. System phase diagram with variable n. Figure 12 presents the particle density distribution of six stationary phase running under specific parameters. The coordinate-axis X is the sequential site in the queue 1/2/3. Points and lines correspond to the results of simulation and analysis respectively. The simulation results are in good agreement with the analytical results. For the LD phase in the system, the density distribution is flat except near the queue boundary. Due to the adaptive adjustment of the YELLOW algorithm, the density distribution of queue 3 can converge to a fixed value around 0.5 regardless of the existence of shock at the queue boundary. Because the yellow algorithm reduces the packet loss rate of queue 1 in order to ensure the forwarding bandwidth, even if both queue 2 and queue 1 are in the LD phase at the same time, the particle density of queue 1 is significantly higher than that of queue 2. The shock appears at the head and tail of queue 3 at the same time, which is caused by the oscillation of the YELLOW algorithm in the time domain. Due to the coexistence phase existing in YELLOW modeling by TASEP, it is difficult to analyze the particle density distribution on each phase boundary by traditional domain-wall theory [50]. Therefore, we will not carry out the experimental analysis here, and will continue to study in the future. Figure 11. System phase diagram with variable n. Figure 12 presents the particle density distribution of six stationary phase running under specific parameters. The coordinate-axis X is the sequential site in the queue 1/2/3. Points and lines correspond to the results of simulation and analysis respectively. The simulation results are in good agreement with the analytical results. For the LD phase in the system, the density distribution is flat except near the queue boundary. Due to the adaptive adjustment of the YELLOW algorithm, the density distribution of queue 3 can converge to a fixed value around 0.5 regardless of the existence of shock at the queue boundary. Because the yellow algorithm reduces the packet loss rate of queue 1 in order to ensure the forwarding bandwidth, even if both queue 2 and queue 1 are in the LD phase at the same time, the particle density of queue 1 is significantly higher than that of queue 2. The shock appears at the head and tail of queue 3 at the same time, which is caused by the oscillation of the YELLOW algorithm in the time domain. Due to the coexistence phase existing in YELLOW modeling by TASEP, it is difficult to analyze the particle density distribution on each phase boundary by traditional domain-wall theory [50]. Therefore, we will not carry out the experimental analysis here, and will continue to study in the future.

Discussion
In this paper, we consider that one of the main factors for congestion and packet loss in ICN routers with an additional end system can be further summarized as the competition between the external forwarding traffic and the internal cache response traffic for limited bandwidth resources. We propose an adaptive active queue management algorithm with dual queues for ICN routers, which predicts and limits the throughput of the response flow of burst request while guaranteeing the throughput of forwarding flow. We prove that ensuring the priority of forwarding traffic when allocating the bandwidth can achieve better benefits mathematically through probability theory. In addition, when bandwidth competition is predicted, YELLOW will limit the rate of request packets to enter the cache module of the router through a token bucket algorithm. In order to study the effect of packet arrival rate on the performance of the YELLOW algorithm, we model YELLOW through the widely studied TASEP model, and give the approximate solution of each stationary phase and the particle density.
Compared with the other two classical AQM algorithms (RED and BLUE) in network scenarios with different ICN cache strategies and load, the simulation results

Discussion
In this paper, we consider that one of the main factors for congestion and packet loss in ICN routers with an additional end system can be further summarized as the competition between the external forwarding traffic and the internal cache response traffic for limited bandwidth resources. We propose an adaptive active queue management algorithm with dual queues for ICN routers, which predicts and limits the throughput of the response flow of burst request while guaranteeing the throughput of forwarding flow. We prove that ensuring the priority of forwarding traffic when allocating the bandwidth can achieve better benefits mathematically through probability theory. In addition, when bandwidth competition is predicted, YELLOW will limit the rate of request packets to enter the cache module of the router through a token bucket algorithm. In order to study the effect of packet arrival rate on the performance of the YELLOW algorithm, we model YELLOW through the widely studied TASEP model, and give the approximate solution of each stationary phase and the particle density.
Compared with the other two classical AQM algorithms (RED and BLUE) in network scenarios with different ICN cache strategies and load, the simulation results showed that the queue of YELLOW can be stable at the expected value, and the significant gains of router with low packet loss rate, robustness and high throughput. We verify the accuracy of the module results of YELLOW through Monte Carlo approximation while we find an interesting phenomenon that there are some symbiotic phases appearing in the system phase diagram. This is because the input/output rate of TASEP queues are adjusted adaptively in time domain when the YELLOW algorithm detects congestion.
We hope that this paper can provide ideas for more research on optimizing the existing ICN transport protocol and congestion control. Moreover, we will continue to study how to analyze and deduce the approximate solution of the existence condition of the symbiotic phase more accurately. In the future, we are interested in further research on how to ensure the fairness of each burst response flow in the router, and how to analyze the particle density distribution on the phase boundary.