An ICN-Based On-Path Computing Resource Scheduling Architecture with User Preference Awareness for Computing Network

: The Computing Network is an emerging network paradigm that aims to realize computing resource scheduling through the intrinsic capabilities of the network. However, existing resource scheduling architectures based on conventional TCP/IP networks for the Computing Network suffer from deficiencies in routing flexibility and a lack of user preference awareness, while Information-Centric Networking (ICN) holds the potential to address these issues. ICN inherently supports dynamic routing in scenarios such as multi-homing and mobility due to its routing mechanism that is based on content names rather than host addresses, and it is further enhanced by the integration with Software-Defined Networking (SDN) technologies, which facilitate convenient network-layer route readdressing, thus offering a conducive environment for flexible routing scheduling. Furthermore, ICN introduces novel routing protocols that, compared with the more rigid protocol designs in conventional TCP/IP networks, offer greater flexibility in field usage. This flexibility allows for the incorporation of customized fields, such as “preference”, enabling the perception of user preferences within the network. Therefore, this paper proposes a novel ICN-based on-path computing resource scheduling architecture named IPCRSA. Within this architecture, an original design for computing resource request packet format is developed based on the IPv6 extension header. Additionally, preference-based computing resource scheduling strategies are incorporated, which employ the technique for order preference by similarity to ideal solution (TOPSIS) combined with the entropy weight method, to comprehensively evaluate computing resource nodes and use a roulette-selection algorithm to accomplish the probability selection of destination nodes. Experimental results indicate that, in comparison to alternative scheduling schemes, IPCRSA exhibits significant advantages in enhancing scheduling flexibility, improving scheduling success rates, and catering to diverse user requirements.


Introduction
The development of the digital economy has driven the generation of massive data, and the processing of this vast amount of data relies on powerful computing resources and widely covered network connections.The convergence of computing and network has become a crucial trend.Against this backdrop, a novel network paradigm, termed the "Computing Network", has been proposed [1][2][3], which integrates computing resource scheduling capabilities into the network.This paradigm aims to interconnect distributed computing resource nodes; coordinate computing resources, network resources, and user requirements; and achieve the integrated scheduling of computing and network resources [4].
Various network service providers, represented by telecommunications operators, have proposed their own Computing Network architectures.Each provider has put forth their respective ideas regarding the collaboration of computing and network resources, and they have implemented computing resource scheduling in either a centralized or distributed manner within these architectures [5][6][7][8].
It is important to note that computing resource scheduling in Computing Network is a broad concept.It can involve the scheduling of fundamental resources such as CPU, memory, and storage, as well as the scheduling of specific computing services provided on computing resource nodes.However, this paper primarily focuses on the scheduling of basic resources under service deployment requests [9].Furthermore, computing resource scheduling, as one of the critical issues in computational systems such as Computing Network, Edge Computing, and Cloud Computing, plays a crucial role in determining the effectiveness of computing resource utilization and user service experience [10].
However, with the rapid increase in cloud-edge-end computing resource nodes, along with diverse computing demands from users, the centralized computing resource scheduling mechanism in a Computing Network faces scalability limitations.It is not suitable for large-scale network deployments [11][12][13].Distributed scheduling methods, on the other hand, offer advantages in terms of scalability and reliability [14].Therefore, this paper places emphasis on the study of distributed computing resource scheduling architectures.
Existing distributed resource scheduling architectures in a Computing Network mostly rely on a flat architecture, performing full-scale dissemination of resource status information and scheduling among all the scheduling nodes [12].Faced with a scenario of highfrequency changes in computing resource node status, it is challenging for resource status information to synchronize rapidly [15,16].Delayed resource status information is prone to scheduling failures.Moreover, current distributed resource scheduling architectures perform scheduling based on conventional TCP/IP networks, determining the final scheduling destination at the user entry node.This "in-out" scheduling method lacks flexibility [13] and cannot effectively address resource competition issues in distributed resource scheduling scenarios, leading to potential scheduling failures.In contrast, in Information-Centric Networking (ICN), the content-name-based routing mechanism, combined with Software-Defined Networking (SDN) technologies, allows for convenient readdressing of routing in the network layer [17][18][19], offering high routing flexibility.Therefore, ICN possesses advantageous conditions for implementing dynamic computing resource scheduling.Additionally, current resource scheduling architectures based on traditional TCP/IP networks often adopt single and static scheduling strategies, such as optimizing the utilization and load balancing of computing resources or minimizing resource usage costs [20].These architectures do not specifically design scheduling schemes based on user preferences, thus failing to meet diverse user requirements.ICN, with its novel routing protocols, which allow flexible definition and use of fields, can achieve flexible routing of interest packets based on user preferences [21,22], demonstrating the potential to implement a preference-based resource scheduling mechanism.
Therefore, this paper proposes an ICN-based on-path computing resource scheduling architecture with the following main contributions: • Designing a hierarchical scheduling architecture that supports on-path resource scheduling: by carrying resource demands and preference parameters in the computing resource request packets, the scheduling nodes can adjust their scheduling destinations promptly during the packet routing process based on demand parameters and current resource status information, thus improving scheduling success rates.

•
Developing computing resource scheduling strategies tailored to different resource request preferences: scheduling nodes read the preference field in the packet header, make decisions, and forward the packet based on the corresponding scheduling strategy, meeting diverse user requirements.
The remaining sections of this paper are organized as follows: Section 2 analyzes relevant research on the Computing Network, existing resource scheduling mechanisms, ICN, and constructing a Computing Network based on ICN.Section 3 describes the overall structure of the proposed distributed computing resource scheduling architecture IPCRSA, presenting the design of computing resource request packets and the typical process of resource scheduling within this architecture.Section 4 expounds on the new mechanisms and optimizations of IPCRSA, encompassing the preference-based computing resource scheduling strategies and the processing and forwarding mechanisms for computing resource request packets.In Section 5, a performance analysis of IPCRSA is conducted through a comparative simulation experiment.Finally, Section 6 summarizes the work presented in this paper.

Related Works
This section introduces the main related works pertinent to this paper.Section 2.1 begins by discussing the related works on the Computing Network from three perspectives: macro-policy, standardization, and academia; it then introduces the existing methods of computing resource scheduling within the Computing Network from two perspectives: distributed and centralized.Finally, it presents some computing resource scheduling mechanisms outside of the Computing Network that utilize multi-criteria decision-making approaches to aid in the allocation of computing resources.Section 2.2 reviews the related works on ICN, explicating the characteristics of the ICN architecture upon which this paper is based.It subsequently provides a brief introduction to the existing methods of building a Computing Network and implementing the scheduling of computing resources based on ICN.

Computing Network and Computing Resource Scheduling
In recent years, the academic and industrial sectors have conducted extensive research and exploration on the Computing Network.On a macro-policy level, Chinese telecommunications operators have issued a white paper on Computing Network research, elaborating on the demands, architecture, and key technologies of the Computing Network.Various Computing Network architectures, including Computing-aware Networking (CAN), Computing Power Network (CPN), Computing First Network (CFN), etc., have been proposed [23][24][25].In terms of standardization, ITU-T has initiated and released a series of standards related to the Computing Network, such as Y.CPN-arch, Y.IMT2020-CAN-req, Q.CPN, Q.BNG-INC, etc. [26][27][28][29].The IETF's Routing Area Working Group (RTGWG) working group has conducted research on CFN, submitting three core proposals that outline the framework, requirements, and use cases of CFN [1,7,30].Additionally, the IRTF has established the Computing in the Network Research Group (COINRG) to conduct relevant research on computing-network integration standards.On the academic front, the authors in [2] first proposed a Computing Network solution based on deep integration of cloud, network, and edge, providing an implementation case for AI applications.The authors in [3] analyzed the research background and key technologies of the Computing Network, demonstrating their advantages in 6G business scenarios.The authors in [11] provided a detailed exposition of the overall composition and routing protocols of the Computing Network architecture: CFN.
As the core of the Computing Network, computing resource scheduling in it has also received extensive attention.In terms of distributed scheduling, the authors in [8] proposed a computing-aware bandwidth allocation method called CABA based on utility optimization.CABA selects the optimal computing resource node and routing path for resource requests in the Computing Network, enhancing the overall utilization of computing resources.Although this method considered user satisfaction as one of the decisive factors in the scheduling of computing resources, it failed to account for the users' preferences for resource requirements.The authors in [12] introduced a CAN-based Computing Network simulation prototype, validating the effectiveness of its computing-aware routing scheduling mechanism and improving the overall utilization of computing resources.The authors in [13] proposed a Computing Network routing protocol, CARP.It comprehensively consid-ered computing and network resource information, scheduled computing resource requests to optimal service nodes, and optimized system capacity and end-to-end average latency.However, both methods above assumed that the resource status information they acquired was real-time, overlooking the resource competition issue in the scenario of distributed scheduling, which was challenging to achieve in real-world network environments.Moreover, the method of always choosing a single optimal destination resource node, in the absence of real-time information synchronization, could not achieve load balancing among nodes, potentially adversely affecting resource scheduling.Additionally, both methods above were based on the TCP/IP network for scheduling and determined the destination resource node at the ingress node, lacking the flexibility to adjust based on resource status information in a timely manner.The authors in [31] presented a Computing Network resource scheduling architecture centered around the IP network, implementing computing orchestration and scheduling at the network layer based on the Segment Routing IPv6 (SRv6).The authors in [32] built a CPN prototype experimental platform based on Kubernetes, achieving dynamic allocation of computing resources and flexible scheduling of resource requests, reducing response latency, and enhancing load balancing performance.Although these methods analyzed and optimized resource scheduling in the Computing Network from an implementation perspective, and the latter balanced the load of resource nodes, there remained an issue of lacking scheduling flexibility.In terms of centralized scheduling, the authors in [33] elaborated on the concept of Segment Identifier as a Service (SIDaaS) based on SRv6 and proposed the Programmable Service System (PSS) based on SIDaaS, providing a vital implementation method for integrated scheduling of computing and network resources.The authors in [34] proposed a lightweight, multi-cluster hierarchical edge resource scheduling solution based on a hybrid architecture of Kubernetes and K3s.This scheme achieved unified management and scheduling of a large number of edge computing resource nodes.The authors in [35] implemented coordinated scheduling of computing and network resources using SDN technologies, orchestrating computing and services on the underlay network.However, with the increase in the number of ubiquitous computing resource nodes and diversified user demands, the aforementioned centralized scheduling methods faced issues of scalability and reliability limitations, which similarly had an adverse impact on resource scheduling.
Furthermore, some computing resource scheduling mechanisms outside of Computing Network have utilized multi-criteria decision-making approaches to aid in the allocation and scheduling of computing resources, which is also of reference value to this paper.For instance, the authors in [36] employed the Analytic Hierarchy Process (AHP) within Kubernetes for resource scaling decisions, optimizing the allocation of computing resources.The authors in [37] used the TOPSIS method for placing Virtual Network Functions (VNFs) within computing resource clusters to complete resource invocation, significantly enhancing the satisfaction rate of requests.In [38], the authors introduced a two-phased Multi-Attribute Decision Making (MADM) method, utilizing the P2P formalism to discover and allocate computing resources.
Overall, due to scalability limitations associated with centralized computing resource scheduling, distributed approaches have gained more widespread application.However, the aforementioned distributed resource scheduling mechanisms still face challenges, such as a lack of flexibility and awareness of user preferences.Computing resource scheduling based on ICN has the potential to address and improve these issues.Consequently, this paper proposes a computing resource scheduling architecture based on ICN, which has advantages in terms of scheduling flexibility and meeting diverse user preferences.A detailed analysis will be further elaborated on in the following sections.
Existing ICN architectures can be categorized into two types based on their implementation approaches: evolutionary and revolutionary.Evolutionary ICN architectures aim to incrementally improve the current network architecture, enhancing network functionalities by introducing ICN concepts and mechanisms within the existing IP network framework.This is typically achieved by adding new protocol layers, utilizing middleware, or extending existing TCP/IP protocols.As a result, they can be relatively easily deployed in current network environments, reducing the need for extensive infrastructure replacement [46,47].On the other hand, revolutionary ICN architectures advocate for a fundamental redesign of the network, with content-centric concepts as the foundation of network construction.This approach often requires the development of entirely new network protocols and infrastructures, which may include new routing algorithms, data naming schemes, caching mechanisms, etc.Therefore, the implementation of revolutionary ICN architectures tends to be more complex, challenging, and costly.The computing resource scheduling architecture proposed in this paper is based on the evolutionary ICN architecture, which facilitates its subsequent implementation in the real world at a lower cost.
ICN adopts an information-centric communication approach, replacing the traditional host-centric model, and routes based on information or content names rather than host addresses.Different ICN architectures may employ various naming schemes for content, which are primarily categorized into hierarchical and flat naming schemes.For instance, NDN uses a human-readable hierarchical structure similar to the URL for naming content.PURSUIT employs a flat Rendezvous Identifier (RID) and a Scope identifier (SID) as a pair for content naming.Mobility First uses flat, globally unique identifiers (GUIDs) for this purpose.However, existing research demonstrates that flat naming schemes offer superior advantages in terms of security, scalability, and flexibility compared with hierarchical schemes [48].Consequently, the ICN architecture described in this paper also utilizes flat identifiers (IDs) for content naming, akin to the aforementioned flat naming schemes, which facilitates routing within the ICN environment based on these IDs.
Since ICN routes based on content names rather than host addresses, it necessitates a name resolution service that matches or translates content names to the locators of content providers.This enables content requests to route to the content provider's location based on content names or locators [49].The name resolution process and the content request routing can be coupled (performing name resolution during the routing process).In this scenario, content requests route directly by content names, and content providers respond along the reverse path of the content request route.This approach is known as the Name-Based Routing (NBR) method, with NDN being a typical example.Alternatively, name resolution can be an independent process, separate from message routing.Content requests first obtain locators through an independent name resolution system, then route based on those locators.This method is referred to as the Standalone Name Resolution (SNR) method, with typical examples including DONA, PURSUIT, MobilityFirst, SEANet, and others.However, NBR methods are often found in revolutionary ICN architectures, requiring substantial modifications to the existing network infrastructure and presenting high deployment challenges.On the other hand, SNR methods, based on a separate name resolution system, offer better scalability and compatibility with the existing TCP/IP network, aligning more closely with the needs of evolutionary ICN architectures.Therefore, the ICN architecture discussed in this paper employs the SNR method.Additionally, it is important to note that IP addresses are utilized as locators for content providers in the ICN environment, which is compatible with TCP/IP in this paper.
Regarding the integration of ICN and Computing Network, the authors in [4] proposed a scheme for constructing a named Computing Network based on ICN, emphasizing that the routing scheduling in both the Computing Network and ICN emphasize "content" rather than the specific location of hosts.However, this method merely presented the concept without delving into specific mechanism designs.The authors in [14] designed a distributed resource scheduling architecture for Computing Network based on ICN, utilizing Conflict-free Replicated Data Types (CRDTs) as the underlying data structure and Remote Method Invocation for ICN (RICE) as the execution environment, demonstrating advantages in simplicity, performance, and fault recovery capability.Yet, this approach, predicated on the revolutionary ICN architecture, poses significant deployment challenges.Additionally, some novel ICNs based on the SNR method, incorporating SDN principles and technologies, have implemented a new feature of readdressing or late-binding for IP in the network layer, further enhancing the flexibility of ICN data routing [45].Consequently, they possess advantageous conditions for realizing the dynamic scheduling of computing and network resources.
Finally, a review of the related works on computing resource scheduling in Computing Network mentioned above is conducted, with the most representative studies selected and summarized in Table 1 below for reference.

Architecture
This section will elaborate on the proposed on-path computing resource scheduling architecture based on ICN, which is named IPCRSA.Section 3.1 presents the overall design of the architecture, including the topology structure and relationships among various entities.Section 3.2 provides an explanation of the design of the computing resource request packet format within this architecture, while Section 3.3 introduces the typical process of computing resource scheduling in IPCRSA.

IPCRSA Overview
Figure 1 illustrates the overall design of IPCRSA.Its network topology is divided into three levels: L 3 , L 2 , and L 1 .The main entities include the Scheduling Node (SN); Gateway Node (GN); Computing Resource Node (CRN); and User Client (Client), User, and Resolution Node (RN).There is a one-to-one connection relationship between SNs and GNs, and GNs establish either a one-to-one or one-to-many connection relationship with CRNs.The L 3 , L 2 , and L 1 level network topologies and various entities within their coverage areas collectively constitute the scheduling domains in IPCRSA at levels L 3 , L 2 , and L 1 .The specific functions of each entity will be explained in the following sections.

Scheduling Node
The SNs in IPCRSA also serve as ICN nodes, implementing computing resource scheduling based on ICN.As depicted in Figure 2, the composition structure of SNs can be divided into the control plane and the data plane.The main function modules in the control plane include the ID Management Module, Topology Construction Module, Resource Status Notification Module, and Computing Resource Scheduling Module.The data plane is responsible for receiving, processing, and forwarding computing resource request packets.
The ID Management Module oversees the management of identifiers and network addresses mapping relationships of SNs, ensuring the addressability of SNs within the ICN.Leveraging the ID Management Module, the Topology Construction Module accomplishes the construction of logical topology for resource status notification and resource scheduling, ultimately forming the hierarchical logical topology structure of IPCRSA, as illustrated in Figure 1 (The logical topology describes the way data packets flow within a network, determined by network protocols and configurations).The Resource Status Notification Module, based on the hierarchical topology structure, synchronizes resource status informa-tion among SNs, forming a resource status information table in each SN.The Computing Resource Scheduling Module, relying on the resource status information table, executes computing resource scheduling strategies.In collaboration with the Computing Resource Scheduling Module, the data plane of each SN processes and forwards computing resource request packets, thereby conducting on-path computing resource scheduling.The specific functionalities of each module and their interrelationships will be further elucidated in the subsequent sections.

• ID Management Module
As shown in Table 2, in IPCRSA, SNs within the scheduling domains at the L x level possess identical and flat resource scheduling service identifiers, denoted as RSS_ID_L x , indicating that they can provide resource scheduling service in the corresponding scheduling domain at level L x .The mapping relationship between the SN's RSS_ID_L x and IP is maintained by a hierarchical ICN name resolution system similar to the one discussed in SEANet [50].The basic structure of the hierarchical name resolution system is illustrated in Figure 3.The resolution domains (the scope within which name resolution services can be provided) are divided into three levels: L 3 , L 2 , and L 1 , adhering to the principle of inter-layer nesting and intra-layer non-intersection.From level L 3 to L 1 , the coverage scope of the resolution domains gradually expands.RNs within each level's resolution domains are the main components of this hierarchical name resolution system.They are responsible for maintaining the mapping relationships between IDs and IP addresses of all entities (such as SNs, Users, etc.) within their resolution domains in the ICN environment.The RNs themselves follow a multi-layered tree structure, providing name resolution services within the scope of different levels' resolution domains.It is important to note that the resolution hierarchy and resolution domains of this hierarchical name resolution system correspond one-to-one with the scheduling hierarchy and scheduling domains in IPCRSA.Each scheduling domain in IPCRSA is also a resolution domain of the hierarchical name resolution system.The RN provides name resolution services within each scheduling domain, as illustrated in Figure 1.The ID Management Module is responsible for interacting with the RN in its scheduling domain, completing the registration, deregistration, and query of the mapping relationship between the SN's RSS_ID_L x and IP.

• Topology Construction Module
The Topology Construction Module utilizes the ID Management Module and ICN name resolution system to build a semi-distributed structured tree-like topology and structured connections among SNs within each scheduling domain at the L 3 and L 2 levels sequentially.The SN with the minimum average communication delay to other SNs within the domain is selected as the root node, and the root node of the lower-level scheduling domain also belongs to the scheduling domain at the upper level, with the corresponding resource scheduling service identifier (L 3 is the lowest level).Subsequently, within the scheduling domain at the L 1 level, the module establishes a fully distributed unstructured topology and unstructured connections among SNs.Ultimately, the hierarchical topology structure depicted in Figure 1 is constructed.
It should be additionally noted that the Topology Construction Module, through the ID Management Module, queries the address information of other SNs within its scheduling domain from the RN and measures the communication delay with other SNs.The construction of tree-like topologies and structured connections among SNs within the L 3 and L 2 level scheduling domains primarily aims to reduce the overhead associated with the flooding of subsequent computing resource status information among SNs, akin to the method discussed in reference [51].The establishment of unstructured topology and connections within the L 1 level scheduling domain primarily aims at load balancing among SNs.
The construction process of the hierarchical topology is as follows: 1.When a new SN comes online, its ID Management Module registers the mapping relationship between RSS_ID_L 3 and IP with the RN within its L 3 level scheduling domain.Ultimately, the construction of the hierarchical logical topology is completed.

• Resource Status Notification Module
The primary responsibilities of the Resource Status Notification Module are as follows: (1) Fetching communication delay information among SNs within the domain as measured by the Topology Construction Module.(2) Receiving and processing local CRNs' status information reported by connected GNs.(3) Receiving CRNs' status information from other SNs within its scheduling domain.(4) Notifying local CRNs' status information to other SNs within the domain.( 5) Aggregating CRNs' status information, performing the fetch of communication delay among SNs and resource status information notification in the corresponding scheduling domain at the upper level, in the condition that the associated SN of the module is the root node of the scheduling domain.( 6) Forming a resource status information table and storing it locally.
The GNs can perceive and monitor the status information of CRNs connected to them and report it to the SNs, as elaborated in Section 3.1.2below.CRNs can be categorized into various types based on their purposes.CRNs of the same type share the same computing resource service identifier, CRS_ID_T x , indicating that these CRNs can provide computing resource services of the corresponding type T x .A detailed introduction is provided in the subsequent Section 3.1.3.The notification of computing resource status information relies on the topology constructed by the Topology Construction Module and is disseminated using a flooding mechanism.The relevant fields of the resource status information table and their meanings are shown in Table 3 above.Each entry in the table records network and computing resource status information related to a CRN.

• Computing Resource Scheduling Module
The Computing Resource Scheduling Module serves as the execution unit for computing resource scheduling strategies.In IPCRSA, user resource requests are sent from the Clients to the SN's data plane in the form of computing resource request packets, and the Clients query the IP addresses of the SNs corresponding to the RSS_ID_L x through the ICN name resolution system, then send the packets to the designated SN.This process will be elaborated in Section 3.3 of the following text.The Computing Resource Scheduling Module, relying on the status information of various CRNs in the local resource status information table, executes the corresponding computing resource scheduling strategy.It precomputes and generates a computing routing flow table, which is then sent to the data plane based on the southbound interface protocol.This flow table guides the processing and forwarding of computing resource request packets.The flow table matching process of in-transit packets is akin to the execution of scheduling strategies.In other words, by on-path matching, processing, and forwarding the request packets, the module seeks CRNs that meet the User's demands, implementing computing resource scheduling.Furthermore, in the ICN environment where routing is based on IDs, the data plane of the SN can adjust the scheduling targets of the request packets based on the current resource status information during their routing process, thereby enhancing the flexibility of resource scheduling.The specific computing resource scheduling strategies and the data plane's processing and forwarding mechanisms for computing resource request packets during the flow table matching process will be elaborated in Sections 4.1 and 4.2, respectively.In IPCRSA, the GNs serve the functions of User and CRN access and management.On the CRN side, the GNs perceive and monitor the real-time resource status information of each CRN.They identify the types of CRNs, generate computing resource service identifiers, and finally report the identifiers and the real-time status information of corresponding CRNs to the connected SNs.On the User side, when computing resource scheduling is successful, the GNs return a scheduling success response packet to the Clients.Based on the information in the response packet, the Clients establish a connection with the GNs.Users operate the CRNs managed by these GNs to complete relevant service deployment operations.

Computing Resource Node
In IPCRSA, CRNs can be classified into three types based on their purposes: Universal-CRN, Super-CRN, and Intelligent-CRN, as shown in Table 4.These nodes can take the form of physical machines, virtual machines, or containers.Specifically, Universal-CRNs are primarily used for routine computing tasks in consumer and industry internet domains.Super-CRNs are mainly employed in the scientific field to handle large-scale scientific and engineering computing tasks.Intelligent-CRNs are mainly dedicated to processing tasks related to artificial intelligence.The same type of CRNs have the same computing resource service identifiers, CRS_ID_T x , indicating that the node can provide corresponding computing resource services with type T x , as shown in Table 2.

CRN Type
Type Name ID It is important to note that IPCRSA does not impose any restrictions on the classification methods for CRNs.The classification methods mentioned above are just examples, and IPCRSA supports other classification methods and custom computing resource service identifiers for various scenarios.

User and User Client
In IPCRSA, Users are the requesters of computing resources and may be individuals or organizations.Within IPCRSA, each User is assigned a unique identifier, USER_ID, which serves not only as an identity marker but also as its ID within the ICN environment, as depicted in Table 2 above.The mapping relationship between USER_ID and the User's IP address is also maintained by the aforementioned hierarchical name resolution system.
The Client is a software or hardware tool that enables Users to access IPCRSA to initiate resource requests or use resources.Its primary functions are as follows: (1) Identifying the User and generating a USER_ID.(2) Receiving input for resource requests from the User, creating corresponding computing resource request packets, and dispatching them to the SN.(3) Interacting with the hierarchical name resolution system to complete tasks such as registering the mapping relationship between USER_ID and User's IP, and querying for the IP of SN associated with RSS_ID_L x .(4) Receiving scheduling success or failure response packets and relaying the scheduling outcome back to the User.If the scheduling is successful, the User's service deployment commands can be forwarded from the Client to GN for execution.

Packet Format Design
Within an evolutionary ICN environment compatible with TCP/IP, IPCRSA designs a computing resource request packet format based on the IPv6 extension header, as illustrated in Figure 4. Firstly, two crucial fields, Source ID and Destination ID, are added to the packet header.The Source ID is filled with USER_ID to identify the User requesting resources, while the Destination ID is filled with the computing resource service identifier CRS_ID_T x to specify the type of computing resource service requested by the User.The inclusion of these two ID fields enables the implementation of an ID-based routing mechanism in the ICN.Combining this mechanism with SDN technologies allows the convenient implementation of late-binding for IP, enabling modification of the Destination IP field during the packet routing process based on actual requirements, which enhances routing and scheduling flexibility.
Following these two ID fields are the QoS field, Cost field, and Resource Request Quantity field.The QoS field is filled with the User's QoS constraint for resource scheduling delay: Scheduling_Delay.The Cost field is filled with the maximum cost that the User is willing to accept for CRN usage: Max_Price.The Resource Request Quantity field is filled with the specific resource request quantities in CPU, memory, storage, and bandwidth dimensions: RRQ_Core, RRQ_Memory, RRQ_Storage, and RRQ_Bandwidth, respectively.The values of these three fields, along with the Destination ID field's computing resource service identifier, collectively constitute the User's hard constraints on resources.
Next is the Preference field, which is filled with the User's specific resource request preference: RRP x , indicating the User's inclination toward resources.This field represents a soft constraint.
Finally, there are the Scheduling State field and the Scheduling Target field.The Scheduling State field, filled with the IS_Scheduled flag, is set to 0 when the computing resource request has not been scheduled, and the corresponding Scheduling Target field is left empty.When set to 1, it signifies that the computing resource request has been scheduled, and the corresponding Scheduling Target field is filled with the sequence number of the destination CRN: Sequence_Number.
Each SN in IPCRSA can receive computing resource request packets, read the relevant fields of resource demands and demand preferences in the packet header, make intelligent decisions, and forward packets based on the corresponding computing resource scheduling strategy to accomplish resource scheduling.

Scheduling Workflow
Taking the overall structure of IPCRSA shown in Figure 1, along with its various entities, as an example, this section describes the typical process of computing resource scheduling in IPCRSA, which is illustrated in Figure 5 below.1.The Client generates a computing resource request packet based on the resource demand parameters the User provides.2. The ICN name resolution system is used to query an SN within the L 3 level scheduling domain that has the identifier RSS_ID_L 3 .The computing resource request packet is then randomly sent to any of these SNs, following the principle of proximity scheduling.3. Through on-path packet matching, processing, and forwarding between SNs, a CRN that satisfies the User's demands and preferences can be identified.4. The GN returns a scheduling success response packet to the Client. 5.The Client establishes a connection with the GN. 6.The User, through the GN, completes the deployment operation of the services on the target CRN.

Mechanisms and Optimizations
The computing resource scheduling mechanisms are the pivotal component of the computing resource scheduling architectures.It determines which destination CRN is selected and governs how CRNs are chosen.It is the synergistic interaction of these two aspects that achieves computing resource scheduling.This section will focus on introducing the new mechanisms and optimizations of IPCRSA, commencing with these two aspects.Section 4.1 delves into the computing resource scheduling strategies based on user preferences.Section 4.2 elucidates the specific processing and forwarding mechanisms of computing resource request packets in the data plane of each SN under the on-path resource scheduling mechanism.

Preference-Based Scheduling Strategies
The computing resource scheduling in IPCRSA is preference-based, with different user preferences corresponding to distinct computing resource scheduling strategies.The execution of the preference-based computing resource scheduling strategies involves two stages: The first stage evaluates various CRNs using a node-scoring algorithm corresponding to the user preferences based on the status information of each CRN in the local resource status information table.The second stage involves selecting the destination CRN based on the scoring results.
In IPCRSA, Users with varying resource request preferences are most concerned about different status parameters of computing resource nodes, which are defined in this paper as priority indicators corresponding to each preference.Currently, IPCRSA has designed specific node-scoring algorithms for various resource request preferences and preference combinations, as outlined in Table 5, with the input parameters being the priority indicators corresponding to them.When a User has no specific preferences, a node-scoring algorithm corresponding to the default preference, i.e., the load-first preference, is employed to score each CRN.Specifically, the SN calculates the load score for each CRN using Equations ( 1)-(4).Here, u i c , u i m , u i s , and u i b represent the normalized utilization rates of CPU, memory, storage, and bandwidth, respectively, for the ith (i ∈ N) CRN.N represents the set of CRNs' ordinal numbers.These parameters correspond to CPU_Usage, Memory_Usage, Storage_Usage, and Bandwidth_Usage in the local resource status information table.ūi is the mean of the normalized utilization rates for these four basic type resources.MaxNodeScore is the predefined maximum score for CRNs, typically set to 10 or 100.s i (i ∈ N) represents the final score for the CRN.The higher the score, the lower the load on the CRN.ūi = (u i c ) × MaxNodeScore (2) When Users have preferences in terms of the resource itself, cost, energy consumption, and scheduling delay, i.e., having one or more preferences from RRP 1 to RRP 7 , the SN scores each CRN based on the TOPSIS comprehensive evaluation method.Firstly, an original data matrix X n×m is constructed based on n CRNs and m input parameters corresponding to user preferences in the local resource status information table.The row vector represents the value vector of m input parameters for the ith (i ∈ N) CRN in this scoring algorithm.The column vector represents the values of the n CRNs on the jth (j ∈ M) input parameter.M represents the set of input parameters' ordinal numbers.
Next, it is necessary to "Benefit-Cost" adjust the original data matrix.Among the input parameters corresponding to the preferences of RRP 1 -RRP 7 mentioned above, Avail-able_CPU_Cores, Available_Memory, Available_Storage, and Available_Bandwidth are beneficial attributes, where larger values are better.As input parameters for the scoring algorithm, their "Benefit-Cost" values are the same as their original values, i.e., x + ij = x ij .On the other hand, Price, PUE, and Communication_Delay are cost attributes, where smaller values are better.If they are used as input parameters, adjustment is required using Equations ( 6) and (7).After adjusting the original data, the "Benefit-Cost" adjusted data matrix X + n×m is obtained. x Then, it is required to utilize Equation ( 9) to standardize the X + n×m , eliminating the influence of different dimensions and obtaining the standardized data matrix Z n×m .
Finally, based on the standardized data matrix, Equations ( 11)-( 15) are employed to calculate the comprehensive score s i (i ∈ N) for each CRN.A higher score indicates a better overall performance of the CRN across the dimensions of input parameters corresponding to user preferences.max z j = max z 1j , z 2j , z 3j , . . .z nj (11) ω j represents the weight of each input parameter, determined through the entropy weight method.Firstly, utilize Equation ( 16) to standardize the original data matrix.min x j is the minimum values of x 1j , x 2j , x 3j , . . .x nj , which is opposite to max x j .Subsequently, for the jth (j ∈ M) input parameter, calculate its information entropy E j using Equations ( 17) and (18).Finally, obtain the weight ω j using ( 19), with 0 ≤ ω j ≤ 1, and the sum of m weights is equal to 1.
x ij − min x j max x j − min x j (16) ) After obtaining scores for each CRN, it is necessary to select the target CRN based on the score results.
Initially, CRNs that meet the hard constraints indicated by the following condition (20) are added to the candidate CRNs set C. Herein, A represents the set of all CRNs.
Subsequently, within the candidate CRNs set, node selection is performed based on the score results.If a greedy algorithm is employed, always choosing the highest-scored node may exacerbate resource competition issues in distributed scheduling scenarios, leading to load imbalance and scheduling failures.On the other hand, adopting a random selection algorithm fails to meet user preferences.Therefore, in IPCRSA, the SN employs a rouletteselection algorithm for CRN selection.In this algorithm, the SN randomly selects CRNs with scores from high to low in descending order of probability, simultaneously balancing the load of CRNs and meeting user preferences.
The roulette-selection algorithm is commonly used in genetic algorithms, where the probability of selecting a candidate node is directly proportional to its composite score.Firstly, the probability p i of selecting each of the n * candidate CRNs is calculated based on their scores.Subsequently, the cumulative probabilities q i for each candidate CRN are computed.Finally, a uniformly distributed random number r is generated within [0, 1], and the range in which r falls determines the ultimately selected CRN.The overall process is illustrated by Equations ( 21)-( 23).
Based on the aforementioned scoring and selection algorithms of CRNs, the preferencebased computing resource scheduling algorithm in IPCRSA is formulated in Algorithm 1.Additionally, it is important to note that the resource scheduling in IPCRSA is not limited solely to the user preferences and corresponding scheduling strategies shown in Table 5.It has the flexibility to incorporate additional user preference options and corresponding scheduling strategies based on new user requirements, demonstrating high scalability.
Finally, the key variables involved in this section are summarized in Table 6 below for reference.The standardized data matrix ω j The weight for the jth input parameter E j The information entropy for the jth input parameter C The set of the candidate CRNs A The set of all CRNs n * Total number of candidate CRNs p i The probability of selecting the ith candidate CRN q i The cumulative probabilities for the ith candidate CRN

Processing and Forwarding Mechanisms for Resource Request Packets
Upon receiving a computing resource request packet in the data plane, the SN first reads the IS_Scheduled, CRS_ID_T x , and RRP x parameters in the packet header to confirm the status of the request, the resource service type requested by the User, and the user preferences.
If the request has not been scheduled, the scheduling strategy corresponding to user preferences is executed, which aims to search for a CRN that satisfies the User's resource demands in all dimensions.

•
If a non-local CRN is found, the Destination IP field of the packet is modified to the destination CRN's Source_IP, the IS_Scheduled is set to 1, the Scheduling Target field is modified to the destination CRN's Sequence_Number, and the packet is finally forwarded.

•
If a local CRN is found, the packet's destination IP is modified to the GN's IP corresponding to the SN and the packet is forwarded.The GN then returns a scheduling success response packet to the Client.

•
If no suitable CRN is found, and the SN has a superior node, the destination IP is modified to the IP of the root node in the scheduling domain, allowing exploration of potential CRNs in other scheduling domains.If there is no superior node, a scheduling failure response packet is directly returned to the Client.
When the computing resource request has been scheduled, the SN first checks whether the available resource quantity of the destination CRN can meet the User's resource demands.

•
If it can, the packet is forwarded to the GN corresponding to the SN and a scheduling success response packet is returned.

•
If it cannot, the SN performs a rescheduling based on the User's resource demands and preference parameters, modifying relevant fields in the packet accordingly.
It is noteworthy that, as a distributed computing resource scheduling architecture, IPCRSA also adheres to the CAP theorem for distributed systems [52].To enhance availability, IPCRSA, like many other distributed resource scheduling systems, follows the BASE principle, which guarantees only partition tolerance while striking a balance between availability and consistency.IPCRSA ensures only the eventual consistency of resource status information across SNs within the same scheduling domain, sacrificing strong consistency to improve the availability of resource scheduling.Consequently, it must grapple with the issue of resource contention resulting from the asynchronous nature of resource status information in the soft state.
For instance, suppose a CRN has 16 CPU cores, 64 GB of memory, 1024 GB of storage, and 10 MB of available bandwidth.Three computing resource requests, A, B, and C, sequentially request resources of this node within the same resource status notification interval.The scheduling results based on the quota allocation of resources are illustrated in Figure 6.After allocating a fixed amount of resources to requests A and B, the remaining available resources are four CPU cores, 8 GB of memory, 256 GB of storage, and 1 MB of available bandwidth, which cannot satisfy the resource demand of request C. In traditional Computing Network resource scheduling architectures, request C would remain pending, resulting in scheduling failure.However, in IPCRSA, the hierarchical scheduling structure accelerates the synchronization of resource status information, and the roulette-selection-based CRN selection mechanism ensures load balancing among CRNs.Both mechanisms alleviate resource contention issues in the soft state to some extent, improving the scheduling success rate.Furthermore, each SN in IPCRSA maintains a computing routing flow table.When the local resource status information table is updated, the computing routing flow table is updated accordingly.Computing resource request packets in the IPCRSA route based on IDs, enabling dynamic late-binding of IP addresses.In the aforementioned scheduling scenario, the resource request packet corresponding to request C can be promptly matched with the updated flow table based on changes in resource status information.This enables timely rescheduling, forwarding the packet to other candidate CRNs for resource invocation, and enhancing the overall scheduling success rate.

Simulation
This section centers on the performance analysis of IPCRSA.Section 5.1 provides a brief overview of the simulation setup and relevant assumptions.Sections 5.2 and 5.3 validate and analyze the scheduling performance of IPCRSA under the combined influence of the on-path scheduling mechanism and the preference-based scheduling strategies from the two metrics of the success rate of scheduling and the average value of priority indicators corresponding to preferences, respectively.Section 5.4 provides a comparative discussion and summary of the method described in this paper and other methods in the simulation experiment.

Simulation Setup and Assumptions
NetworkX is a Python library designed for the creation, analysis, and visualization of complex networks [53].It offers Users a concise API that is easy to learn and use.The simulation experiment utilized NetworkX to create a hierarchical topology, as depicted in Figure 7, to simulate the topology structure of IPCRSA.In the diagram, blue and red nodes represent root SNs in the L 3 and L 2 level scheduling domains, respectively, while black nodes are leaf SNs within the L 3 level scheduling domains.Overall, the simulation topology includes nine L 3 level domains, three L 2 level domains, and one L 1 level domain.The intradomain link delays between SNs in L 3 , L 2 , and L 1 level domains are set to random values within the ranges [1 ms, 5 ms], [5 ms, 15 ms], and [15 ms, 30 ms], respectively.Considering the nanosecond-level high-speed matching, processing, and forwarding capabilities of the SN's data plane for packets [54], the processing delay of packets at SNs can be neglected.Additionally, in this simulation experiment, it is assumed that each SN has a stable and minimal communication delay with its corresponding GN, which is negligible compared to the link delays between SNs.Therefore, the link delays between SNs become the primary factor that affects resource scheduling delays.It is important to note that, to focus on the computing resource scheduling process itself and make simulated topology more straightforward and understandable, GNs and CRNs are not labeled in the simulated topology diagram in Figure 7.However, they still exist and maintain the connectivity relationships with the SNs, as described in Section 3.1 in this simulation experiment.Furthermore, given that the simulated topology is situated within an evolutionary ICN environment that routes based on IDs, this simulation experiment introduces a hashing algorithm to generate flat IDs for Users, SNs, and CRNs within the simulation environment.When implementing the resource scheduling function on each SN, a separate subprocess is included to facilitate the late-binding for IP addresses, allowing for the late-binding of scheduling targets in subsequent experiments.Additionally, the simulation experiment also incorporates a simulated hierarchical name resolution system designed to maintain the mapping relationship between IDs and IP addresses within the simulation environment, thereby aiding in the scheduling of computing resources in this simulation experiment.
In this simulation experiment, computing resource requests and the status information of CRNs are sourced from an open dataset provided by Alibaba Cloud [55].In terms of requests, the dataset only includes the quantity of resources requested by Users in dimensions of CPU, memory, and storage.Therefore, it can be assumed that Users have no strict constraints on bandwidth, price, scheduling delay, or computing resource service types in this simulation experiment.Regarding the status information of CRNs, the dataset offers the total amount of resources and corresponding utilization rates in terms of CPU, memory, and storage for each CRN.For subsequent performance validation, the PUE values and usage prices for each CRN are set to random values within the ranges (1,4] and (0, 20], respectively.It is also assumed that all CRNs can provide the same computing resource service types and share identical computing resource service identifiers.
Under the aforementioned assumptions, based on the request information provided by the dataset, 5 w/10 w/15 w/20 w computing resource request packets are formed.These resource request packets are randomly dispatched to various SNs within the topology illustrated in Figure 7.The intervals between the arrival of each request at the SN's data plane follow an exponential distribution with a scale parameter set to 40 w.The dataset's CRNs are randomly linked to the GNs corresponding to the SNs in the simulated topology, and they report resource status information to the SNs through the GNs.The SNs then exchange resource status information to form a resource status information table.For cross-level aggregation, the TOPSIS and entropy weight method are employed to select CRNs with relatively higher values in all resource dimensions, representing lower-level scheduling domains.Each SN, based on the demand information from the computing resource request packets and the state information from the resource status information table, executes the corresponding computing resource scheduling strategy to accomplish the scheduling.Detailed experimental parameter configurations are provided in Table 7 for reference.The main metrics emphasized in this simulation experiment include the success rate of scheduling and the average value of priority indicators corresponding to user preferences.The former refers to the proportion of successfully scheduled request packets in all request packets.In the context of distributed resource scheduling, a higher success rate of scheduling indicates a stronger capability of the scheduling architecture to address resource contention issues when the total quantity of request packets is held constant.The latter denotes the average value of the preference-associated priority indicators on the CRNs occupied by successfully scheduled request packets.When the number of successfully scheduled request packets is the same, the magnitude of the average value of priority indicators reflects the capability of the computing resource scheduling strategies to meet diverse user demands and preferences.

Success Rate of Scheduling
In the context of Computing Network, reference [56] compares three scheduling schemes: CFN-dyncast, Native DNS scheduling, and Compute-Aware DNS scheduling.CFN-dyncast directs requests to a single "optimal" CRN based on the status information of both computing and network resources.The Native DNS scheduling scheme guides requests to various CRNs according to statically configured weights.Meanwhile, the Compute-Aware DNS scheduling scheme dynamically adjusts weights based on the status information of resources, facilitating dynamic load balancing.It is noteworthy that these schemes operate within a flat architecture and rely on the traditional TCP/IP network for scheduling.
This simulation experiment compares the three aforementioned schemes with the IPCRSA.To ensure comparability, the same simulation topology, as shown in Figure 7, is employed for all three schemes, treating the hierarchical topology as a flat topology for scheduling.In terms of specific scheduling strategies, IPCRSA utilizes the computing resource scheduling strategy corresponding to the default preference.This involves scoring each CRN based on its load status and subsequently selecting the CRN using the rouletteselection algorithm.Under the same scoring method, the CFN-dyncast scheme directly chooses the CRN with the highest score.The Compute-Aware DNS scheduling scheme adopts the same scheduling strategy as IPCRSA, achieving dynamic weight adjustments.The Native DNS scheduling scheme defaults to equal static weights, i.e., round-robin scheduling.However, in contrast to the scheduling of service requests in [56], the candidate set of CRNs for basic resource requests in service deployment scenarios undergoes high dynamic changes, making it challenging to perform effective scheduling based on static weights.Additionally, considering the similarity in the performance of random scheduling and round-robin scheduling when the number of requests is large, this experiment replaces the static-weight setting method with a random selection method for scheduling.Finally, the performance of the four schemes is compared under the Scheduling_Delay of 10 ms/50 ms/100 ms/150 ms.
Figure 8 present the scheduling performance comparison of the four schemes under different scheduling delay constraints.It can be observed from the figures that, regardless of the scheduling delay constraint, IPCRSA consistently achieves an overall higher scheduling success rate compared with the other three schemes.The advantage of IPCRSA becomes more pronounced when the number of computing resource requests is relatively small.Particularly, when the total number of requests is 5 w, IPCRSA demonstrates a minimum improvement of 24% and a maximum improvement of 63% in scheduling success rate compared with those of the other three schemes.As the number of requests increases, resource competition intensifies, and constrained by the total available resources of CRNs, the scheduling success rates of all four schemes decrease.However, IPCRSA maintains its advantageous position.These conclusions thoroughly demonstrate the superiority of IPCRSA in improving scheduling success rates, primarily attributed to its on-path computing resource scheduling mechanism.This mechanism enables computing resource request packets to dynamically adjust their scheduling destinations during the routing process based on changes in resource status information, enhancing the flexibility of routing and scheduling.Additionally, this experiment analyzes the impact of different scheduling delay constraints on the scheduling success rate of IPCRSA.The results from Figure 9 indicate that as the constraints on scheduling delay gradually loosen, the scheduling success rate of IPCRSA increases accordingly.The reason behind this observation is that relaxed scheduling delay constraints provide computing resource request packets with more opportunities for matching, processing, forwarding, and rescheduling, thereby enhancing the scheduling success rate.

Average Value of Priority Indicators Corresponding to Preferences
To validate the scheduling performance of the preference-based computing resource scheduling strategies in IPCRSA, this simulation experiment takes the examples of two strategies corresponding to two user preferences: Efficiency-First and Cost-First (referred to as the Efficiency-First strategy and Cost-First strategy, respectively).These strategies are implemented in the simulated topology shown in Figure 7 and compared against the random-scheduling strategy.The total number of computing resource request packets is set to 10 w, with a Scheduling_Delay of 100 ms.The vital comparative metrics are the average values of the priority indicators corresponding to the above two user preferences, i.e., the average PUE and the average usage price.
Figure 10a presents the average PUE of CRNs occupied by requests under the Efficiency-First strategy and random-scheduling strategy, with successfully scheduled packet quantities of 1 w/3 w/5 w.The figure shows that the Efficiency-First strategy achieves an overall lower average PUE compared with the random-scheduling strategy.Particularly, when the number of successfully scheduled requests reaches 1 w, the average PUE of occupied CRNs is reduced by up to 31%.Similarly, Figure 10b provides a comparison of the Cost-First strategy and random-scheduling strategy.It shows that the Cost-First strategy obtains a lower overall CRN usage price compared with the random-scheduling strategy.Especially with 1 w successfully scheduled packets, the average usage price is reduced by approximately 28%.Regarding scheduling delay, Figure 11 reveals that although the average scheduling delay for both Efficiency-First and Cost-First strategies is slightly higher than that of the random-scheduling strategy, the difference does not exceed 1 ms.This difference is less than 1% of the scheduling delay constraint, which is significantly higher than the average scheduling delay.Given the obtained benefits, such overhead is evidently acceptable.The conclusions above strongly demonstrate the advantages of the preference-based computing resource scheduling strategies in IPCRSA.It simultaneously takes into account Users' preferences for resource demands and the load balancing of resource scheduling, meeting diverse user requirements with low scheduling delay overhead.

Discussion
Regarding the success rate of scheduling, the summary of the comparative schemes in the aforementioned simulation experiment is presented in the following Table 8.The scheduling scheme based on CFN-dyncast always selects the optimal CRN, which does not allow for load balancing among CRNs and may even exacerbate the problem of resource competition in distributed resource scheduling scenarios.The scheme based on Native DNS randomly selects the CRN, which, to some extent, balances the load and alleviates the resource competition issue.However, due to the limited flexibility of resource scheduling in the TCP/IP network, the resource competition problem cannot be further improved.The scheduling scheme based on Compute-Aware DNS makes a probabilistic selection of the CRN using the roulette-selection algorithm, which also allows for load balancing.However, it also faces the limitation of resource scheduling flexibility in the TCP/IP environment.The IPCRSA scheme proposed in this paper also uses the roulette-selection algorithm for the probabilistic selection of resource nodes, but its on-path computing resource scheduling mechanism in the ICN environment allows the computing resource request packets to adjust the scheduling targets in a timely manner based on the current resource status information during the routing process, thereby greatly enhancing the flexibility of routing and scheduling.Therefore, it can further alleviate the resource competition problem while achieving load balancing, thus having an advantage in the success rate of scheduling compared with that of the previous three schemes.
In terms of the average value of priority indicators, while the scheme of randomly selecting CRNs can effectively balance the load, it fails to accommodate user preferences.On the other hand, the IPCRSA's computing resource scheduling mechanism, based on user preferences, can balance the load among CRNs with minimal and negligible scheduling latency overhead while also taking into account user preferences, thus meeting diverse user requirements.
Finally, a retrospective analysis of the simulation experiment is conducted.Although it successfully validated and analyzed the scheduling performance of the architecture discussed in this paper, demonstrating its superiority, there are still areas for improvement.Firstly, due to the scarcity of accessible open-source datasets of real-world computing resource clusters, the simulation experiment was based on a single dataset, which may lack generality due to its limited information.In future work, more comprehensive datasets could be employed to further validate the performance of the proposed architecture.Secondly, some modules within the current simulation environment, such as the hierarchical ICN name resolution system, were developed in previous works.While these modules fulfill basic name resolution functions, they may not match the performance of the latest standalone name resolution systems.Future efforts could focus on upgrading and improving these modules to align with the latest research findings.Lastly, the computing resource scheduling architecture discussed in this paper is based on an evolutionary ICN architecture, with the core aim of deploying the proposed scheduling methods in the real world at a lower cost.Future work could involve validating these methods in the real world to achieve a more "realistic" "simulation environment".

Conclusions
This paper proposes a novel ICN-based on-path computing resource scheduling architecture, IPCRSA, in the context of the Computing Network, primarily addressing the fundamental resource scheduling issues under service deployment requests.IPCRSA adopts a hierarchical structure.By embedding resource requirements and preference parameters in the computing resource request packets, the SNs can dynamically adjust their scheduling destinations based on demand parameters and current resource status information during the packet routing process.This effectively alleviates resource contention issues in distributed scheduling scenarios, enhancing the overall scheduling success rate.Additionally, IPCRSA introduces preference-based computing resource scheduling strategies, where SNs execute corresponding resource scheduling strategies based on the Preference field in the packet header, catering to diverse user requirements.Finally, this paper presents a simulation implementation and performance analysis of IPCRSA.Experimental results demonstrate that IPCRSA has advantages in improving scheduling success rates.Moreover, as the scheduling delay constraints become loose, it exhibits even better scheduling performance.The preference-based computing resource scheduling strategies in IPCRSA achieve a balance between user preferences and resource scheduling load balancing, meeting diverse user requirements with minor overhead.
Although IPCRSA shows significant advantages in scheduling performance, there are still some challenges that need to be addressed in future research.For instance, as a hierarchical scheduling architecture, if the root SN within an L 3 or L 2 level scheduling domain experiences faults or becomes disconnected, rapid fault detection and handling are critical issues.Typically, a viable solution involves monitoring faults and implementing corrective measures by periodically sending heartbeat signals between root and leaf SNs.Additionally, efficient aggregation of resource status information at the root SNs within L 3 and L 2 level scheduling domains is another pressing concern that requires attention.Furthermore, IPCRSA is based on an evolutionary ICN architecture that employs the SNR method.Given the significant differences between the SNR method and the NBR method, IPCRSA cannot be directly applied to ICN architectures that utilize the NBR method.However, the concept of integrating computing resource scheduling capabilities into the network itself, as well as the idea of conducting computing resource scheduling based on ICN, are equally applicable to ICN architectures based on the NBR method.Therefore, how to modify the IPCRSA under an NBR-based ICN architecture to enhance its broad applicability is also one of the issues worth exploring in the future.

Figure 2 .
Figure 2. Architecture and function modules of SN.

Figure 4 .
Figure 4. Format of the computing resource request packet.

Figure 5 .
Figure 5. Workflow of computing resource scheduling in IPCRSA.

Figure 6 .
Figure 6.An example of computing resource scheduling in IPCRSA.

Figure 8 .
Figure 8. Success rate of scheduling with different Scheduling_Delay.

Figure 9 .
Figure 9. Success rate of scheduling in IPCRSA with different Scheduling_Delay.

Figure 10 .
Figure 10.Average value of priority indicators in IPCRSA with different scheduling strategies.

Figure 11 .
Figure 11.Average scheduling delay in IPCRSA with different scheduling strategies.

Table 1 .
Related works on computing resource scheduling.
2. Within each scheduling domain at the L 3 level, the Topology Construction Module of each SN queries the IP address information of other SNs within its scheduling domain from the RN via the ID Management Module and measures the communication delay with other SNs. 3. The SN with the smallest average communication delay to other SNs within the domain is selected as the root SN for the corresponding L 3 level scheduling domain, with other SNs serving as leaf nodes.4. Adjacency relationships are established between the root SN and all leaf SNs to form a tree-structured logical topology.5.The ID Management Module of the root SN of each scheduling domain at the L 3 level registers the mapping relationship between RSS_ID_L 2 and IP with the RN within its L 2 level scheduling domain.6.Similarly, within each scheduling domain at the L 2 level, the Topology Construction Module of each SN sequentially completes the tasks of measuring the communication delay between SNs, selecting the root SN, and constructing the tree-structured logical topology.7. The ID Management Module of the root SN of each scheduling domain at the L 2 level registers the mapping relationship between RSS_ID_L 1 and IP with the RN within its L 1 level scheduling domain.8. Within the scheduling domain at the L 1 level, each SN establishes adjacency relationships with each other, forming a fully distributed, unstructured topology.

Table 3 .
Fields of the resource status information table and their meanings.

Table 5 .
Resource request preferences in IPCRSA.
Output: CRN that meets the User's resource requirements 1: if RRP x = RRP 1 then x ; CRNs with Status Information Table;

Table 8 .
Comparative schemes in the simulation experiment.