Next Article in Journal
ZT-IoTrust: A Quantum-Resistant Zero Trust Framework for Secure IoT Access Control
Previous Article in Journal
Balancing Benefits and Risks of Embedding System Knowledge in Linear Active Disturbance Rejection Control
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust and Adaptive Service Recommendation Framework for Distinct Behavioral Services in Dynamic Multi-Cloud Environments

School of Computer Science and technology, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(22), 4468; https://doi.org/10.3390/electronics14224468
Submission received: 20 October 2025 / Revised: 13 November 2025 / Accepted: 14 November 2025 / Published: 16 November 2025

Abstract

Effective service recommendation is essential in multi-cloud environments, directly influencing system performance, resource utilization, and service composition quality. However, existing methods often overlook key issues: (i) the inability to identify distinct behavioral services (DBS), leading to inaccurate or inefficient recommendations; (ii) the neglect of temporal dynamics in service behaviors, which results in unstable or outdated decisions; and (iii) the lack of effective strategies for handling cold-start and data sparsity, causing incomplete or biased scoring. To address these challenges, this study proposes a behavior-aware recommendation framework that explicitly integrates DBS characteristics. First, three types of DBS are defined and identified through a two-phase topology-based detection algorithm. Second, a sliding window with role-stability judgment is introduced to capture behavioral evolution and ensure consistent role labeling. Third, a Markov-based scoring model is designed to propagate compatibility scores across the invocation topology, enabling infrequently invoked services to obtain nonzero scores during early iterations. Extensive simulations demonstrate notable gains in accuracy and robustness, showing a 22.85% improvement in precision and a 41.69% improvement in F1-score, effectively mitigating cold-start and sparsity challenges.

1. Introduction

Cloud services offer scalable, flexible, and on-demand resources that enable enterprises and individuals to deploy applications, store data, and apply artificial intelligence without dedicated infrastructure. As cloud computing evolves, organizations increasingly adopt multi-cloud environments, utilizing services from multiple providers to enhance reliability and performance. The diversity of service providers and offerings introduces significant complexity in service discovery and selection [1]. To address this, cloud service recommendation has become a vital mechanism for assisting users in identifying appropriate services efficiently and accurately.
Existing cloud service recommendation approaches [2] can be broadly categorized into the following three paradigms: (i) Collaborative Filtering (CF) methods [3] analyze user-service interaction patterns to infer preference but suffer from cold-start and data sparsity problems; (ii) Context-aware methods [4] incorporate external information such as time, location, or device context to enhance personalization, yet often ignore structural characteristics of services [5]; and (iii) QoS-aware methods [6] focus on predicting non-functional service attributes such as response time and availability but typically treat services as static entities without modeling their invocation behavior [7]. While these approaches significantly improve recommendation accuracy from different angles, they generally neglect the behavioral characteristics of services, which can directly affect both the relevance and efficiency of recommendations.
Motivations. Despite the prevalence of service recommendation systems, these methodologies often overlook an important aspect: the behavioral patterns of services within invocation workflows, which significantly impact their practical effectiveness, especially in complex, dynamic environments such as multi-cloud service ecosystems [8]. In such environments, services are often distributed across different cloud providers, leading to heterogeneous deployment standards, isolated logging mechanisms, and fragmented service invocation paths. These characteristics make it more likely for services to exhibit behaviorally irregular patterns—such as never being invoked together with others, only appearing at the end of workflows, or lacking historical records altogether. We define such anomalies as DBS. Unlike traditional single-cloud or centralized systems, multi-cloud settings amplify the occurrence and impact of DBS, making it imperative to design recommendation mechanisms that are explicitly aware of such behavioral irregularities.
Several key reasons contribute to this oversight: (i) Ignoring service behavior patterns reduces recommendation efficiency. Some services, such as Isolated Services, rarely interact and may introduce noise, while Ending Services frequently appear as workflow endpoints and should be prioritized. New Member Services, despite lacking historical records, may be crucial in specific contexts. Failing to differentiate these Distinct Behavioral Services (DBS) results in inefficient recommendations and redundant computations. (ii) Ignoring the dynamic nature of service behavior leads to misclassification and reduced recommendation accuracy. Services transition [9] between behavioral states due to workload shifts [10], cloud migration [11], or user demand changes [12]. Static models  [13] fail to capture these shifts, making it difficult to distinguish evolving patterns. For example, an Isolated Service may become highly interactive, or an Ending Service may transition to an intermediate role in cross-domain workflows. (iii) Neglecting DBS cold start and data sparsity issues hinders their effective utilization in recommendation. Many DBSs have low or no historical invocation frequency, making conventional approaches ineffective. However, some DBSs, despite their sparse historical records, should still be prioritized over structurally similar services due to their contextual significance in workflows. In a word, DBSs exhibit fragmented, unstable, or irregular invocation behaviors in multi-cloud service environments, making them difficult to model using conventional scoring techniques. DBSs often suffer from inaccurate or missing recommendation scores due to their structural uniqueness which not only reduces the overall recommendation quality, but also aggravates common issues such as cold-start and data sparsity. To address this, we propose a DBS identification and scoring framework that explicitly models such behavioral structures, offering both improved recommendation fairness and enhanced robustness against sparsity-related challenges. Our approach integrates dynamic monitoring mechanisms to track behavioral shifts over time, ensuring DBS classification remains adaptive to multi-cloud environments.
Challenges. The proposed framework addresses the following challenges (i): Differentiating DBS with similar static invocation patterns. DBS categories can have similar invocation records, making them hard to distinguish. Ending and Isolated Services may both exhibit low frequency, yet serve different roles. New Member Services may resemble inactive services. Effective identification must go beyond static similarity and incorporate topological and temporal features. (ii): Ensuring dynamic adaptability of DBS classification. Service behaviors change over time, requiring adaptive classification. A service may transition from Isolated to Interactive or Ending to Mid-Workflow based on evolving workflows. The challenge is designing an identification mechanism that updates DBS classification in response to real-time behavioral shifts. (iii): Addressing cold start and data sparsity issues. DBS often have limited historical data, making classification difficult. Traditional collaborative filtering and graph-based methods rely on sufficient invocation records, rendering them ineffective for DBS. A robust method should leverage contextual and topological intelligence to infer relevance even in sparse environments.
Contributions. To address these challenges, we make the following contributions:
  • A Novel DBS Identification Framework. We define Distinct Behavioral Services (DBS) and introduce an invocation topology-based approach to classify services into Isolated, Ending, and New Member Services. Our method captures both topological and temporal features to improve classification accuracy beyond conventional invocation-based methods.
  • A Dynamic Monitoring and Adaptative Detection of Behavioral Role Shifts based on Sliding Windows. We implement a sliding window mechanism to track behavioral shifts, ensuring DBS classification adapts to evolving interactions. A Markov-based transition model predicts state changes in DBS, preventing misclassification due to short-term fluctuations.
  • A Markov-Based Scoring Strategy against Cold-Start and Sparse Issues. We design a topology-aware scoring model that leverages the homogeneous and irreducible properties of Markov Chains. This enables complete and stable score propagation for DBS services even when invocation data is sparse or unavailable, effectively addressing cold-start limitations.
  • Comprehensive Validation Supporting the Proposed Framework. To validate the effectiveness of our framework, we design five dedicated simulations that assess its accuracy, stability, and robustness under cold-start and sparsity scenarios. These evaluations demonstrate the practical viability of our behavior-aware strategies in dynamic service environments.
The remainder of this paper is structured as follows. Section 2 reviews the literature on collaborative filtering, context-aware, QoS-aware, behavior-aware service recommendation together with cold-start and data sparsity issues. Section 3 presents the overall architecture of the proposed DBS-aware framework. Section 4 details the system design, including DBS definitions, invocation topology construction, the DBS-based filtering algorithm, and the sliding time window mechanism. Section 5 describes the simulation settings, dataset preprocessing, evaluation metrics, and baseline methods. Section 6 reports and analyzes the experimental results across dynamic environments, temporal stability, data sparsity, and cold-start scenarios. Section 7 concludes the paper and discusses limitations and future research directions.

2. Literature Review

2.1. Collaborative Filtering-Based Service Recommendation

Collaborative Filtering (CF) relies on user-service interaction patterns to infer preferences. Traditional CF methods, including user-based and item-based techniques, estimate similarity by comparing co-usage behaviors. To overcome data sparsity and cold-start problems, matrix factorization and deep learning variants have been proposed to enhance representation learning. For example, Pan et al. [14] proposed a hybrid recommendation system for university libraries that combines user-based CF, content-based filtering, and K-means clustering to alleviate data sparsity. Zhao et al. [15] further enhanced CF by integrating knowledge graph (KG) embeddings to capture richer semantic relationships between users and items. However, these approaches treat services uniformly and overlook their behavioral roles within service invocation networks, which limits their ability to filter or prioritize structurally significant services.

2.2. Context-Aware Service Recommendation

Context-aware recommendation methods enhance accuracy by integrating factors such as user location, time, and environment. These approaches often use embedding models or deep collaborative filtering to incorporate contextual information into the recommendation process. For instance, Wang et al. [16] introduced DFORM, a deep learning-based model that fuses contextual signals with collaborative features for IoT service recommendation. Fan et al. [4] proposed CASR-TSE, which models spatial relevance and temporal decay to improve QoS prediction in web service recommendation. However, these methods mainly emphasize external context and still treat services as static entities, failing to capture or utilize their internal behavioral dynamics and structural roles in workflows.

2.3. QoS-Aware Service Recommendation

QoS-aware approaches aim to improve service selection by predicting non-functional attributes such as latency, throughput, and reliability. These methods often employ machine learning or probabilistic models to estimate QoS values under varying user conditions. Wang et al. [17] proposed WAMTL, a multi-task learning framework that jointly predicts multiple QoS attributes by leveraging their interdependencies. Zhang et al. [18] combined trust propagation and QoS modeling in SIoT environments to improve reliability and satisfaction. Despite these innovations, existing QoS-aware models generally ignore service behavior patterns and offer no pre-filtering mechanism based on structural or temporal analysis. While QoS modeling is critical in traditional service composition or SLA optimization, it is less effective when service invocation behaviors exhibit instability, fragmentation, or sparsity—characteristics often observed in multi-cloud environments. Moreover, most QoS-aware approaches lack structural analysis capabilities and do not consider behavioral roles of services in invocation workflows, which limits their applicability in scenarios involving DBSs.

2.4. Behavior-Aware Service Recommendation

Behavior-aware service recommendation refers to approaches that explicitly model and leverage the behavioral patterns of services or users (such as interaction sequences, role changes, or multi-type behaviors) to improve recommendation accuracy and relevance. For instance, Kim et al. [19] proposed a graph convolutional network (GCN) framework for multi-behavior sequence-aware recommendation, capturing dependencies among different user behaviors such as viewing, adding to cart, and purchasing. While effective in modeling behavioral correlations, this approach relies on black-box embeddings and does not provide mathematically defined behavior roles, leaving the interpretation of behavior categories implicit. Similarly, Su et al. [20] introduced a personalized behavior-aware Transformer (PBAT) that employs attention mechanisms to distinguish between multiple behavior types and personalize recommendations. Despite its advanced modeling capabilities, PBAT also remains largely opaque and does not integrate service invocation topology or role-based classification into its framework. However, most existing behavior-aware methods still rely on black-box feature learning and lack mathematically defined service roles or explicit integration with service invocation topology.

2.5. Cold Start in Service Recommendation

2.5.1. Root Causes

Cold start arises when newly deployed or rarely invoked services lack behavioral traces in the invocation graph. Without such traces, collaborative or graph-based models cannot estimate their relevance with confidence. In multi-cloud settings, fragmented logs and domain boundaries intensify the problem, so new services may appear as isolated nodes with no exploitable connectivity.

2.5.2. Mainstream Solutions

Content-based methods [21] infer initial relevance from textual descriptions, metadata, or side QoS; they can operate without history, but quality hinges on the richness of side information and its cross-domain consistency. Initialization-based approaches [22] assign non-zero priors or meta-learned warm starts; this ensures early visibility in ranking, yet the priors may be arbitrary and unstable under domain shift. Context-aware models [23] leverage exogenous signals (e.g., time, location, social ties) to compensate for missing interactions; these signals are often noisy or misaligned across providers. Cross-domain/transfer-learning techniques [24] reuse knowledge from data-rich domains to the sparse target; they help when feature spaces align, but can suffer negative transfer when domain gaps are large.
Current practices lack a clear, topology-grounded rule for identifying newly joined services and for tracing their early role transitions. Most approaches provide no guarantee of nonzero scoring for cold-start or isolated services when the invocation graph is fragmented across domains. A unified mechanism that integrates role-aware filtering with a provably convergent propagation process remains largely absent.

2.6. Data Sparsity in Service Recommendation

2.6.1. Root Causes

Sparse interactions between users and services, often amplified by long-tail usage and multi-cloud fragmentation, lead to weak connectivity and even disconnected subgraphs; this undermines both similarity estimation and path-based propagation. Over time, sparsity not only reduces accuracy but also narrows system coverage, hiding potentially useful yet less popular services.

2.6.2. Mainstream Solutions

Matrix factorization [25] compresses the interaction matrix into low-rank factors to fill missing values; it works at moderate sparsity but degrades when observations are extremely scarce. Clustering-based methods [26] densifies interactions within groups and improves coverage; however, coarse clusters may blur individual differences. Graph-based propagation [27] exploits indirect paths to collect multi-hop evidence; this helps with few links but remains sensitive to graph disconnection and restart bias. Deep models (e.g., GNNs, autoencoders) [28] capture higher-order patterns beyond co-occurrence and can be robust in some sparse regimes; they nevertheless require large-scale training and still depend on sufficient structural signals.
Under extreme sparsity and cross-domain fragmentation, similarity estimation and path-based propagation become unreliable, leading to brittle recommendations. Existing solutions seldom combine role-aware filtering with formal coverage guarantees when the graph is disconnected. Practical assurances of nonzero scoring and stability in severely sparse regimes are still limited.

2.7. Summary and Research Gap

Existing research in CF, context-aware, QoS-aware and behavior-aware service recommendation has led to significant advances by leveraging interaction history [29], contextual signals [30], performance metrics [31] and behavioral patterns. However, these approaches commonly lack mechanisms to model or monitor service behavior in dynamic invocation environments. Specifically, few methods attempt to identify services with distinct behavioral roles, such as isolated, ending, or newly introduced services or consider how these roles evolve over time.In contrast, our work introduces a behavior-aware and structure-sensitive filtering mechanism based on the notion of DBS, explicitly modeling service roles within the invocation topology and enabling dynamic behavior tracking. This framework complements existing recommendation strategies by improving recommendation fairness, robustness against sparsity, and interpretability in multi-cloud environments.
In order to provide a clearer comparison of existing solutions, Table 1 summarizes the representative methods discussed in this section along five key capability dimensions, including role modeling, pre-filtering, dynamic adaptation, cold-start handling, and topology-awareness. As shown in the table, most methods focus on a single aspect of the problem, such as service semantic modeling or QoS-aware selection, while none of them simultaneously address the temporal dynamics, behavioral distinctions, and topology-aware adaptation in multi-cloud environments. In particular, existing works rarely incorporate topology-based behavioral similarity or dual-sided dynamic evolution (i.e., dynamic user behaviors and dynamic service pools), which leaves a methodological gap for robust and adaptive service recommendation under highly dynamic settings. This gap motivates the design of our DBS-aware framework, which integrates DBS identification, temporal sliding-window adaptation, and Markov-based compatibility scoring to provide a more holistic and adaptive solution.

3. Framework Overview

This section presents the overall architecture of our DBS-aware service recommendation system in multi-cloud environments. As illustrated in Figure 1, the architecture can be conceptually divided into two major components: the core backbone architecture and the behavioral parameter system derived from DBS strategies.

3.1. Backbone Architecture

The backbone of the system handles the full lifecycle of a user request, from input ingestion to recommendation output. It is composed of the following coordinated modules:
  • User Interface and API Gateway: The process begins when a user initiates a service request via the user interface. The request is routed through the API Gateway, which standardizes access and forwards requests to the backend logic.
  • Unified Data Management Module: This module serves as a bridge between user interaction and cloud resources. It is responsible for gathering invocation data from multiple cloud providers, synchronizing service descriptions, and triggering the construction of invocation topologies.
  • Data Integration and Standardization: Based on collected logs, the system incrementally constructs a time-aware service invocation graph G t = ( V t , E t ) within a sliding window, and performs initial domain assignment for each service node.
  • DBS-Recognition Module: Once the topology is established, the system identifies behaviorally significant services, referred to as Distinct Behavioral Services (DBS), by analyzing their connectivity and invocation roles.
  • DBS-Aware Filtering: A sliding time window strategy is applied to monitor role transitions over time. Services whose behavioral roles remain consistent over k consecutive windows are marked as stable and exempt from repeated updates. This dynamic judgment reduces noise and accelerates recommendation updates.
  • Recommendation Engine: Finally, the refined and behavior-aware service pool is passed to the recommendation engine, which generates personalized ranking results and returns them to the user.

3.2. Behavioral Parameter System via DBS

The yellow block in Figure 1 represents the DBS-aware parameter system, which encodes the behavioral semantics extracted during the filtering process. The system maintains a sliding time window of invocation topologies T [ b i j ] to sketch multi-cloud interactions, computes the Markov-based base score of each service, and extracts behavioral sets such as S n , S u e , S n e , and S i s . These symbols together support the identification and role maintenance of each service node, where the evolving role Role t ( s ) serves as the key indicator for filtering control. This lightweight symbolic system enables precise behavioral representation and supports strategy triggering without full recomputation, forming a compact yet expressive interface between topology dynamics and recommendation logic.

4. System Design

4.1. Distinct Behavioral Services: Definition and Mathematical Formulation

To capture the structural and behavioral characteristics of services within a dynamic multi-cloud environment, we define a special class of services, referred to as Distinct Behavioral Services (DBS). These services deviate from typical invocation patterns and are categorized into three types: Isolated Services, Ending Services, and New Member Services. Their classification is based on the topological features observed within a service invocation graph constructed over a sliding time window. The key mathematical symbols are summarized in Table 2.
The time window [ t Δ , t ] defines the temporal scope for constructing the invocation graph. Here, Δ represents the window size, and t is the current evaluation time. All service invocation records falling within this interval are considered valid when building G t . This design enables the system to capture short-term behavioral trends while adapting to the dynamic nature of service environments. The value of Δ can be tuned according to service invocation density or domain-specific time granularity.
Let the service invocation graph at time t be denoted as:
G t = ( V t , E t ) ,
where:
  • V t is the set of services active in the time window [ t Δ , t ] ;
  • E t V t × V t is the set of directed edges representing invocation relationships.
We define the following DBS categories:

4.1.1. Isolated Service

A service s V t is classified as Isolated if it has no interaction with other services in the time window:
deg t i n ( s ) + deg t o u t ( s ) = 0 .
Optionally, s may only invoke itself: ( s , s ) E t . Isolated Services are completely disconnected from the service invocation graph in the current time window, either due to inactivity or self-invocation only.

4.1.2. Ending Service

A service s V t is an Ending Service if it only appears at the terminal position of invocation workflows:
deg t o u t ( s ) = 0 and u V t , ( u , s ) E t .
Ending Services act as leaf nodes in invocation workflows, frequently terminating invocation chains, and are categorized based on whether the final invocation comes from the same or a different domain. We further distinguish:
  • Cross-domain Ending: the last invocation to s originates from a different domain;
  • Intra-domain Ending: the last invocation to s originates from the same domain.

4.1.3. New Member Service

A service s is considered a New Member if it has not appeared in any invocation record within the time window:
s V t and ( u , s ) E t ( s , u ) E t .
New Member Services are newly appeared services with no prior records, representing cold-start entities. Each service s is assigned a behavioral role based on these definitions:
Role ( s ) { Normal , Isolated , Ending , New Member } .
These definitions provide deterministic and reproducible DBS classification based on strict graph-theoretic conditions and forms the basis for downstream service filtering and behavior-aware scoring.

4.1.4. Impact of DBSs on Recommendation Accuracy

The three types of DBS introduce distinct disruptions to recommendation accuracy by interfering with the structural integrity and information flow within invocation topologies. Isolated Services, having no outward or inward connections, receive little to no propagation of collaborative information, resulting in severe score underestimation or exclusion. Ending Services, which often situated at the tail of invocation sequences and act as dead ends in the graph, preventing score propagation to earlier nodes and causing an asymmetry in score distribution. New Member Services, lacking any prior interactions, face absolute cold-start conditions, where traditional history-based models cannot even assign initial values. These conditions lead to information asymmetry in the recommendation system, where certain services are disproportionately undervalued, and introduce scoring imbalance across the pool. Without explicit recognition and compensation of these behavioral anomalies, the system risks biased rankings, missed high-quality candidates, and overall degradation in recommendation robustness. To mitigate these effects, we introduce a DBS identification and structure-aware scoring mechanism that explicitly models the behavioral roles of atypical services. By leveraging the invocation topology and its propagation paths, the mechanism compensates for data deficiencies and behavioral irregularities, ensuring that structurally significant services are assigned more reasonable and reliable scores.

4.2. Service Pool Topology Construction and Initial Scoring

To support behavior-aware analysis and recommendation, we first construct a directed invocation topology from service logs within a given sliding time window. Each invocation record is denoted as:
Record ( u , s , t ) ,
where user u invokes service s at time t.
From the filtered records, we build a directed service invocation graph:
G t = ( V t , E t ) ,
where:
  • V t represents the set of services active during the sliding time window [ t Δ , t ] ;
  • E t V t × V t represents the directed edges that denote invocation transitions. If service s i is followed by s j in a user session, then an edge ( s i , s j ) is added to E t .
To compute the preliminary importance or activation level of each service, we define a transition probability matrix M derived from G t :
M i , j = w i , j k w i , k ,
where w i , j is the frequency of invocation from s i to s j . This matrix captures the probability that a service call transitions from one service to another.
Let π t ( 0 ) denote the initial invocation probability vector at time t. A standard Markov Chain iteration computes the steady-state distribution as:
π t ( k + 1 ) = π t ( k ) M .
Once the sequence { π t ( k ) } converges, the resulting vector represents the base importance score for each service:
π t = lim k π t ( k ) .
Each score π t ( s ) reflects the centrality and activation likelihood of service s in the current topology. These scores are subsequently adjusted based on DBS classification in the filtering and recommendation stages.
Algorithm 1 is designed to compute the base importance scores for all services within a sliding time window, serving as the initialization for downstream recommendation and filtering tasks. The algorithm transforms raw invocation records into a directed service invocation graph, constructs a transition probability matrix based on invocation frequencies, and propagates service influence scores using Markov Chain dynamics until convergence. The core steps of the algorithm can be interpreted as follows:
Algorithm 1 Initial Service Scoring via Markov-Based Topology Analysis
  1:
Input: Invocation records R, time window [ t Δ , t ]
  2:
Output: Base importance score base _ score ( s ) for each service s
  3:
Filter R to obtain records in [ t Δ , t ]
  4:
Construct invocation graph G t = ( V t , E t )
  5:
Count transition frequencies w i j from E t
  6:
Compute transition matrix M i j = w i j k w i k
  7:
Initialize vector v 0
  8:
for all  s j V t  do
  9:
       v 0 [ j ] 1
10:
end for
11:
for all  s j V t  do
12:
       v 0 [ j ] ε                               ▹ ε > 0 to avoid zero-score propagation
13:
end for
14:
while not converged do
15:
       v t + 1 M T · v t
16:
       t t + 1
17:
end while
18:
for all  s j V t { unreachable nodes }  do
19:
       base _ score ( s j ) v t [ j ]
20:
end for

4.2.1. Initial Scoring via Markov Propagation

The Markov-based scoring mechanism operates on the invocation records within a sliding time window. The scoring procedure, as detailed in Algorithm 1, can be interpreted as follows:
  • Lines 1–4: Filter the raw invocation records R within the window [ t Δ , t ] , and construct a directed invocation graph G t = ( V t , E t ) .
  • Lines 5–6: Count transition frequencies w i j and normalize them into a row-stochastic transition matrix M, where M i j represents the probability of invoking service s j after  s i .
  • Line 7: Declare and initialize the score vector v 0 .
  • Lines 8–13: For each service, assign v 0 [ j ] = 1 if s j V t ; otherwise assign a small constant ε > 0 to avoid zero-score contamination.
  • Lines 14–17: Perform iterative matrix-vector multiplication v t + 1 M T · v t until convergence.
  • Lines 18–19: Assign the converged scores v t [ j ] as the base scores b a s e _ s c o r e ( s j ) for each service.
  • Line 20: End of loop.

4.2.2. Time and Space Complexity

Let n = | V t | be the number of services, m = | E t | the number of edges, | R | the number of raw invocation records, and K the number of iterations until convergence. The time complexity of each component is:
  • Lines 1–2: Filtering and graph construction takes O ( | R | ) .
  • Line 3: Transition frequency counting takes O ( m ) .
  • Lines 4–5: Score vector initialization takes O ( n ) .
  • Line 6: Transition matrix normalization takes theoretically O ( m 2 ) in the worst case, but O ( | R | + K · m ) in the proposed mechanism.
  • Lines 7–9: Score value assignment for reachable and unreachable nodes takes O ( n ) .
  • Lines 10–13: Markov propagation iterations takes O ( K · m ) .
  • Lines 14–18: Final score assignment takes O ( n ) .
Hence, the total time complexity is:
O ( | R | + m 2 + K · m )
The space complexity, dominated by the storage of the sparse transition matrix and the score vector, is:
O ( n + m )

4.2.3. Practical Complexity with Fixed-Invocation Windows

Although the theoretical worst-case complexity may reach O ( m 2 ) during the construction of the transition matrix M in Line 6, this situation only arises when the invocation graph becomes densely connected, leading to a large number of non-zero entries in each row. However, in our design, the sliding window mechanism is based on a fixed number of invocation operations (i.e., Δ = 100 ), rather than a fixed time span. This design ensures that the number of recorded invocations per window is controlled, effectively limiting the number of edges and keeping the resulting graph sparse.
As a result, the actual complexity of Markov scoring is O ( | R | + K · m ) , where | R | denotes the number of invocations in the window, K is the number of Markov iterations, and m is the number of active services. Since | R | is fixed and the graph remains sparse across all windows, the matrix M is efficiently computable and the overall scoring process maintains near-linear complexity in practical scenarios.

4.3. DBS-Based Service Filtering Algorithm

After assigning behavioral roles to services based on the topology structure, we apply a DBS-based filtering algorithm to refine the candidate service pool prior to recommendation. This step improves both computational efficiency and semantic relevance by eliminating low-value services and prioritizing behaviorally significant ones.
The filtering process is executed in two phases.

4.3.1. Phase I: Behavioral Role Sorting

In the first phase, each service is examined within its current time-slot topology matrix. The row and column activity vectors are computed to detect null interactions. Services are classified as Isolated, Ending, or New Member based on their invocation activity and, in cases beyond the first time-slot, whether they appeared in previous topologies. This classification is formalized in Algorithm 2.
Algorithm 2 DBS Sorting Algorithm
  1:
Input:  T k [ a i j ] : Topology matrix at time slot k
  2:
Output:  S i s , S e , S n : Sets of Isolated, Ending, and New Member Services
  3:
for  i = 1 to n do
  4:
     r o w [ i ] any ( T k [ a i j ] , 1 )
  5:
     c o l [ i ] any ( T k [ a i j ] , 2 )
  6:
end for
  7:
for  i = 1 to n do
  8:
    if  r o w [ i ] = 0  then
  9:
        if  c o l [ i ] = 0  then
10:
           if  k = 1  or  i T k 1  then
11:
                S n [ i ] i
12:
           else
13:
                S i s [ i ] i
14:
           end if
15:
        else
16:
            S e [ i ] i
17:
        end if
18:
    end if
19:
end for
Algorithm Explanation
  • Lines 3–6: Filter invocation relationships using the topology matrix T k , and traverse all service nodes to prepare for behavioral classification.
  • Lines 7–14: For services with neither incoming nor outgoing edges, determine if they are newly appeared or belong to the first time slot; classify them as New Members or Isolated Services accordingly.
  • Lines 15–17: For services that are only invoked by others but never invoke any service, classify them as Ending Services.
Complexity Analysis
Let n be the number of services and m be the number of time slots.
  • Computing the row and column activity vectors takes O ( n ) for each time slot.
  • The unified iteration over n services involves only constant-time checks and conditional assignments.
  • Overall time complexity per time slot: O ( n ) .
  • Total complexity over m time slots:
    O ( m · n ) .

4.3.2. Phase II: Ending Role Refinement

In the second phase, the previously detected Ending Services are further divided into Cross-Domain Ending, -, or Universal Ending types by analyzing domain boundaries and inter-domain invocation patterns. This additional semantic distinction is detailed in Algorithm 3.
Algorithm 3 Ending Services Sorting Algorithm
  1:
Input:  T k [ a i j ] : Topology matrix at time k; S e [ i ] : Ending Services set; D i : Domain of i
  2:
Output:  S c e : Cross-Domain Endings; S u e : Universal Endings; S w e : Within-Domain Endings
  3:
for  i = 1 to n and  i D i  do
  4:
    if  c o l [ i ] = 0  then
  5:
         S c e [ i ] i
  6:
    else
  7:
         S u e [ i ] i
  8:
    end if
  9:
end for
10:
for each  i D i  do
11:
    if  c o l [ i ] = 0  then
12:
         S w e [ i ] i
13:
    else
14:
         S u e [ i ] i
15:
    end if
16:
end for
Algorithm Explanation
  • Lines 3–9: For each ending service, examine cross-domain interactions. If it receives calls from services outside its domain, it is labeled as Cross-Domain Ending; otherwise, it is initially assigned to Universal Ending.
  • Lines 10–15: Within the same domain, further classification determines whether the service is Within-Domain Ending or remains Universal.
Complexity Analysis
The domain-based classification requires a single comparison per service. Assuming domain membership lookup is O ( 1 ) , the overall complexity remains:
O ( m · n ) .

4.3.3. Impact of Filtering

The results from both phases determine the behavioral status of each service, which is then used to control its inclusion and impact in the recommendation engine. Isolated Services are typically filtered out due to their lack of structural engagement. Ending Services receive score amplification due to their contextual significance, and New Member Services are conditionally retained with neutral or default weights.
This filtering step serves as a lightweight, rule-based pre-processing module that enhances the interpretability and effectiveness of the final recommendation phase without adding substantial computational overhead.

4.4. Sliding Time Window Strategy and Role Judgment

To support dynamic adaptation in service behavior analysis, we implement a sliding time window strategy that incrementally updates the invocation topology and behavioral roles of services over time, shown in Figure 2, this mechanism enables the system to reflect temporal patterns and evolving invocation contexts without requiring full historical recomputation.
Let Δ denote the window size and τ the sliding step. At each evaluation timestamp t, the system constructs the service invocation graph G t = ( V t , E t ) using records within the time interval [ t Δ , t ] . After computing the behavioral roles role t ( s ) for all s V t , the system waits for τ time units before updating the window to [ t + τ Δ , t + τ ] , and recomputes G t + τ and role t + τ ( s ) accordingly.
To capture role evolution and reduce noise due to temporary fluctuations, we define a stability condition that monitors whether a service maintains the same role across multiple consecutive windows. Formally, for a service s, its role is considered temporally stable if:
role t ( s ) = role t τ ( s ) = = role t k τ ( s ) ,
where k is a predefined stability threshold. Once a service satisfies this condition, its role label can be frozen, avoiding unnecessary recomputation or misclassification due to transient invocation patterns.
This sliding mechanism enables the DBS filtering module to dynamically respond to real-time behavior while preserving consistency for structurally stable services. It also allows for efficient incremental updates to recommendation scores when combined with streaming-friendly architectures.
  • Algorithm Explanation
Algorithm 4 maintains the temporal evolution of service behavioral roles using a sliding time window approach. At each update step, it extracts recent invocation records to reconstruct the topology, recomputes behavioral roles using the DBS filtering logic (Algorithms 2 and 3), and checks the stability of each service’s role across previous windows.
  • Lines 3–5: Initialize the current timestamp and construct the current invocation graph G t from logs in [ t Δ , t ] .
  • Line 6: Behavioral roles for all active services are recomputed based on current topology and domain-aware DBS rules.
  • Lines 7–11: For each service, check whether its current role matches the previous k roles over τ -spaced windows. If so, the role is marked as frozen to avoid redundant updates in future steps.
  • Line 12–13: Move the time window forward by τ and repeat the process.
Algorithm 4 Sliding Window Update Procedure for DBS Role Maintenance
  1:
Input: Service logs R = u i , s j , t k ; window size Δ ; step size τ ; stability threshold K
  2:
Output: Updated role labels role t ( s ) for each service s
  3:
Initialize t Δ
  4:
while new data arrives do
  5:
    Filter R within window [ t Δ , t ] to construct G t = ( V t , E t )
  6:
    Compute role t ( s ) for all s V t using Algorithms 2 and 3
  7:
    for all  s V t  do
  8:
        if  role t ( s ) = role t τ ( s ) = = role t K τ ( s )  then
  9:
           Mark role t ( s ) as frozen
10:
        end if
11:
    end for
12:
     t t + τ
13:
end while
  • Time Complexity
Let n be the number of services, m the number of invocation edges within the window, and T the number of total time units (i.e., number of windows processed). The complexity per update step includes:
  • Filtering records and constructing G t : O ( | R t | ) , where | R t | is the number of logs in window [ t Δ , t ] .
  • Computing roles via Algorithms 2 and 3: O ( n ) (assuming per-slot DBS filtering is linear).
  • Checking stability for all n services across k prior windows: O ( k · n ) .
Hence, the time complexity per iteration is:
O ( | R t | + k · n ) ,
and the total complexity over T updates is:
O T · ( | R t | + k · n ) .
  • Space Complexity
The algorithm maintains:
  • The current invocation graph G t : O ( n + m ) ;
  • A short history of previous k roles per service: O ( k · n ) ;
Thus, the total space complexity is:
O ( n + m + k · n ) .
  • Stability Efficiency
The use of role freezing significantly reduces redundant recomputation for behaviorally stable services. Once a service is frozen, its role does not need to be re-evaluated in subsequent iterations unless its topology context changes, thus enhancing performance in dynamic but stable environments.

5. Simulations

5.1. Dataset

5.1.1. Brief Introduction

All simulations are conducted on the Santander Smart City dataset [33] from the SmartSantander EU Project, which records fine-grained service invocations across multiple urban domains. Although the dataset originates from IoT-based city services, its multi-domain invocation structure resembles heterogeneous service deployments commonly observed in multi-cloud environments, where domain-specific services may be operated by different departments or vendors [34]. This characteristic enables us to reasonably extend the dataset to simulate multi-cloud scenarios.
To adapt Santander for our evaluation, we apply a randomized but balanced allocation strategy to distribute services across several logical clouds, ensuring that each cloud hosts a comparable number of services while preserving the dataset’s inherent invocation topology. This design allows us to simulate both inter-cloud and intra-cloud interactions in a controlled manner without introducing artificial structural bias. Furthermore, Santander provides large-scale, dynamic, and diverse invocation traces, which makes it particularly suitable for evaluating behavior-aware recommendation frameworks like ours, especially under settings involving temporal evolution and cross-domain behavioral dynamics.

5.1.2. Preprocessing

Each invocation record is structured as a tuple u i , s j , t k , indicating that user u i invoked service s j at time t k . To adapt the dataset for our framework, we preprocess the original fields and semantically map them to cloud service concepts. The mapping is summarized in Table 3. We adopted a randomized yet balanced allocation strategy: services are randomly assigned to clouds while ensuring each cloud hosts a comparable number of services and retains a diverse mix of service types. This design avoids the creation of ‘single-function clouds’, which could distort compatibility evaluation, and instead mimics real-world multi-cloud ecosystems where services are deployed heterogeneously across providers.
To simulate a multi-cloud deployment, we partition the full service pool into two logical clouds, C 1 with 8115 services and C 2 with 8101 services, based on service ID ranges or domain attributes. This separation enables the modeling of inter-cloud invocations while maintaining controllable experimental scope.
For dynamic analysis, we adopt a sliding window mechanism with window size Δ and step size τ to segment the data temporally. Within each window, we extract sequential invocation paths and construct a directed graph G t = ( V t , E t ) , where nodes represent services and edges represent temporal invocations. This evolving topology forms the basis for DBS identification, scoring, filtering, and recommendation evaluation in our simulations.

5.2. Simulation Design

To comprehensively evaluate the effectiveness of our DBS-aware framework, we design five targeted groups of simulations. Each group focuses on a specific aspect of system performance, ranging from recommendation accuracy to behavioral stability and robustness under real-world constraints. The design intentionally aligns with the role semantics and structural innovations proposed in this paper.
  • Group I: Ablation Study on DBS-Aware Recommendation. To validate the effect of DBS-aware mechanisms, we performed an ablation study using synthetic invocation data. The dataset simulated a service pool with 100 services and 1000 invocation records, where a configurable proportion (10%, 20%, 30%) of services were artificially marked as DBS. These included Isolated Services, Ending Services, and New Member Services, all exhibiting behavioral anomalies such as lack of invocations or unbalanced connectivity.
    We compared two system variants:
    NoDBS: The baseline system without any DBS identification or filtering.
    FullDBS: The enhanced system that integrates full DBS detection and filtering workflows.
    For each setting, a fixed user group was used to simulate invocation-based recommendation, and Precision, Recall, and F1-score were calculated based on the overlap between recommended and relevant services. The simulation was repeated under three different DBS ratios to assess the robustness and contribution of the DBS module under varying degrees of service pool contamination.
  • Group II: Recommendation Performance. This group evaluates how the DBS-enhanced service filtering affects the final recommendation quality. Specifically, we assess the ranking performance using Precision, Recall, and F1-score under varying Top-N recommendation thresholds. These metrics jointly reflect the accuracy, completeness, and overall reliability of the recommendation results. Comparisons are made against both classical and state-of-the-art baselines.
  • Group III (a): Window Size Sensitivity Simulation. This group evaluate the stability of behavior role classification under sliding window mechanisms and pick the most suitable length of the window size, we designed a window size sensitivity simulation based on synthetic data. Instead of using fixed time intervals, this simulation adopts a fixed number of service invocations as the segmentation unit, thereby eliminating biases introduced by uneven invocation densities.
  • Group III (b): Temporal Role Stability. This group measures the stability of recommendation results under the sliding time window. It investigates whether the role judgment and filtering mechanisms introduce volatility or maintain consistent service outputs across adjacent windows. Top-N list overlap and Jaccard Similarity are used as core indicators.
  • Group IV: Coverage under Data Sparsity. This group validates whether the Markov-based scoring process, combined with DBS filtering, can sustain service recommendation coverage when invocation records become sparse. It simulates controlled data sparsity ratios and measures the impact on overall service recommendation coverage.
  • Group V: Cold-Start Scoring Capability. This group specifically targets cold-start services—particularly those identified as New Member and Isolated by the DBS mechanism. It measures whether these behaviorally low-frequency services can still be scored and ranked effectively, using our irreducible Markov process. The evaluation metric is cold-start scoring rate (CSR).
These four simulation groups together provide a well-rounded evaluation of our framework from the perspectives of performance, consistency, robustness, and adaptability.
Finally, to ensure that the performance improvements of the proposed framework over baseline methods are not only empirical but also statistically validated, we planned a paired t-test as part of the simulation design. This statistical test compares the proposed approach with all baselines (GNN, TCF, TREP, and MCTSE) under both experimental scenarios, providing a formal significance assessment to support the reliability of the conclusions.
All simulations were implemented in Matlab 2024b and conducted on a workstation equipped with an Intel i7 CPU and 32 GB RAM.

5.3. Evaluation Metrics

To enable targeted evaluation, we adopt four types of metrics across the four simulation groups, as summarized in Table 4. Each metric is designed to evaluate a specific aspect of the framework, such as recommendation quality, temporal stability, scoring coverage, and cold-start support.

5.4. Baseline

To prove the effectiveness of our proposed framework, we choose four state-of-the-art research works to be the baseline of our simulations.
Trust-Based Approach (TCF) [35]: Trust-Based Approach: A service recommendation that relies on a distributed collaborative filtering rating system using friendship and social relationships as the filtering criteria.
Interface Matching Graph Neural Network (GNN): [36] This GNN-based service recommendation method combines an interface-matching graph construction process with a two-layer GCN scoring module. In our study, we adapted IM GNN by using invocation graphs as input to fairly compare its GNN-based scoring with our DBS framework.
Trust and Reputation Approach (TREP) [37]: This approach to service recommendation which integrates trust and reputation within the recommendation process.
Mutual Context-Aware Trustworthy Service Evaluation (MCTSE) [38]: This framework is another trust-based recommendation system that leverages the time and location of services to establish trust among them.
A comparative evaluation with representative state-of-the-art methods—TCF, TREP, MCTSE and GNN—is conducted using precision, recall, and F1-score metrics, as summarized in Section 6.

6. Results and Analysis

6.1. Ablation Study on DBS-Aware Recommendation

To evaluate the contribution of DBS-aware mechanisms to recommendation performance, we conducted an ablation study comparing two variants: the baseline without DBS filtering (NoDBS), and the full version integrating DBS identification and filtering (FullDBS). The proportion of DBS in the service pool was set to 10%, 20%, and 30%, and evaluation was performed based on Precision, Recall, and F1-score.
As shown in Figure 3, FullDBS consistently outperforms NoDBS across all metrics and DBS ratios. In particular, the F1-score exhibits the most significant improvement, with relative gains exceeding 130% on average. Precision and Recall also show substantial improvements, generally ranging between 85% and 135%. These results clearly demonstrate that DBS-aware filtering helps eliminate low-quality services and enhance the relevance of recommended items.
Notably, as the DBS ratio increases, the performance gap between FullDBS and NoDBS becomes more pronounced, confirming that proper detection and handling of behavioral outliers (e.g., isolated or newly joined services) is essential for maintaining recommendation accuracy and stability.

6.2. Recommendation Accuracy Under Dynamic Environments

To evaluate the robustness and adaptability of the proposed DBS-based model in such scenarios, we design two complementary simulations:
  • Dynamic Service Pool: Simulates a gradually expanding set of services, increasing from 1000 to 10,000. This setting tests the model’s ability to score and recommend effectively in the presence of newly added, low-visibility services.
  • Dynamic User Crowd: Simulates a growing population of service requesters, ranging from 1000 to 10,000. This scenario evaluates how the model responds to shifts in user behavior and the emergence of sparse interaction patterns.
In both simulations, we evaluate five methods: TCF, TREP, MCTSE, GNN, and the proposed method. Precision, Recall, and F1-score are used as performance metrics. All methods share a unified recommendation pipeline to ensure fair comparison, with DBS detection performed uniformly before scoring and ranking.

6.2.1. Performance Under Dynamic Service Pool

Shown in Figure 4, as the number of available services increases, most baseline methods experience a clear decline across all three metrics. TREP and GNN show the most severe degradation, with precision and recall dropping by more than 20%, highlighting their inability to cope with newly added, sparsely connected services. TCF and MCTSE maintain relatively better performance but still exhibit instability. Notably, MCTSE suffers from a gradual decline in precision, likely due to its bias toward highly connected nodes, which becomes problematic when the pool expands and the average connectivity drops.
In contrast, the proposed method demonstrates outstanding robustness. Precision remains above 0.9 even as the pool size increases tenfold, while recall and F1-score show minimal variance. Compared with the best-performing baseline, our model achieves a 46.7% improvement in precision, 42.79% in recall, and 43.76% in F1-score under this setting. This is largely due to the DBS filtering mechanism, which removes unstable or uninvoked services—such as Isolated or Ending types—thereby improving the overall quality of the candidate set and mitigating cold-start effects.

6.2.2. Performance Under Dynamic User Crowd

Figure 5 visualizes the performance trends under user-side dynamics. As user count increases from 1000 to 10,000, most baseline methods show a notable drop in recall. GNN, in particular, falls below 0.65 in high-volume settings, revealing its limited ability to generalize to unseen user behaviors. MCTSE maintains competitive performance in smaller-scale settings but shows increasing volatility in recall and F1-score as the user base grows. This suggests that while it handles structural noise reasonably well, it does not adapt effectively to behavioral variation introduced by dynamic users.
In contrast, the proposed method maintains high and stable performance, with all three metrics remaining above 0.9 across the entire user range. Notably, the model delivers a 40.5% increase in precision, 41.2% in recall, and 39.62% in F1-score compared to the best-performing baseline in this setting. This robustness is attributed to two key factors: (i) DBS filtering helps stabilize the service graph regardless of user fluctuation, and (ii) the Markov scoring ensures smooth and fair score propagation even when user-service interactions are sparse or newly formed.

6.3. Temporal Role Stability Under Sliding Window

6.3.1. Window Size Sensitivity

To further evaluate the stability of service behavior identification under different sliding window configurations, we conducted a sensitivity analysis based on window size. Instead of using fixed time intervals, we adopted the number of invocation operations as the unit of segmentation to mitigate density imbalance caused by usage fluctuations. Specifically, we simulated a service pool with 100 services and 1000 total invocations, and divided the timeline into windows of varying sizes: 25, 50, 75, 100, 125, 150, 175, and 200 operations per window. Within each window, the system identified behavioral roles for services (i.e., Isolated, Ending, and New Member), and we computed the role consistency across adjacent windows. Two metrics were used for evaluation:
  • Jaccard Similarity: measures the overlap of role-specific service sets between adjacent windows.
  • Preservation Rate: reflects how consistently a service maintains its assigned role across windows.
As shown in the Figure 6, both metrics peak at a window size of 100, reaching 0.91 and 0.88 respectively. These results indicate that a window size of 100 offers optimal stability for role classification, and it is thus adopted as the default configuration in subsequent simulations.

6.3.2. Temporal Role Stability

This simulation evaluates the temporal stability of recommendation results when applying the proposed sliding time window strategy and behavioral role freezing. We aim to test whether filtering services based on stable DBS roles reduces noise and prevents recommendation list volatility over time.
We segment the service invocation data using a sliding window strategy based on operation counts rather than time intervals, aligning with the findings of our window size sensitivity analysis. Specifically, we set the window size to 100 service invocations and a sliding step size of 25 invocations. For each window, the system constructs an invocation graph G t = ( V t , E t ) and generates Top-N recommendations for active users ( N = 10 ). We compare two variants of the system across 20 consecutive windows and evaluate the consistency of recommendation outputs.
Two variants of the system are compared:
  • Ours (Sliding Windows(SW) + freeze): we refer to a common strategy in temporal smoothing and behavioral stability analysis, where consistency observed across three or more consecutive observations is typically considered a strong indicator of stable behavior. Services whose roles remain unchanged for k = 3 consecutive windows are frozen, and their roles are reused in future filtering.
  • Ours (Slinding Windows (SW)): Role judgment is recomputed independently in every window, without considering temporal consistency.
Figure 7 shows the Jaccard@10 scores over 20 sliding windows. The method with role freezing consistently achieves higher overlap (typically above 0.75), while the version without freezing fluctuates at lower values (around 0.62). This indicates that the freezing strategy stabilizes the recommendation list across time.
In Figure 8, we visualize the Preservation@10 metric. Similar to Jaccard, the method with role freezing shows higher preservation rates, indicating that previously recommended services are more likely to persist in subsequent windows. This reinforces the idea that the role stability mechanism prevents premature removal of useful services due to transient behavior shifts. Both metrics confirm that our sliding window strategy with behavioral role freezing improves the temporal stability of recommendations. Compared to the setting without freezing, our proposed strategy achieves a 22.85% improvement across sliding windows, demonstrating its effectiveness in maintaining recommendation stability over time. It helps retain high-confidence services and avoids fluctuations caused by unstable DBS classifications, resulting in more consistent and reliable recommendation behavior over time.

6.4. Coverage Under Data Sparsity

We evaluate scoring coverage under five predefined sparsity levels: {10%, 30%, 50%, 70%, 90%}, representing increasingly severe data sparsity in the service invocation records. For each sparsity level, a corresponding proportion of invocation data is randomly removed from the original dataset to simulate incomplete service histories. Based on the resulting topology, we apply the scoring process and calculate the Scoring Coverage Rate (SCR), which measures the proportion of services receiving non-zero scores. A higher SCR indicates better robustness to sparse data conditions.
The results are shown in Figure 9. As the sparsity level increases, both TCF and TREP exhibit consistent drops in scoring coverage. For example, at 90% sparsity, their SCRs decrease to 32% and 38%, respectively, indicating that a large portion of services fail to receive valid scores due to insufficient invocation data or limited connectivity.
In contrast, our proposed Markov Chain-based method demonstrates significantly stronger resilience. Even under the most extreme sparsity condition (90%), it achieves an SCR of 88%, and steadily improves as more data is available—reaching 97% when sparsity drops to 10%. This consistent high coverage is aligned with the theoretical guarantee of our method, which ensures complete score propagation as long as the service topology remains connected and irreducible.
To highlight this point, we include a horizontal reference line representing the theoretical upper bound of 100% SCR, derived from the homogeneous irreducibility of our Markov Chain model. The minor gap between the actual and ideal coverage stems primarily from occasional topological fragmentation during random edge removal. Nevertheless, our approach consistently outperforms TCF and TREP, showing clear superiority in handling sparse service environments.

6.5. Cold-Start Scoring Rate on DBS Services

DBS exhibit strong cold-start characteristics due to their limited or nonexistent invocation histories. Traditional recommendation methods, such as Collaborative Filtering (TCF) and trust-based approaches (TREP), often fail to assign scores to these services, resulting in poor coverage and ineffective recommendations. To assess the cold-start handling capability of our model, we define the Cold-Start Scoring Rate (CSR) as the proportion of DBS services receiving non-zero scores. All three types of DBS services are considered, and the CSR is evaluated across 20 randomized simulations to ensure statistical stability. The compared methods include: (i) TCF, (ii) TREP, and (iii) our proposed Markov Chain-based scoring method.
The average CSR results and standard deviations over 20 experimental runs are summarized in Figure 10. Our proposed model significantly outperforms baseline methods across all DBS categories. In particular, it achieves a CSR of 95% on New Member services, whereas TCF and TREP score only 22% and 23%, respectively. Similarly, for Isolated and Ending services, our method maintains CSR above 88%, while the baselines drop below 45%.
These results validate the model’s strong cold-start robustness, which stems from its reliance on the topological structure of the service invocation graph rather than historical co-occurrence. Furthermore, the low standard deviation across all runs demonstrates the consistency of the scoring mechanism. Although the theoretical CSR of our model is expected to reach 100% due to its homogeneous and irreducible Markov Chain properties, slight deviations may still occur. These are primarily caused by fully disconnected nodes in the DBS set or by numerical thresholds that treat extremely low scores as zero. Nevertheless, the achieved performance remains superior in both coverage and stability.

6.6. Statistical Significance Test

To further validate the reliability of our experimental results, we conducted a paired t-test to examine whether the performance improvements of the proposed method over baseline approaches are statistically significant. The test was performed on the F1-scores obtained across all device count settings under both experimental scenarios: Dynamic User Crowd and Dynamic Pool.
Specifically, we compared the Proposed framework against four representative baselines (GNN, TCF, TREP, and MCTSE) using paired samples from identical experimental configurations. Shown in Table 5, the resulting p-values are consistently far below the commonly accepted threshold of 0.05 (in the order of 10 9 to 10 12 ), confirming that the observed improvements are not due to random variation but are statistically significant.
This statistical validation reinforces the conclusion that the proposed method consistently and significantly outperforms existing baselines across diverse experimental settings, providing stronger evidence for the robustness and generalizability of our framework.

7. Conclusions

7.1. Summary of Contributions

In this paper, we introduced the concept of Distinct Behavioral Services (DBS) and proposed a structural classification framework that identifies special service types such as New Member, Isolated, and Ending services based on invocation topology. These services typically suffer from extreme sparsity or lack of co-invocation records, posing serious challenges for conventional recommendation systems. To address this, we designed a unified scoring model that integrates DBS detection with a Markov Chain-based scoring mechanism. This approach ensures that even under cold-start or sparse conditions, services can receive consistent and topology-aware scores. The proposed framework is scalable and domain-agnostic, making it suitable for complex multi-cloud service environments.
Extensive simulations were conducted under four perspectives: recommendation accuracy, stability under sliding time windows, scoring coverage under data sparsity, and cold-start performance on DBS services. Results confirm the superiority of our approach over existing collaborative filtering and topology-based methods. In particular, our method consistently achieves higher precision, recall, and F1 score, while maintaining robustness in dynamic user and service environments.

7.2. Limitations and Future Work

Although the proposed DBS-aware recommendation framework demonstrates strong performance and robustness across different dynamic scenarios, several limitations remain.
(1)
Dataset Generalizability.
The Santander Smart City dataset reflects urban IoT-style service interactions, which are not fully equivalent to real-world multi-cloud service invocation ecosystems. While it provides representative dynamic behaviors and invocation patterns, the absence of genuine cross-cloud workflow traces may limit the direct generalizability of the evaluation results.
(2)
Topology Reconstruction Assumptions.
The framework assumes that invocation relations can be accurately extracted from logs to construct topology matrices. In large-scale multi-cloud deployments, log completeness, monitoring granularity, or privacy restrictions may introduce missing or noisy topology information.
(3)
Temporal Window Sensitivity.
The sliding-window mechanism involves selecting a fixed window size. Although experiments demonstrate stable performance across different settings, extreme fluctuations in service pool dynamics may require adaptive or learned window sizes.
Future work will focus on collecting real multi-cloud workflow traces, designing adaptive temporal windows, and exploring learning-based topology extraction to further improve the applicability of DBS-aware recommendation in large-scale cloud environments.

Author Contributions

Conceptualization, S.M.; Methodology, S.M.; Software, S.M. and X.G.; Validation, S.M.; Formal analysis, S.M.; Investigation, L.X. and Z.D.; Writing—original draft, S.M.; Writing—review & editing, S.M. and X.D.; Visualization, S.M.; Supervision, X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major Research plan of the National Natural Science Foundation of China (No. 92267204).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, X.; Zheng, J.; Ren, B.; Dong, X.; Shen, Y. Microthingschain: Blockchain-based controlled data sharing platform in multi-domain iot. J. Netw. Netw. Appl. 2021, 1, 19–27. [Google Scholar]
  2. Ko, H.; Lee, S.; Park, Y.; Choi, A. A Survey of Recommendation Systems: Recommendation Models, Techniques, and Application Fields. Electronics 2022, 11, 141. [Google Scholar] [CrossRef]
  3. Peng, J.; Gong, J.; Zhou, C.; Zang, Q.; Fang, X.; Yang, K.; Yu, J. KGCFRec: Improving Collaborative Filtering Recommendation with Knowledge Graph. Electronics 2024, 13, 1927. [Google Scholar] [CrossRef]
  4. Fan, X.; Hu, Y.; Zheng, Z.; Wang, Y.; Brézillon, P.; Chen, W. CASR-TSE: Context-Aware Web Services Recommendation for Modeling Weighted Temporal-Spatial Effectiveness. IEEE Trans. Serv. Comput. 2021, 14, 58–70. [Google Scholar] [CrossRef]
  5. Liu, Z.; Sheng, Q.Z.; Xu, X.; Chu, D.; Zhang, W.E. Context-Aware and Adaptive QoS Prediction for Mobile Edge Computing Services. IEEE Trans. Serv. Comput. 2022, 15, 400–413. [Google Scholar] [CrossRef]
  6. Mohapatra, S.S.; Kumar, R.R.; Alenezi, M.; Zamani, A.T.; Parveen, N. QoS-Aware Cloud Service Recommendation Using Metaheuristic Approach. Electronics 2022, 11, 3469. [Google Scholar] [CrossRef]
  7. Darabkh, K.A.; Al-Akhras, M.; Zomot, J.N.; Atiquzzaman, M. RPL routing protocol over IoT: A comprehensive survey, recent advances, insights, bibliometric analysis, recommendations, and future directions. J. Netw. Comput. Appl. 2022, 207, 103476. [Google Scholar] [CrossRef]
  8. Aladiyan, A. Efficient Data Structures and Algorithms for Cloud Computing Platforms. In Proceedings of the 2024 4th International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Greater Noida, India, 14–15 May 2024; pp. 1717–1721. [Google Scholar] [CrossRef]
  9. Chen, W.; Chen, Y.; Liu, J. Service migration for mobile edge computing based on partially observable Markov decision processes. Comput. Electr. Eng. 2023, 106, 108552. [Google Scholar] [CrossRef]
  10. Addya, S.K.; Satpathy, A.; Ghosh, B.C.; Chakraborty, S.; Ghosh, S.K.; Das, S.K. Geo-distributed multi-tier workload migration over multi-timescale electricity markets. IEEE Trans. Serv. Comput. 2023, 16, 3385–3396. [Google Scholar] [CrossRef]
  11. Yang, C.; Liu, Y.; Ding, Y.; Liang, H. Secure data migration from fair contract signing and efficient data integrity auditing in cloud storage. J. Netw. Comput. Appl. 2025, 239, 104173. [Google Scholar] [CrossRef]
  12. Zhang, H.; Yang, D. Research on User Demand-Driven Service Matching Methodology. In Proceedings of the 2021 IEEE 6th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 24–26 April 2021; pp. 511–516. [Google Scholar] [CrossRef]
  13. Theodoropoulos, T.; Makris, A.; Boudi, A.; Taleb, T.; Herzog, U.; Rosa, L.; Cordeiro, L.; Tserpes, K.; Spatafora, E.; Romussi, A.; et al. Cloud-based xr services: A survey on relevant challenges and enabling technologies. J. Netw. Netw. Appl. 2022, 2, 1–22. [Google Scholar] [CrossRef]
  14. Pan, T. Personalized Recommendation Service in University Libraries using Hybrid Collaborative Filtering Recommendation System. In Proceedings of the 2024 International Conference on Intelligent Algorithms for Computational Intelligence Systems (IACIS), Hassan, India, 23–24 August 2024; pp. 1–5. [Google Scholar] [CrossRef]
  15. Zhao, X.; Liu, C.; Zhang, S.; You, X. A Novel Science and Technology Resource Recommendation Service based on Knowledege Graph and Collaborative Filtering. In Proceedings of the 2022 International Conference on Service Science (ICSS), Zhuhai, China, 13–15 May 2022; pp. 196–202. [Google Scholar] [CrossRef]
  16. Wang, Z.; Sun, C.A.; Aiello, M. Context-aware IoT Service Recommendation: A Deep Collaborative Filtering-based Approach. In Proceedings of the 2022 IEEE International Conference on Web Services (ICWS), Barcelona, Spain, 10–16 July 2022; pp. 150–159. [Google Scholar] [CrossRef]
  17. Wang, J.; Zhang, X.; Wang, Q.; Zheng, W.; Xiao, Y. QoS Prediction Method via Multi-Task Learning for Web Service Recommendation. In Proceedings of the 2024 IEEE International Conference on Web Services (ICWS), Shenzhen, China, 7–13 July 2024; pp. 1353–1355. [Google Scholar] [CrossRef]
  18. Zhang, S.; Zhang, D.; Wu, Y.; Zhong, H. Service Recommendation Model Based on Trust and QoS for Social Internet of Things. IEEE Trans. Serv. Comput. 2023, 16, 3736–3750. [Google Scholar] [CrossRef]
  19. Kim, D.; Tanwar, S.; Kang, U. Accurate multi-behavior sequence-aware recommendation via graph convolution networks. PLoS ONE 2025, 20, e0314282. [Google Scholar] [CrossRef]
  20. Su, J.; Chen, C.; Lin, Z.; Li, X.; Liu, W.; Zheng, X. Personalized Behavior-Aware Transformer for Multi-Behavior Sequential Recommendation. arXiv 2024, arXiv:2402.14473. [Google Scholar]
  21. Herce-Zelaya, J.; Porcel, C.; Bernabé-Moreno, J.; Tejeda-Lorente, A.; Herrera-Viedma, E. New technique to alleviate the cold start problem in recommender systems using information from social media and random decision forests. Inf. Sci. 2020, 536, 156–170. [Google Scholar] [CrossRef]
  22. Li, Z.; Amagata, D.; Zhang, Y.; Maekawa, T.; Hara, T.; Yonekawa, K.; Kurokawa, M. HML4Rec: Hierarchical meta-learning for cold-start recommendation in flash sale e-commerce. Knowl.-Based Syst. 2022, 255, 109674. [Google Scholar] [CrossRef]
  23. Jeevamol, J.; Renumol, V. An ontology-based hybrid e-learning content recommender system for alleviating the cold-start problem. Educ. Inf. Technol. 2021, 26, 4993–5022. [Google Scholar] [CrossRef]
  24. Hao, B.; Yin, H.; Zhang, J.; Li, C.; Chen, H. A Multi-strategy-based Pre-training Method for Cold-start Recommendation. ACM Trans. Inf. Syst. 2023, 41, 31. [Google Scholar] [CrossRef]
  25. Wang, R.; Cheng, H.K.; Jiang, Y.; Lou, J. A novel matrix factorization model for recommendation with LOD-based semantic similarity measure. Expert Syst. Appl. 2019, 123, 70–81. [Google Scholar] [CrossRef]
  26. Zhang, F.; Qi, S.; Liu, Q.; Mao, M.; Zeng, A. Alleviating the data sparsity problem of recommender systems by clustering nodes in bipartite networks. Expert Syst. Appl. 2020, 149, 113346. [Google Scholar] [CrossRef]
  27. Taheri, M.; Farnaghi, M.; Alimohammadi, A.; Moradi, P.; Khoshahval, S. Point-of-interest recommendation using extended random walk with restart on geographical-temporal hybrid tripartite graph. J. Spat. Sci. 2023, 68, 71–89. [Google Scholar] [CrossRef]
  28. Zhu, J.; Li, B.; Wang, J.; Li, D.; Liu, Y.; Zhang, Z. BGCL: Bi-subgraph network based on graph contrastive learning for cold-start QoS prediction. Knowl.-Based Syst. 2023, 263, 110296. [Google Scholar] [CrossRef]
  29. Khababa, G.; Bessou, S.; Seghir, F.; Harun, N.H.; Almazyad, A.S.; Jangir, P.; Mohamed, A.W. Collaborative Filtering Techniques for Predicting Web Service QoS Values in Static and Dynamic Environments: A Systematic and Thorough Analysis. IEEE Access 2025, 13, 45350–45376. [Google Scholar] [CrossRef]
  30. Mezni, H.; Benslimane, D.; Bellatreche, L. Context-Aware Service Recommendation Based on Knowledge Graph Embedding. IEEE Trans. Knowl. Data Eng. 2022, 34, 5225–5238. [Google Scholar] [CrossRef]
  31. Lu, S.; Gu, R.; Jin, H.; Wang, L.; Li, X.; Li, J. QoS-Aware Task Scheduling in Cloud-Edge Environment. IEEE Access 2021, 9, 56496–56505. [Google Scholar] [CrossRef]
  32. Qi, L.; Song, H.; Zhang, X.; Srivastava, G.; Xu, X.; Yu, S. Compatibility-aware web API recommendation for mashup creation via textual description mining. ACM Trans. Multimid. Comput. Commun. Appl. 2021, 17, 20. [Google Scholar] [CrossRef]
  33. SmartSantander: The City of the Future. 2025. Available online: https://smartsantander.eu/ (accessed on 10 January 2025).
  34. Achir, M.; Abdelli, A.; Mokdad, L.; Benothman, J. Service discovery and selection in IoT: A survey and a taxonomy. J. Netw. Comput. Appl. 2022, 200, 103331. [Google Scholar] [CrossRef]
  35. Chen, R.; Guo, J.; Bao, F. Trust management for SOA-based IoT and its application to service composition. IEEE Trans. Serv. Comput. 2014, 9, 482–495. [Google Scholar] [CrossRef]
  36. Zhao, T.; Chen, T.; Sun, Y.; Xu, Y. IM-GNN: Microservice Orchestration Recommendation via Interface-Matched Dependency Graphs and Graph Neural Networks. Symmetry 2025, 17, 525. [Google Scholar] [CrossRef]
  37. Kokoris-Kogias, E.; Voutyras, O.; Varvarigou, T. TRM-SIoT: A scalable hybrid trust & reputation model for the social internet of things. In Proceedings of the 2016 IEEE 21st International Conference on Emerging Technologies and Factory Automation (ETFA), Berlin, Germany, 6–9 September 2016; pp. 1–9. [Google Scholar]
  38. Khani, M.; Wang, Y.; Orgun, M.A.; Zhu, F. Context-aware trustworthy service evaluation in social internet of things. In Proceedings of the Service-Oriented Computing: 16th International Conference, ICSOC 2018, Hangzhou, China, 12–15 November 2018; Proceedings 16. Springer: Berlin/Heidelberg, Germany, 2018; pp. 129–145. [Google Scholar]
Figure 1. System architecture and key behavioral strategies of the DBS-aware recommendation framework.
Figure 1. System architecture and key behavioral strategies of the DBS-aware recommendation framework.
Electronics 14 04468 g001
Figure 2. Sliding time window mechanism and temporal stability judgment.
Figure 2. Sliding time window mechanism and temporal stability judgment.
Electronics 14 04468 g002
Figure 3. Performance Comparison under Varying DBS Ratios. The figure illustrates the Precision, Recall, and F1-Score for both NoDBS and FullDBS variants at 10%, 20%, and 30% DBS levels. FullDBS consistently outperforms across all metrics.
Figure 3. Performance Comparison under Varying DBS Ratios. The figure illustrates the Precision, Recall, and F1-Score for both NoDBS and FullDBS variants at 10%, 20%, and 30% DBS levels. FullDBS consistently outperforms across all metrics.
Electronics 14 04468 g003
Figure 4. Distinct Behavioral Services via Dynamic Pool (Precision, Recall, and F1 Comparison).
Figure 4. Distinct Behavioral Services via Dynamic Pool (Precision, Recall, and F1 Comparison).
Electronics 14 04468 g004
Figure 5. Distinct Behavioral Services via Dynamic User Crowd (Precision, Recall, and F1 Comparison).
Figure 5. Distinct Behavioral Services via Dynamic User Crowd (Precision, Recall, and F1 Comparison).
Electronics 14 04468 g005
Figure 6. Window Size Sensitivity Analysis on DBS Classification.
Figure 6. Window Size Sensitivity Analysis on DBS Classification.
Electronics 14 04468 g006
Figure 7. Top-N recommendation stability over time measured by Jaccard@10.
Figure 7. Top-N recommendation stability over time measured by Jaccard@10.
Electronics 14 04468 g007
Figure 8. Top-N service preservation rate over time measured by Preservation@10.
Figure 8. Top-N service preservation rate over time measured by Preservation@10.
Electronics 14 04468 g008
Figure 9. SCR comparison across different sparsity levels.
Figure 9. SCR comparison across different sparsity levels.
Electronics 14 04468 g009
Figure 10. Cold-Start Scoring Rate (CSR) comparison on three types of DBS: New Member, Isolated, and Ending services.
Figure 10. Cold-Start Scoring Rate (CSR) comparison on three types of DBS: New Member, Isolated, and Ending services.
Electronics 14 04468 g010
Table 1. Comparative analysis of related methods and capabilities. Note: ✓ indicates support, while ✗ indicates the feature is not supported.
Table 1. Comparative analysis of related methods and capabilities. Note: ✓ indicates support, while ✗ indicates the feature is not supported.
MethodApproach CategoryRole ModelingPre-FilteringDynamic AdaptationCold Start HandlingTopology-Aware
DFORM [16]Context-aware
CASR-TSE [4]Context-aware
Hybrid [14]CF
CA-WebAPI [32]CF + KG
WAMTL [17]QoS-aware
TrustQoS-SIoT [18]QoS-aware
OursDBS-aware
Table 2. Key Mathematical Notations.
Table 2. Key Mathematical Notations.
SymbolDescription
RSet of invocation records within the time window. Each record is a tuple ( u i , s j , t k ) indicating user u i invoked service s j at time t k .
[ t Δ , t ] Time window ranging from t Δ to t, used to construct the invocation graph.
G t = ( V t , E t ) Invocation graph for time window t, where V t is the set of services and E t is the set of invocation edges.
w i j Transition frequency from service s i to service s j within E t .
M i j Transition probability from service s i to s j , computed as M i j = w i j k w i k .
M T Transpose of the transition matrix M (used for Markov iteration with column vectors).
v t Column vector representing service scores at iteration t.
ϵ Small initialization constant for unreachable services, ensuring irreducibility of the Markov chain.
b a s e _ s c o r e ( s j ) Base importance score for service s j after convergence.
Table 3. Mapping from Santander City Entities to Cloud Service Concepts.
Table 3. Mapping from Santander City Entities to Cloud Service Concepts.
Original EntityMapped EntityMapped Entity Description
id_deviceid_serviceIdentifier of cloud service instances
device_typeservice_typeService category/type
id_req_applicationid_req_serviceRequested cloud service from the user’s perspective
id_off_applicationid_off_serviceOffered cloud service from the provider side
Table 4. Summary of Evaluation Metrics Used in Validations and Simulations.
Table 4. Summary of Evaluation Metrics Used in Validations and Simulations.
MetricPurposeDefinition/FormulaUsed in
PrecisionMeasures correctness of Top-N recommendation Precision @ N = | Recommended Relevant | | Recommended | Group I
RecallMeasures completeness of Top-N recommendation Recall @ N = | Recommended Relevant | | Relevant | Group I
F1-scoreHarmonic mean of Precision and Recall F 1 @ N = 2 · Precision @ N · Recall @ N Precision @ N + Recall @ N Group I
Jaccard SimilarityMeasures Top-N list stability over time Jaccard @ N = | L t L t τ | | L t L t τ | Group II
Coverage RatioMeasures scoring reach under sparse data Coverage = | Scored Services | | All Services | Group III
Cold-Start Scoring Rate (CSR)Measures scoring success on cold-start services CSR = | Cold - start services with scores | | All cold - start services | Group IV
Table 5. Paired t-test p-values comparing proposed with baselines.
Table 5. Paired t-test p-values comparing proposed with baselines.
BaselineDynamic User CrowdDynamic Pool
GNN 1.88 × 10 10 3.89 × 10 11
TCF 5.04 × 10 10 5.04 × 10 9
TREP 2.26 × 10 12 3.97 × 10 9
MCTSE 6.15 × 10 11 2.53 × 10 12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, S.; Guo, X.; Xue, L.; Dong, Z.; Dong, X. A Robust and Adaptive Service Recommendation Framework for Distinct Behavioral Services in Dynamic Multi-Cloud Environments. Electronics 2025, 14, 4468. https://doi.org/10.3390/electronics14224468

AMA Style

Ma S, Guo X, Xue L, Dong Z, Dong X. A Robust and Adaptive Service Recommendation Framework for Distinct Behavioral Services in Dynamic Multi-Cloud Environments. Electronics. 2025; 14(22):4468. https://doi.org/10.3390/electronics14224468

Chicago/Turabian Style

Ma, Shiyang, Xiaojie Guo, Lingtao Xue, Zesong Dong, and Xuewen Dong. 2025. "A Robust and Adaptive Service Recommendation Framework for Distinct Behavioral Services in Dynamic Multi-Cloud Environments" Electronics 14, no. 22: 4468. https://doi.org/10.3390/electronics14224468

APA Style

Ma, S., Guo, X., Xue, L., Dong, Z., & Dong, X. (2025). A Robust and Adaptive Service Recommendation Framework for Distinct Behavioral Services in Dynamic Multi-Cloud Environments. Electronics, 14(22), 4468. https://doi.org/10.3390/electronics14224468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop