5.2. Experimental Setup and Communication Cost Modeling
In the evaluation of our study, we assessed the performance of our proposed metascheduling algorithm in time-triggered systems, focusing on communication cost as the primary metric. For the genetic algorithm, we utilized Python (version 3.10) along with the DEAP library (version 1.4.3) [
31]. The choice is due to the flexibility in implementing both value- and permutation-encoded genomes [
29]. The detailed configuration of the genetic algorithm used in our experiments is summarized in
Table 1.
We used the Stanford SNAP Library [
32] to generate Directed Acyclic Graphs (DAGs) conforming to application models with job sizes ranging from 10 to 60. These DAGs, created using SNAP’s efficient graph representation capabilities and based on a random model, allowed us to simulate various workload scenarios for metascheduler evaluation. The models varied in nodes, edges, indegrees, and outdegrees to reflect a diverse set of application scenarios. The structural format of the application and platform models used in our experiments is defined using abstract JSON schemas. The application schema captures jobs, messages, and an application deadline, while the platform schema defines nodes and interconnecting links. These schemas are illustrated in
Figure 6.
The use of Directed Acyclic Graphs (DAGs) in scheduling and resource allocation research is well-established and widely supported within the community. Tools such as Task Graphs for Free (TGFF) [
33] and Synchronous Dataflow For Free (SDF3) [
34] were specifically developed to generate and analyze DAG-based workloads, allowing researchers to simulate and evaluate scheduling strategies under controlled and repeatable conditions.
TGFF offers parameterizable task graphs with adjustable structural properties, including degree, critical path length, and communication volume. It is extensively used for benchmarking task allocation and scheduling algorithms in heterogeneous systems. Similarly, SDF3 offers advanced modeling capabilities for synchronous dataflow applications, supporting the generation of consistent, deadlock-free graphs that closely resemble digital signal processing (DSP) and multimedia workloads.
These tools, along with the underlying DAG abstraction, serve as standard baselines in the design and evaluation of schedulers for embedded, multiprocessor, and MPSoC platforms. Therefore, our use of DAGs generated through the SNAP library aligns with these best practices, enabling comparative analysis across a broad design space.
We applied the algorithm to a range of datasets, ensuring a consistent platform model across all experimental runs to isolate performance variations to application model differences. Each dataset, encompassing diverse job sizes and structures generated via the SNAP library, was processed by the algorithm, enabling an evaluation across varying complexities and scenarios.
In our study, the computation of communication cost plays a crucial role in evaluating our proposed metascheduling algorithm. This cost is fundamentally associated with the propagation of context information among end systems (ES) within the network. To accurately compute this cost, we adopted the following approach:
- (a)
We categorized end systems requiring context information as ’recipients’. This categorization is based on Equations (
1) and (
2).
- (b)
We incorporated a dedicated time slot for the exchange of context information prior to the commencement of task scheduling on the end systems. Following the occurrence of a context event, we established the current time for each affected end system (denoted as ).
We then reserved time slots following the initiation of a broadcast of context events to all recipient end systems. This process included the assumption of a specified delay, denoted as
, which varies based on the path index of each context message, reflecting the network’s topology and the data flow paths. The overall communication cost associated with each scheduling decision is given as
Equation (
3) captures the cumulative path delays (
) based on routing decisions made during schedule construction. Other sources of communication latency, such as serialization time or software-induced processing delays, are not explicitly modeled. We assume fixed-size messages and static link characteristics, where serialization is either negligible or embedded within the path delay values used in the model. Moreover, such delays remain constant across all schedules and are not influenced by the optimization variables. As the genetic algorithm ranks individuals by relative makespan, including such constants would not change the selection process and can therefore be safely omitted to simplify the model.
The value of (
) in Equation (
3) is derived from the platform model using a set of precomputed multicast paths. These paths are generated by solving the k-shortest-path problem between relevant end-system pairs. For each message, the genetic algorithm selects one of these precomputed paths through its path index
P genome. The associated path delay
is calculated as the number of hops (i.e., the total number of links and switches traversed) along the selected path. This value is retrieved from the platform data structure during schedule reconstruction and used to compute the message delay contribution in the makespan.
The communication cost for periodic context information exchange, denoted as
, is mathematically expressed in Equation (
4). It is proportional to the ratio of the makespan
M over the sampling frequency
, scaled by the context information cost
. The makespan
M represents the total time to complete tasks, while
is the rate of system data collection, and
covers the resources required to process the sampled data.
5.3. Comparative Evaluation with Periodic Sampling Approaches
To benchmark our algorithm against existing solutions, we conducted a comparison focusing on the prevalent approach of periodic sampling for context information exchange, a standard in the current state of the art. Our methodology for this comparison involved the following steps:
We simulated the conventional approach by implementing periodic sampling across a range of application profiles. Sampling points were calculated at intervals varying from 2 to 10 time units (represented as 2TU to 10TU), reflective of typical practices observed in existing solutions. By sampling points, we refer to the periodic exchange of context information. The range (2TU to 10TU) is chosen to encompass a broad spectrum of scenarios, from more frequent to less frequent information exchanges, providing a robust basis for comparison. We describe the context information utilized in the test as a tuple. The elements of the tuple are as follows: (, , and ).
Each application profile was analyzed under both our proposed algorithm and the traditional periodic sampling method. The primary focus of this comparative analysis was to assess the impact on two crucial metrics: communication cost and makespan (the total time taken to complete all jobs). By evaluating these metrics, we aimed to understand not only the efficiency of our algorithm in reducing communication overhead but also its effectiveness in optimizing overall system performance.
Table 2 presents the total communication costs associated with different context exchange methods as the number of jobs increases for each MSG. Our proposed algorithm exhibits the most efficient performance, with a gradual increase in communication cost, ranging from 490 time units for 10 jobs (10J) to 147,420 time units for 60 jobs (60J). In contrast, sampling every 2 time units incurs the highest costs, especially noticeable at higher job counts (e.g., 4,862,588 time units for 60 jobs). As the sampling interval increases (from 2 TU to 10 TU), there is a consistent reduction in communication cost, demonstrating the impact of less frequent context exchanges in reducing communication overhead. However, sampling at a less frequent interval poses the challenge of omitting responses to context events. Compared to the 2TU sampling method, the proposed algorithm reduces communication overhead by significant factors across different job sizes, as shown in
Table 2 below.
The plot in
Figure 7 compares the communication costs associated with different scheduling algorithms as the number of jobs increases. Using a logarithmic scale for the y-axis (Total Communication Cost in Time Units, TU), the plot demonstrates a clear trend: as the number of jobs grows, the communication cost escalates in all cases. The data illustrate that our proposed algorithm (blue) consistently incurs lower communication costs compared to other sampling intervals, particularly at higher job counts, thus emphasizing its efficiency.
Figure 8 illustrates the observed makespan for the previous example with and without the application of our algorithm. The values in time units are given in
Table 3. It can be observed in
Figure 8 that the makespan when our proposed algorithm is implemented (blue plot) is slightly increased compared to when the algorithm is not applied (the case of no context exchange (red plot)). It is important to note that the makespan in the case “No Context Exchange” does not include the cost of sampling the exchange of context information. However, in a scenario where the communication cost associated with different sampling frequencies is included, the makespan can be impacted, as shown in the results obtained in
Table 4, if the same communication channel is used for exchanging context information. Nevertheless, if a dedicated communication infrastructure is used for exchanging context information, similar communication costs will still be incurred on the dedicated network, which includes extra hardware implementation for its realization.
Table 3 and
Table 4 provide a broader comparison of makespan outcomes under various context exchange strategies. The periodic sampling methods reveal a consistent pattern: shorter intervals, such as 2TU, introduce substantial communication overhead, leading to inflated makespan values. Conversely, longer intervals, such as 10TU, lessen this overhead, resulting in comparatively lower makespans. Nonetheless, these periodic strategies still underperform when contrasted with our proposed algorithm, which seamlessly incorporates context information exchange during scheduling. For example, in the case of 30 jobs, the proposed method completes scheduling in 210,652 time units, outperforming even the most efficient periodic method (10TU) by a considerable margin. While the no-context-exchange setup yields the lowest makespan (e.g., 177,685 for 30 jobs), it entirely omits adaptivity, rendering it unsuitable for dynamic environments. These comparative results underscore the practical advantage of our algorithm in managing communication overhead while maintaining system responsiveness.
5.4. Statistical Consistency Across GA Executions
To assess the consistency of the genetic algorithm (GA) under repeated executions, we conducted a statistical stability analysis using a 10-job application. The GA was executed 100 times independently for both the proposed method and a benchmark configuration without context exchange. For each run, the resulting makespan was recorded, and key statistical metrics were computed, including the mean, standard deviation, variance, and the coefficient of variation (CoV).
The CoV, defined as the ratio of standard deviation to the mean, serves as a normalized measure of dispersion. As shown in
Table 5, the makespan CoV for the 10-job for our proposed method and the benchmark case remained well below 7%, indicating that the GA consistently converges to similar solutions across runs. This analysis exemplifies the statistical stability of the proposed method and supports its applicability in reliable and repeatable scheduling scenarios.
5.5. Real-World Benchmark Implementation
To support the evaluation of our scheduling method, we used an application model based on the automotive benchmark described by Kramer et al. [
35]. This benchmark, derived from real engine control software, provides detailed information about execution times, communication sizes, and task dependencies. We selected and adapted only the attributes that are directly relevant to our scheduler.
Each job in the application represents a schedulable software unit. The key attributes are
execution_time(): This represents the worst-case execution time (WCET) of each job, in microseconds. We generated these values using a Weibull distribution to reflect runtime variability. For each job, we selected a period group (e.g., 10 ms, 20 ms) and used the corresponding minimum, average, and maximum execution times from Table IV in [
35]. The distribution was scaled to match the average, and the values were constrained to remain within the reported bounds.
core_node: This field indicates which processing core each job is mapped to. In our method, this mapping is not predefined but is determined as part of the genetic algorithm. The scheduler explores different job-to-core assignments to minimize makespan and reduce contention. While the original benchmark assumes fixed task-to-core mappings for multi-core execution platforms (as discussed in [
35]), we reinterpret core-level mapping as an optimization parameter targeting an integrated MPSoC architecture.
communications: These define data dependencies between jobs. Each message includes:
The system runs on a 2 × 2 MPSoC mesh with four processing cores connected via routers. While the original benchmark targets distributed ECUs connected via a CAN bus, we adapted the workload to run on a tiled MPSoC platform. This allows us to evaluate mapping and scheduling decisions in the context of on-chip multi-core communication and resource contention.
As shown in
Table 6, the metascheduler employing selective context exchange (proposed algorithm) achieves the most favorable timing performance. Its makespan of 7724
is markedly lower than that of any periodic-sampling alternative and is only marginally higher than the “No Context Exchange’’ reference, whose result ( 6853
) is interesting yet practically infeasible because it violates the consistency guarantees mandated by functional-safety standards. Periodic sampling with a 2 TU interval increases the makespan to 23,986
, a degradation of approximately 211 relative to the proposed approach. Even at the longest interval considered ( 10 TU), the makespan (10,279
) remains 33% larger than that obtained with selective exchange.
Table 7 clarifies the origin of these timing effects. The selective strategy transmits only 880 byte of context data, whereas 2 TU periodic sampling injects 17,132 byte, representing an increase by a factor of nearly 20. Although longer sampling intervals reduce traffic volumes, they never approach the communication efficiency of the proposed method. This demonstrates that the selective context-exchange mechanism effectively balances communication cost and timing performance, offering a more scalable and resource-efficient solution compared to fixed-interval policies.
5.6. Benchmark Against State-of-the-Art Adaptation Techniques
We recognize the importance of benchmarking against recent adaptive metaschedulers, such as the discrete particle swarm optimization (DPSO)-based method mentioned in [
36] and the genetic algorithm (GA)-based metascheduler in [
37]. However, a direct quantitative comparison is currently not feasible due to the unavailability of their source code, which is not publicly accessible. Additionally, these implementations are specifically tailored to niche domains, namely wireless IoT-based wireless sensor networks (IoT-WSNs) and network-on-chip (NoC)-based multi-processor system-on-chips (MPSoCs), each with unique architectural assumptions, fault models, and evaluation metrics.
Despite this limitation, a qualitative comparison reveals several key differences. The metascheduler proposed in [
36] concentrates on fault recovery in the event of single and double component failures, utilizing DPSO to precompute schedule graphs. While effective, this method treats adaptation and context synchronization as distinct layers and relies on periodic task re-allocation. Conversely, our approach seamlessly integrates timing-aware context communication into the metascheduling process. This integration allows for proactive and selective synchronization that minimizes communication overhead and preserves system-wide determinism.
Similarly, the approach in [
37] constructs multi-schedule graphs using a GA and implements path reconvergence to address state-space explosion. However, it does not consider context propagation within the scheduling loop. In contrast, our algorithm jointly optimizes both schedule construction and the selective transfer of context information, creating a communication-efficient adaptation framework. Although we present a simplified “No Context Exchange” scenario to illustrate a theoretical lower bound, the results demonstrate that our approach offers structural and functional advantages over those in [
36,
37], thereby supporting our primary claims of efficiency and predictability.