Next Article in Journal
High-Speed Video System for Micro-Expression Detection and Recognition
Next Article in Special Issue
A Police and Insurance Joint Management System Based on High Precision BDS/GPS Positioning
Previous Article in Journal
Centralized Networks to Generate Human Body Motions
Previous Article in Special Issue
Obstacle Detection as a Safety Alert in Augmented Reality Models by the Use of Deep Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Energy Efficient Real-Time Scheduling Using DPM on Mobile Sensors with a Uniform Multi-Cores

1
Da Vind College of Genneral Education, Chung-Ang University, Heuksuk-ro 84, Dongjak-gu, Seoul 06974, Korea
2
Department of Computer Science and Engineering, Chung-Ang University, Heuksuk-ro 84, Dongjak-gu, Seoul 06974, Korea
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(12), 2906; https://doi.org/10.3390/s17122906
Submission received: 31 October 2017 / Revised: 1 December 2017 / Accepted: 12 December 2017 / Published: 14 December 2017
(This article belongs to the Special Issue Mobile Sensing Applications)

Abstract

:
In wireless sensor networks (WSNs), sensor nodes are deployed for collecting and analyzing data. These nodes use limited energy batteries for easy deployment and low cost. The use of limited energy batteries is closely related to the lifetime of the sensor nodes when using wireless sensor networks. Efficient-energy management is important to extending the lifetime of the sensor nodes. Most effort for improving power efficiency in tiny sensor nodes has focused mainly on reducing the power consumed during data transmission. However, recent emergence of sensor nodes equipped with multi-cores strongly requires attention to be given to the problem of reducing power consumption in multi-cores. In this paper, we propose an energy efficient scheduling method for sensor nodes supporting a uniform multi-cores. We extend the proposed T-Ler plane based scheduling for global optimal scheduling of a uniform multi-cores and multi-processors to enable power management using dynamic power management. In the proposed approach, processor selection for a scheduling and mapping method between the tasks and processors is proposed to efficiently utilize dynamic power management. Experiments show the effectiveness of the proposed approach compared to other existing methods.

1. Introduction

WSNs consist of a number of moblile sensor nodes which are tiny, multi-functional, and low-power. Table 1 lists mobile sensing platforms with various sensors. It is widely used in various applications to collect and process data, such as various types of physical and environment information. Recently, sensor nodes in WSNs have evolved for multimedia streaming and image processing. In response to these high performance demands, sensor nodes with multi-processors have emerged. A multi-processor sensor node platform, mPlatform, which is capable of parallel processing for computationally intensive signal processing, was proposed by Lymberopoulos et al. [1]. These platforms operate with limited batteries, as shown in Table 1. The use of a multi-cores in the sensor node makes energy consumption more serious. Power management among sensor nodes is of critical importance for several reasons: limited energy batteries and ensuring longevity [2,3,4], meeting performance requirements [2,5,6], inefficiency arising because of over provisioning resources [2], power challenges posed by CMOS scaling [2,7], and enabling green computing [2]. Recent advances in CMOS technology have improved the density and speeds for on-chip transistors. These trends limit the fraction of chips that can be used at maximum speeds within limited power. Therefore, power challenges in CMOS have been addressed for processor performance. Transistor performance scaling in the future may end if left unaddressed [8,9]. Battery-operated embedded systems are sensitive to high power consumption, which leads to heating and reduced battery lifetime. Thus, energy-efficient management is essential in several embedded systems such as wearable devices. Improving energy efficiency leads to scale performance without violating the power budget. In recent advances, mobile computing devices with a multi-cores have been dynamically increased for mobile convergence applications (e.g., video streaming and web browsing). Power management in embedded systems contributes to achieve nearly 3% of the overall carbon footprint in green computing [10]. Energy efficiency scheduling algorithms for the sensor node with multi-processors are necessary. The scheduling algorithms must be able to keep battery lifetime longer while meeting the time constraints.
The Asymmetric Multi-core Platform (AMP) is capable of parallelism with different performance levels. The examples of AMP include mobile phones, tablets, and high-end mobile sensor nodes. These devices are equipped with cores capable of handling tasks requiring high-performance processing. Note that not all tasks need the high-performance processing and energy efficient schemes are adopted even for the cores consuming low power. The problem of scheduling AMP for high-performance mobile sensors is important in terms of performance and energy efficiency. The scheduler can switch the high-performance cores to a low power state by assigning tasks to the low power cores when processing the tasks requiring low loads. It is also possible that powerful cores are changed into simpler cores to adapt the system to varying loads. ARM’s big.LITTLE architecture [17,18] is a representative example. In ARM’s big.LITTLE architecture, there are three modes for task migration: cluster migration, CPU migration, and global task scheduling. The scheduler improves energy efficiency by migrating tasks between big and little cores. In this paper, we discuss real-time scheduling problems in the context of AMP. We adopt a scheme using the T-Ler plane to develop energy-efficient scheduling algorithms for real-time tasks on uniform multi-core systems.
The T-Ler plane extends the T-L plane using a T-L abstraction strategy to fit uniform multi-core systems. The Voltage Frequency Scale (VFS) is exploited on energy-efficient scheduling algorithms using the T-L plane. On the other hand, there are not many studies related to Dynamic Power Management (DPM). Sensor network applications with varying loads depending on the situation can take advantage of the energy by switching the state of unnecessarily used processors. Kim et al. [19,20] proposed several T-L plane based energy-efficient algorithms using DPM for sensor nodes with identical multi-processors. However, these algorithms are not suitable for uniform multi-processor systems. In particular, it is hard to select the set of processors with the lowest power consumption among the multiple sets of processors that have the same capacity. We propose a new algorithm suitable for sensor node with uniform multi-processors, called Uniform-DPM. More specifically, we extend the previous approaches [19,20] by considering the characteristics of uniform multi-processors in terms of energy efficiency as follows:
  • At the beginning of each T-Ler plane, select the processors operating with a low frequency and minimize the processing capacity as much as possible.
  • Reduce the complexity of scheduling and fragments of idle time, and classify the processors and tasks into processor sets and task sets at the beginning of the T-Ler plane, respectively.
  • At each event in the T-Ler plane, utilize constrained migration to reduce the complexity of scheduling and fragments of idle time.
The first extension is to reduce the power loss caused by uniform multi-processors that consist of processors with difference processing capacities. The previous approach [20], as shown in Section 2, focuses solely on minimizing the number of processors. It is not suitable for uniform multi-processors. In the case of uniform multi-processors, the processors must be selected considering the processing capacity and the frequency of each processor. The second extension is to classify processors and tasks for limited scheduling, where tasks in a set are only scheduled to processors in the according processor set. The third extension is to adjust the sets in each event and to assign tasks to the processors using the limited scheduling. These prevent the unnecessary migration of tasks and enables the collection of idle time on particular processors.
We organize this paper as follows. In Section 2, we introduce related works, including the approaches previously based on T-L plane targeting uniform multi-processors. In Section 3, we propose mechanisms to select processors and allocate tasks for energy-efficient scheduling in uniform multi-processors. We extend the proposed events in identical multi-processors to ones in uniform multi-processors. In Section 4, we perform experimental evaluations by comparing our proposed algorithms with other algorithms. Lastly, we present the conclusions and future works in Section 5.

2. Related works

2.1. Power Management Techniques

Due to the advancements in semiconductor process technologies, there have been more high-end processors available that integrate more transistors. Recently, real-time embedded systems have been increasingly adopting high-end processors. In addition, to improve the performance, real-time embedded systems are also adopting multi-processors. However, this increases the processor power consumption significantly. The power consumption of CMOS chips is as follows [21]:
P t o t a l = P s t a t i c + P d y n a m i c .
P s t a t i c is the static power consumption which is calculated as the sum of the leakage power and short current power. P d y n a m i c is the dynamic power consumption by charging and discharging of the output capacitance for processing time. It is not easy to reduce the static power consumption which depends on various parameters in the semiconductor process. Therefore, we focus on reducing the dynamic power consumption. Dynamic power is defined as:
P d y n a m i c = α C V 2 f ,
where f is the frequency, α is the switching activity factor, V is the supply voltage, and C is the capacitive load. DVFS is a method used to adjust the supply voltage and frequency of a CMOS chip by utilizing the slack time that occurs when scheduling tasks. On the other hand, DPM is a method of reducing energy consumption by switching to a low power state when slack time occurs. However, if a sufficient slack time is not guaranteed over the break-even time, the energy overhead caused by the state transition will cause loss. The break-even time B E T s l e e p is determined by Equation (3) [22].
B E T s l e e p = max ( t s w , E s w P s l e e p · t s w P i d l e P s l e e p )
The transition energy overhead and recovery time are denoted as E s w and t s w , respectively. P i d l e denotes the idle power. The sleep power is denoted by P s l e e p . The break-even time should be considered when developing a scheduling algorithm that not only uses the sleep mode, but also guarantees real-time responsiveness.

2.2. Global Scheduling Approaches on Multi-Processors

Scheduling disciplines can be categorized by considering the complexity of the priority mechanisms and the degree of job migration. Considering how task priorities are determined, Carpenter et al. [23] have categorized the schemes to static, dynamic but fixed within a job, or fully dynamic.
  • Static: A single fixed priority is applied to all jobs for each task in the system. e.g., Rate Monotonic (RM) scheduling.
  • Dynamic but fixed within a job: Different priorities may be assigned for the jobs of a task, but a job has a fixed priority at different times. e.g., Earliest Deadline First (EDF) scheduling.
  • Fully dynamic: Different priorities may be assigned for a single job at different times, e.g., Least Laxity First (LLF) scheduling.
Depending on the degree of job migration, Carpenter et al. [23] have categorized the migration criterion to no migration, restricted migration, and unrestricted migration.
  • No migration: The set of tasks in the system is partitioned into some subsets for available processors, a scheduler schedules a subset on a unique processor. The jobs of a task in a subset are executed on the corresponding processor.
  • Restricted migration: Each job of a task must be scheduled entirely on a single processor. However, other jobs of the same task may be executed on different processors. Therefore, migrations among processors are allowed at the task-level context, but not at job boundaries.
  • Unrestricted migration: Any jobs is also allowed to migrate among processors during its lifetime.
Note that our proposed scheduling algorithm supports fully dynamic and unrestricted migration.
Various global scheduling algorithms for multi-processors have been studied. In global scheduling, all eligible jobs waiting for execution are in a single priority-ordered queue shared by all of the processors in the system; the highest priority job is dispatched from this queue by the global scheduler. Most of early the studies on global scheduling extended optimal scheduling algorithms known well for a single processor, such as RM and EDF, to multi-processors. However, these extensions can result in wasted utilization of resources. The fluid scheduling model with fairness notion, where each task is always executed at a fixed rate, emerged to overcome the limitation [24]. Figure 1 compares the fluid scheduling concept and the practical scheduling. There is a gap between fluid scheduling and practical scheduling, as shown in Figure 1. There are some algorithms extending the fluid scheduling model for achieving optimality on multi-processors. Proportionate fair (P-fair) scheduling has produced a feasible schedule for periodic tasks on multi-processors, and it has shown considerable promise in multi-processor scheduling [25]. However, extensive amount of migrations and preemption are needed to follow the fluid schedule. Much effort has been made to overcome this problem in global optimal scheduling. Thereafter, Deadline Partitioning-fair (DP-fair) and Deadline Partitioning-warp (DP-wrap) algorithms were proposed, and they exhibited better performance with respect to preemption in [26]. The method of allocating tasks to the processors supported by these scheduling algorithms is not suitable for uniform multi-processors. Cho et al. [27] proposed Largest Nodal Remaining Execution-time First (LNREF) using a T-L plane abstraction and it performs well with uniform multi-processors. Funk and Meka [28] proposed a T-L plane based scheduling algorithm, U-LLREF, that extends LNREF algorithm for uniform parallel machines. In U-LLREF, a uniform multi-processors provides a condition for determining event-c. Chen et al. [29] proposed Precaution Cut Greedy (PCG), a T-L plane based scheduling algorithm for uniform multi-processors. PCG uses a modified T-L plane, a T-Ler plane. Figure 2 shows how the PCG schedules in the first T-Ler plane. When event-c occurs, τ 3 is assigned to p 2 until the end of the T-Ler plane. Thus, in PCG, a task monopolizes a single processor, thereby preventing unnecessary task migration.

2.3. T-L Plane Based Energy-Efficient Global Optimal Scheduling Approaches

Energy-efficient scheduling based on the T-L plane for uniform parallel machines has been proposed due to the demand for energy efficiency. Uniform RT-SVFS [30] reduces the energy consumption by scaling the frequency of all processors with a constant rate. By scaling the height of the T-L plane, as shown in Figure 3, scheduling is enabled at the changed frequency. α k represents the normalized frequency of the processor. In addition, energy-efficient T-L plane based scheduling algorithms for unrelated parallel machines have emerged. Independent RT-SVFS [30] determines the frequency by statically scaling each processor. This algorithm has been proposed to overcome the heavy task bottlenecks that can occur when using the frequency scaling technique. The Growing Minimum Frequency (GMF) [31], which is a state-of-the-art algorithm for T-L plane based non-uniform frequency scaling for saving energy on VFS embedded multi-processors, has been proposed for the frequency control of multi-processors using U-LLREF, and the global optimal frequency can be determined. RT-DVFS [32] allows you to dynamically adjust the frequency of each processor in the event of scheduling.
It is difficult to consider DPM due to the idle time fragmentation problem that occurs when using the T-L plane based scheduling algorithm. In addition, since scheduling is performed using all processors existing in the system, a considerable energy overhead due to unnecessary state transition occurs when DPM is used. TL-DPM [19] solves the idle time fragmentation problem by using a new event to retrieve tokens, which is performed in the next plane. However, since only the token of the next plane is targeted, there is room for solving the idle time fragmentation problem. Kim et al. [20] proposed a generalized method for executing tokens to be scheduled in the later plane in the current plane in order to solve this problem. To reduce the number of state transitions, scheduling is performed using only the minimum number of processors.

3. Proposed Energy Efficient Approach on Uniform Multi-Processors

3.1. Feasibility Conditions

Theorems 1 and 2 represent the conditions that must be met to obtain schedules satisfying the time constraints when uniform multi-processors are used for scheduling the given task set.
Theorem 1.
(Horvath et al. [33]) The level algorithm constructs a minimal length schedule for the set of independent tasks with service requirements e 1 e 2 e n on the processing system π = ( c 1 c 2 c m ) , where m n . The schedule length is given by
max { max 1 i m ( e i c i ) , e n c m } .
Theorem 2.
(Funk et al. [34]) Consider a set τ = { τ 1 , , τ n } of periodic tasks indexed according to non-increasing utilization (i.e., u i u i + 1 for all i, 1 i n , where u i = e i / p i ). Let U i = j = 1 i u i for all i, 1 i n . Let π denote a system of m n uniform processors with capacities c 1 , c 2 , …, c m , c i c i + 1 for all i, 1 i m . Periodic task system τ can be scheduled to meet all deadlines on the uniform multi-processor platform π if and only if the following constraints hold:
U n c m ,
where U k c k , for all k = 1 , 2 , , m .
Selecting processors for the scheduling tasks at the beginning of each T-L plane is divided into the case where tasks are allocated to the processors with the same capacity as the utilization, and the case where they are not.

3.2. Processor Selections and Classification

3.2.1. Simple Case: Exact Match

Table 2 shows some examples of processor selections for scheduling the task set shown in Table 3. In a CMOS chip, power consumption is determined by the operating frequency and supply voltage. The relationship between the power consumption and the supply voltage in the processor is as follows.
P d y n a m i c V 2 .
In addition, according to the relationship between the supply voltage and the operating frequency in a processor, shown in Equation (7), a processor operating at a higher frequency requires a higher supply voltage than that operating at a lower frequency. Therefore, as shown in Table 2, a processor with a higher supply voltage will have a higher capacity.
f ( V V t h ) β V ,
where V t h is the threshold voltage of transistors and β is a measure of the velocity saturation in COMS transistors.
S 1 , S 2 , and S 3 shown in Table 2 satisfy Theorems 1 and 2 presented above. Since the processing capacity of S 1 , S 2 , and S 3 is equal to the total utilization of the task set, there is no idle time when the task set is scheduled. However, since the number and capacity of processors is not the same in each processor, the power consumed by S 1 , S 2 , and S 3 is different. The energy consumption for scheduling the task set in Table 3 on S 1 , S 2 , and S 3 is shown in Table 4. The lowest power consumption can be observed on S 2 , where each task is independently assigned to a processor whose capacity is equal to the utilization of each task in the task set. If the total capacity of a processor set is equal to the total utilization of a task set, then there is no idle time because all processors always perform their tasks. Therefore, the power consumption of each processor is dependent on the processed workload. High-capacity processors can handle more work in terms of processor workloads. Equation (8) shows the power consumption E e ( V k ) needed to process e i in a processor whose operating frequency and supply voltage are f k and V k , respectively.
E e ( V k ) = α C V k 2 f k ( e i f k ) = α C V k 2 e i ,
where V k and f k denote the voltage and capacity of the k-th processor, respectively. Lemma 1 shows the power consumption characteristics of processor sets whose total capacity is equal to the total utilization of a task set.
Lemma 1.
If U t o t a l = c n = c i + c j , when scheduling a task set with U t o t a l on two processor sets S 1 = { c n } and S 2 = { c i , c j } , the power consumption satisfies α C V n 2 e n > α C V i 2 e i + α C V j 2 e j .
Proof of Lemma 1.
According to Equation (8), the power consumption measures of S 1 and S 2 are α C V n 2 e n and α C V n 2 e i + α C V i 2 e j respectively. Since c n = U t o t a l and c i + c j = U t o t a l , there is no idle time when the tasks are scheduled. In addition, V n 2 e n = V n 2 ( e i + e j ) > V i 2 e i + V j 2 e j , where e n = e i + e j and V n > V i , V j . Hence, α C V n 2 e n > α C V i 2 e i + α C V j 2 e j . ☐
According to Lemma 1, selecting the 0.8 capacity processor for scheduling the 0.6 and 0.2 utilization tasks in the task set, as shown in Table 1, will result in higher power consumption than selecting the 0.6 and 0.2 capacity processors for the scheduling the tasks. Therefore, assigning each task to the processor sets whose capacity is equal to its utilization is the most energy-efficient way when there are enough processors. Under the condition of c i u i , the processor whose capacity is equal to u i shows the lowest power consumption to execute the task with the utilization u i . Lemma 2 shows these characteristics.
Lemma 2.
When a task with utilization u i is executed on two processors under the condition of c n > c j = u i , their power consumption for processing the allocated workload during the task period is E e ( V n ) > E e ( V j ) .
Proof of Lemma 2.
When two processors with capacities of c n and c j perform the workload e i during the period p i , their power consumption measures are E e ( V n ) and E e ( V j ) respectively. If c n > c j , V n > V j is satisfied by Equation (7). Hence, E e ( V n ) > E e ( V j ) is satisfied by Equation (8). ☐
When a task set with the total time of i = 1 n e i is scheduled on n processors whose capacity is different, the power consumption required for processing the allocated workload on n processors is shown in Equation (9). e 1 , e 2 , …, e n represents the workload assigned to each processor.
E e = α C V 1 2 f 1 ( e 1 f 1 ) + α C V 2 2 f 2 ( e 2 f 2 ) + + α C V n 2 f n ( e n f n ) = α C V 1 2 e 1 + α C V 2 2 e 2 + + α C V n 2 e n = i = 1 n E e ( V i ) .
If the total capacity of n processors is greater than the total utilization of a task set to be scheduled, there will be idle time during task scheduling. This means that the power consumption during the idle time should be taken into account to measure the processors’ power consumption required for scheduling the task set. The power consumption of n processors is shown in Equation (10). α i denotes the power consumption of the i-th processor during the idle time.
E d = i = 1 n ( E e ( V i ) + α i ) .
Lemma 3 and Theorem 3 show the power consumption required for scheduling a task set on a set of n processors with different capacities.
Lemma 3.
When the task set is scheduled with the processor set S 1 , the lowest power is consumed, where the total capacity of S 1 is τ i τ u i = p i S 1 c i .
Proof of Lemma 3.
If τ i τ u i = p i S 1 c i , scheduling involves no idle time, so the processor power consumption is p i S 1 E e ( V i ) . If τ i τ u i < p i S 1 c i , scheduling involves some idle time, so the processor power consumption based on Equation (10) is p i S 1 ( E e ( V i ) + α i ) . Hence, if τ i τ u i = p i S 1 c i , then the lowest power consumption will be observed. ☐
Theorem 3.
Independently assigning each task in the task set τ to processors whose capacity is equal to the utilization of the task u i shows the lowest power consumption for scheduling a set of tasks.
Proof of Theorem 3.
This is easily proven by Lemmas 1–3. ☐
Therefore, selecting processors whose capacity is equal to the utilization of each task shows the lowest power consumption for scheduling a set of tasks.

3.2.2. Generalized Solution

When not assignable to a processor with the same capacity as the task’s utilization, it is necessary to select a processor set available for scheduling with the limited processors. Table 5 shows the characteristics of processors used for task scheduling. Table 6 shows the processor sets selected from the processors in Table 5 for scheduling the task set shown in Table 7.
Since the processor sets S 1 , S 2 , S 3 , and S 6 shown in Table 6 satisfy Theorems 1 and 2, they can be used for task scheduling. However, since the processor sets are differently configured, the idle time during the task scheduling and the difference in their supply voltages result in their different power consumption. Therefore, the following two strategies should be considered to select energy-efficient processors.
Selecting a processor set for task scheduling in consideration of all the problems presented above is a NP-hard problem. Therefore, in this paper, we propose a heuristic method for selecting an energy-efficient processor set. In the proposed method, if the size of the current plane is smaller than C s l e e p , the processor in active mode is added to a processor set for task scheduling because it cannot be switched to sleep mode at the end of the previous plane. If the preferentially selected processors are not enough for scheduling the given task set, additional processors will be selected. Processors for scheduling are selected in terms of the local utilization of tasks from highest to lowest. Selecting processors for scheduling depends on the difference between the total local utilization of tasks in τ r e a d y at the start time t 0 in each plane τ j τ r e a d y r j ( t 0 ) and the total capacity p j P s e l e c t e d c j of the processors in P s e l e c t e d . The selected processors are moved to P s e l e c t e d . The following describes how to select processors.
  • If 0 τ j τ r e a d y r j ( t j ) p j P s e l e c t e d c j < r i ( t 0 ) , the processor with the smallest capacity is selected for scheduling in the given processor set, { p j | c j r i ( t 0 ) τ j τ r e a d y r j ( t 0 ) p j P s e l e c t e d c j   w h e r e   p j { P a l l P s e l e c t e d } } .
  • If τ j τ r e a d y r j ( t 0 ) p j P s e l e c t e d c j r i ( t 0 ) , the previously selected processor is used for scheduling without selecting an additional processor.
P a l l is the set of all the processors in the system. P s e l e c t e d is the set of the selected processors for task scheduling. Algorithm 1 shows how to select processors for scheduling at the beginning of each plane. The function getMinimumCapacityProcessor(availableCapacity, τ, P t e m p ) takes the available capacity (availableCapacity) of the previously selected processor into account to return the lowest capacity processor for scheduling the task set τ from the given processor set P t e m p . The function add() adds elements to the set, and the function erase() removes elements from the set. The processors in P s l e e p indicate the processor in the sleep state in the plane. It is necessary to ensure a break-even time longer than the idle time in order to use DPM techniques for switching the state of a processor. To ensure the idle timeis long enough to enter the sleep mode, the idle time in the plane should be generated as much as possible on a single processor. To prevent unnecessary power consumption, a task is assigned to the selected processor whose capacity is the lowest for scheduling the task. For this reason, in the proposed method, the processors in the selected processor set are classified into the following categories: processors that can be used to the maximum extent in the plane and processors that can be used exclusively by a single task in the plane. That is, the processors in P s e l e c t e d are classified into the following categories: P f i x e d , P m a x , and P s l a c k . The processors in P f i x e d represent a set of processors exclusively used by a single task. P m a x is the set of processors used to the maximum extent in the plane. P s l a c k is the set of processors that may result in idle time in the plane during task scheduling. The tasks to be executed on the classified processor sets are divided into the following categories: τ f i x e d , τ m a x , and τ s l a c k . Tasks assigned to a processor set cannot be moved to another processor set. The following describes how to classify the processor sets.
Algorithm 1 Processor selection of the beginning time of a T-L plane
  1:
Input: P a l l , P s l e e p , τ a l l , p s i z e
  2:
Output: P s e l e c t e d , τ r e a d y
  3:
psize—Size of the T-L plane
  4:
P a l l —The set of processors in the system
  5:
P s l e e p —The set of processors to be sleep mode
  6:
P s e l e c t e d —The set of processors selected for scheduling tasks
  7:
P t e m p —The temporary set of processors
  8:
τ a l l —The set of all tasks in the system
  9:
τ r e a d y —The set of ready tasks
10:
τ —Temporary variable for tasks
11:
p—Temporary variable for processors
12:
availableCpacity—Temporary variable for available capacity
13:
for p P a l l P s l e e p do
14:
    if p s i z e < p . C s l e e p then
15:
        add( p , P t e m p );
16:
    end if
17:
end for
18:
for τ τ a l l do
19:
    if τ . e > 0 then
20:
        add( τ , τ r e a d y );
21:
    end if
22:
end for
23:
repeat
24:
     τ = getFirstLocalUtilizationTask( τ r e a d y );
25:
    availableCapacity = p i P s e l e c t e d p i . c τ i τ r e a d y τ i . r ( t 0 )
26:
    if availableCapacity τ . r ( t 0 ) then
27:
        contitue;
28:
    else
29:
        p = getMinimumCapacityProcessor( a v a i l a b l e C a p a c i t y , τ , P t e m p );
30:
        if p is null then
31:
           p = getMinimumCapacityProcessor( a v a i l a b l e C a p a c i t y , τ , P s l e e p );
32:
           if p . c > p 1 . c then
33:
               erase( p , P s l e e p )
34:
           end if
35:
           erase( p , P s l e e p );
36:
        else
37:
           erase( p , P t e m p );
38:
        end if
39:
        add( p , P s e l e c t e d );
40:
         p 1 = p ;
41:
    end if
42:
until τ is not null
43:
return P s e l e c t e d , τ r e a d y
  • To select a processor for scheduling a task τ i in τ r e a d y where the difference between the total local utilization of the tasks in τ s l a c k at t 0 ( τ j τ s l a c k ( t 0 ) ) and the total capacity of the processors in P s l a c k ( τ j τ s l a c k r j ) is greater than zero: τ j τ s l a c k r j ( t 0 ) p j P s l a c k c j > 0 .
    • If τ j τ s l a c k r j ( t 0 ) p j P s l a c k c j r i ( t 0 ) , the task is additionally assigned to a previously selected processor without selecting an additional processor. The assigned task is moved from τ s l a c k to τ r e a d y .
    • If τ j τ s l a c k r j ( t 0 ) p j P s l a c k c j < r i ( t 0 ) , the task is additionally assigned to a previously selected processor without selecting an additional processor. The assigned task is moved from τ s l a c k to τ r e a d y .
  • If τ j τ s l a c k r j ( t 0 ) p j P s l c j = r i ( t 0 ) ,
    • All processors and tasks in P s l a c k and τ s l a c k are moved to P m a x and τ m a x .
    • The processor whose capacity is the lowest for scheduling a task τ i , is selected from the following processor set, { p j | c j r i ( t 0 ) w h e r e p j P s e l e c t e d } . If the capacity of the selected processor is equal to the local utilization ( r i ( t 0 ) ) of the task t i , the processor and the task are moved to P f i x e d and τ f i x e d , respectively. Otherwise, they are moved to P s l a c k and τ s l a c k , respectively.
Algorithm 2 shows how to classify the processor set P s e l e c t e d into the following categories: P f i x e d , P m a x , and P s l a c k . The task with the highest local utilization is considered first to classify the processor set and the task set. The function getFirstLocalUtilizationTask( τ r e a d y ) returns the task with the highest local utilization in τ r e a d y . The function getMinimumCapacityProcessor(availableCapacity, τ, P s e l e c t e d ) takes availableCapacity into account to return the processor whose capacity is the lowest for scheduling a task τ in P s e l e c t e d . If the capacity of the returned processor is equal to the local utilization of the task, the processor and the task is moved to P s l a c k and τ s l a c k , respectively. If availableCapacity is 0, the processors in P s l a c k and the tasks in τ s l a c k are moved to P m a x and τ m a x .

3.3. Scheduling Strategy

In the paper written by Chen et al. [29] , there are two suggested methods for scheduling on a uniform multi-processors. However, event-t, event-s, and event-r presented above are not taken into account in these scheduling methods. In this section, we propose a new T-L plane based scheduling method in which event-t, event-s, and event-r are used to reduce the power consumption of a uniform multi-processors. When the τ f i x e d , τ m a x , and τ s l a c k tasks are scheduled with the P f i x e d , P m a x , and P s l a c k processor sets, the tasks cannot be moved from one processor set to another in order to generate no idle time on the processors in P f i x e d and P m a x . The remaining part shows the movement of elements between task sets and processor sets and the processor assignment when a rescheduling event occurs. Since event-t as defined above targets identical multi-processors is not suitable for uniform multi-processors, it is redefined as in Definition 1.
Definition 1.
An event-t in uniform multi-processors occurs at t t if the following conditions are met.
  • t f t t C s l e e p .
  • τ i τ a c t i v e r i ( t t ) = ( p i P s l a c k c i ) c j where p j P s l a c k .
Algorithm 3 shows the process of assigning tasks to processors when a rescheduling event occurs. All the tasks in the set τ a c t i v e are moved to the set τ r e a d y , and all the tasks in the set τ a c t i v e are deleted. The function eraseAll( τ a c t i v e ) removes all elements in the set τ a c t i v e . Tasks are assigned to processors in each processor set in the following order: P s l a c k , P m a x , and P f i x e d . The function getMaximumLocalUtilizationTask(p.c, τ f i x e d , τ r e a d y ) returns the task with the highest local utilization in τ f i x e d and τ r e a d y where the task can be performed on the processor with the capacity of p.c. The function getFirstLocalUtilizationTask( τ f i x e d , τ r e a d y ) returns the task with the highest local utilization in τ f i x e d and τ r e a d y . The function allocateTaskToProcessor( τ , p) assigns the task τ to the processor p.
Algorithm 2 Classification of selected processors for scheduling
  1:
Input: P s e l e c t e d , τ r e a d y
  2:
Output: P f i x e d , P m a x , P s l a c k , τ f i x e d , τ m a x , τ s l a c k
  3:
P f i x e d —The set of processors fixed by a task
  4:
P m a x —The set of processors having maximum utilization
  5:
P s l a c k —The set of processors to be able to have slack time
  6:
τ f i x e d —The set of tasks fixed to a processor on on P f i x e d
  7:
τ m a x —The set of tasks scheduled on P m a x
  8:
τ s l a c k —The set of tasks scheduled on P s l a c k
  9:
τ 1 —Temporary variable for tasks
10:
τ 2 —Temporary variable for tasks
11:
p 1 —Temporary variable for processors
12:
p 2 —Temporary variable for processors
13:
repeat
14:
     τ 1 = getFirstLocalUtilizationTask( τ r e a d y );
15:
    availableCapacity = p i P s l a c k p i . c τ i τ s l a c k τ i . r ( t 0 ) ;
16:
     p 1 = getMinimumCapacityProcessor( a v a i l a b l e C a p a c i t y , τ 1 , P s e l e c t e d );
17:
    if p 1 . c = τ 1 . r ( t 0 ) then
18:
        add( p 1 , P f i x e d );
19:
        add( τ 1 , τ f i x e d );
20:
    else if availableCapacity = 0 then
21:
        for τ 2 τ s l a c k do
22:
           add( τ 2 , τ m a x );
23:
           erase( τ 2 , τ s l a c k );
24:
        end for
25:
        for p 2 P s l a c k do
26:
           add( p 2 , P m a x );
27:
           erase( p 2 , P s l a c k );
28:
        end for
29:
        add( τ 1 , τ s l a c k );
30:
        add( p 1 , P s l a c k );
31:
    else
32:
        add( τ 1 , τ s l a c k );
33:
        add( p 1 , P s l a c k );
34:
    end if
35:
    erase( p 1 , P s e l e c t e d );
36:
until τ 1 is not null
37:
return P f i x e d , P m a x , P s l a c k , τ f i x e d , τ m a x , τ s l a c k
Algorithm 4 shows the movement of the elements between processor sets and task sets. When an event-b occurs, all the tasks which have triggered an event-b are moved to τ d o n e and are removed from τ a c t i v e . The function getEventTasks() returns all the tasks that have triggered the event-b. When an event-c or an event-f occurs, all the tasks that have triggered the event are moved to τ f i x e d , and the processors that have triggered the event are moved to P f i x e d . The function getProcessor( τ . r ( t 0 ) , P m a x ) returns the processor with the capacity τ . r ( t 0 ) in P m a x . When an event-t occurs, the processors which can be switched to sleep mode are moved to P s l e e p and are removed from P s l a c k . When an event-s or an event-r occurs, all the tasks that have triggered the event are moved to τ d o n e and are removed from τ a c t i v e . The function reallocateProcessorTime() assigns the available processing time to a task with remaining execution time in τ d o n e . The assigned task is moved to τ r e a d y .
Algorithm 3 Assignment of tasks to processors at rescheduling
  1:
Input: P f i x e d , P m a x , P s l a c k , τ f i x e d , τ m a x , τ s l a c k
  2:
Output: τ f i x e d , τ m a x , τ s l a c k
  3:
for τ τ a c t i v e do
  4:
    add( τ , τ r e a d y );
  5:
end for
  6:
eraseAll( τ a c t i v e );
  7:
for p P s l a c k do
  8:
     τ = getMaximumLocalUtilizationTask( p . c , τ s l a c k , τ r e a d y );
  9:
    if τ is null then
10:
         τ = getFirstLocalUtilizationTask( τ s l a c k , τ r e a d y );
11:
    end if
12:
    allocateTaskToProcessor( τ , p );
13:
    erase( τ , τ r e a d y );
14:
    add( τ , τ a c t i v e );
15:
end for
16:
for p P m a x do
17:
     τ = getMaximumLocalUtilizationTask( p . c , τ m a x , τ r e a d y );
18:
    if τ is null then
19:
         τ = getFirstLocalUtilizationTask( τ m a x , τ r e a d y );
20:
    end if
21:
    allocateTaskToProcessor( τ , p );
22:
    erase( τ , τ r e a d y );
23:
    add( τ , τ a c t i v e );
24:
end for
25:
for p P f i x e d do
26:
     τ = getMaximumLocalUtilizationTask( p . c , τ f i x e d , τ r e a d y );
27:
    allocateTaskToProcessor( τ , p );
28:
    erase( τ , τ r e a d y );
29:
    add( τ , τ a c t i v e );
30:
end for
31:
return τ f i x e d , τ m a x , τ s l a c k
Figure 4 shows the scheduling in the first plane from the proposed method when scheduling the tasks of Table 8 on the processors listed in Table 9. Algorithm 2 is used to categorize the processor sets and ready tasks selected by Algorithm 1 at t 0 . Task τ 5 that has triggered an event-c at τ 1 and the processor p 3 whose capacity is equal to the local utilization of τ 5 are moved to τ f i x e d and P f i x e d , respectively. At the same time, the processor p 4 is moved to P s l e e p by event-t. Task τ 1 that has triggered an event-b at t 2 is moved to τ d o n e . Task τ 3 that has triggered an event-b at t 3 is moved to τ d o n e . At the same time, task τ 3 that has triggered an event-c and the processor p 1 whose capacity is equal to the local utilization of τ 3 are moved to τ f i x e d and P f i x e d , respectively. Table 10 shows the elements added to the processor and task sets by Algorithm 4 at each event in the 1st plane. Tasks are assigned to processors by Algorithm 3. As shown in Figure 4, tasks assigned to processors move diagonally along the slope of the processor capacity and tasks unassigned to processors will move horizontally.
Algorithm 4 Movement of elements during rescheduling in the T-L plane
  1:
Input: P f i x e d , P m a x , P s l a c k , τ f i x e d , τ m a x , τ s l a c k
  2:
Output: P s l e e p , τ r e a d y , τ a c t i v e , τ d o n e
  3:
τ a c t i v e —The set of tasks to be excuted
  4:
τ d o n e —The set of tasks to be done
  5:
if event-b then
  6:
    T = getEventbTasks();
  7:
    for τ T do
  8:
        add( τ , τ d o n e );
  9:
        erase( τ , τ a c t i v e );
10:
    end for
11:
else if event-c ∣ event-f then
12:
    T = getEventcOrEventfTasks();
13:
    for τ T do
14:
        add( τ , τ f i x e d );
15:
        if τ τ m a x then
16:
           p = getProcessor( τ . r ( t 0 ) , P m a x );
17:
           erase( τ , τ m a x );
18:
           erase( p , P m a x );
19:
        else
20:
           p = getProcessor( τ . r ( t 0 ) , P s l a c k );
21:
           erase( τ , τ s l a c k );
22:
           erase( p , P s l a c k );
23:
        end if
24:
        add( p , P f i x e d );
25:
    end for
26:
else if event-t then
27:
    capacity = p i P s l a c k p i . c τ i τ s l a c k τ i . r ( t 0 )
28:
    p = getProcessor(capacity, P s l a c k );
29:
    add( p , P s l e e p );
30:
    erase( p , P s l a c k );
31:
else if event-s ∣ event-r then
32:
    T = getEventsOrEventrTasks();
33:
    for τ T do
34:
        add( τ , τ d o n e );
35:
        erase( τ , τ a c t i v e );
36:
    end for
37:
    reallocationProcessorTime();
38:
    for τ τ d o n e do
39:
        if τ . l ( t c u r ) = 0 then
40:
           continue;
41:
        else
42:
           add( τ , τ r e a d y );
43:
           erase( τ , τ d o n e );
44:
        end if
45:
    end for
46:
end if
47:
return P s l e e p , τ r e a d y , τ a c t i v e , τ d o n e

4. Energy Efficiency on Uniform Multi-Processors

In this section, the performance of the proposed algorithm is compared with the major algorithms previously developed for power management. We implemented a simulator operating in Windows 10 using the Ruby language (version 2.4.1) for the experiments. Figure 5 illustrates the architecture of the simulator. The results of the simulation show the energy consumption for task executions, as well as the energy overheads associated with the state transitions.

4.1. Experiment Environment

The characteristics of the cortex-A7 core in Marvell’s MV78230, which is the Multi-Core ARMv7 system based on the chip processor, is used to set the experimental parameters of the processor in the simulator. This core supports dynamic frequency scaling and dynamic power down options. Table 11 and Table 12 show that cortex-A7 supports six frequency levels and five processor states. Run thermal is used in the stress test of the CPU. The deep idle and sleep modes consume the same energy with respect to the CPU. We consider the run typical, idle, and sleep modes in Table 12 for our experiment. WolfBot [16], which is a distributed mobile sensing platform, has ARMv7 based cortex processors.
To confirm the scalability of the proposed algorithm, we change the number of available processors within the range 8–32. Then, we use the Emberson procedure to construct 100 task sets on each available processor. The total utilization of the task set is equal to 8, and the task has a utilization within 0.01–0.99. The period of each task is evenly distributed within 10–150 and simulated for 1000 system units.

4.2. Experiment Results and Analysis

Table 13 shows the platform type and power management technique of the algorithm to be simulated. The algorithm’s platform type is called “non-uniform” when the associated frequency of each processor is independently adjustable, and is called “uniform” when it can change all the frequencies at a constant rate when scaling the frequency of the processor. It is possible for each processor among the uniform multi-processors to operate at a different frequency. A job has a different execution time depending on which processor is allocated. These platforms are otherwise called “unrelated”.
Figure 6 shows the power efficiency obtained by simulating the five algorithms mentioned in Table 13 while varying the number of available processors and the number of tasks. We implement our proposed algorithm as well as the following models: PCG, the original uniform algorithm without any power management [29]; Uniform-DPM, our proposed scheduling algorithm for DPM-embedded uniform multi-processors; GMF [31]; Independent RT-SVFS [30]; and Uniform RT-SVFS [30]. The x-axis of Figure 6 represents the number of available processors, and the y-axis represents the normalized power consumption (NPC). The power consumption consumed by the PCG is measured by the reference consumption and the power consumption rate of each algorithm. Figure 6 show the results when the number of tasks composing a task set is 12, 16, 20, and 24, respectively. All of the algorithms to be simulated is global optimal scheduling. Thus, since the total utilization of the task set used in the simulation is fixed at 8, the power efficiency of all algorithms shows 100% energy consumption in all scheduling using eight processors. As shown in Figure 6, the GMF and RT-SVFS algorithms change the power efficiency according to the number of tasks, while the proposed algorithm, Uniform-DPM, consumes the same a mount of power. This is because they always generate the same idle time. In addition, in the case of many available processors, the proposed algorithm shows high power efficiency by preventing unnecessary processor activation and idle time fragmentation, and by preventing frequent state transitions of the processor. GMF and independent RT-SVFS have similar power efficiencies because they determine the frequency of each processor independently. GMF finds a global optimal solution in the search spaces, but not Independent-SVFS. Thus, GMF is better than Independent-SVFS, as shown in Figure 6. Uniform RT-SVFS adjusts the frequency of all processors to a certain ratio, so if the number of tasks is small, the energy efficiency is not good because the work can be concentrated on some processor and the frequency of the processor cannot be lowered. However, as the number of tasks increases, the number of tasks can be divided and processed simultaneously by multiple processors, which can reduce the frequency of the processor. Table 14 and Table 15 show the energy efficiency characteristics of this proposed algorithm. Table 14 shows that Uniform-DPM always shows constant energy efficiency regardless of the number of tasks. Table 15 shows that the energy efficiency increases as the number of processors increases.

5. Conclusions and Future Works

The lifetime of WSNs is closely related to the management of sensor nodes operating at limited energy. In this paper, we propose a power management method for sensor nodes supporting DPM-enabled uniform multi-processors. In the proposed approach, the selection of processors to process a set of tasks and the assignment of tasks to the selected processors have been proposed in terms of energy efficiency. In addition, we implement a simulator to measure the power consumption of various scheduling algorithms. The experimental results show that the proposed algorithms provide better scalability to the number of available processors than DVFS-based approaches. Currently, our proposed algorithms can handle periodic tasks with implicit deadlines. In future work, we plan to extend our algorithms to handle sporadic tasks with time constraint. We are very interested in combining the DVFS and DPM approaches for T-L plane abstraction as well. In addition, studies on trade-offs between the power usage and computational complexity, as well as performance evaluations on overloaded situations, would be interesting problems for potential future research.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant (NRF-2017R1D1A1B03029552, NRF-2017R1E1A1A01075803) funded by the Korea government (MSIP).

Author Contributions

Youngmin Kim and Chan-Gun Lee conceived and developed the algorithm; Ki-Seong Lee performed the experiments; Youngmin Kim and Ki-Seong Lee analyzed the data; Youngmin Kim and Chan-Gun Lee verified the results and finalized the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lymberopoulos, D.; Priyantha, N.B.; Zhao, F. mPlatform: A reconfigurable architecture and efficient data sharing mechanism for modular sensor nodes. In Proceedings of the 6th International Conference on Information Processing in Sensor Networks, Cambridge, MA, USA, 25–27 April 2007. [Google Scholar]
  2. Mittal, S. A survey of techniques for improving energy efficiency in embedded computing systems. Int. J. Comput. Aided Eng. Technol. 2014, 6, 440–459. [Google Scholar] [CrossRef]
  3. Tutuncuoglu, K.; Yener, A. Optimum transmission policies for battery limited energy harvesting nodes. IEEE Trans. Wirel. Commun. 2012, 11, 1180–1189. [Google Scholar] [CrossRef]
  4. Zhang, Y.; He, S.; Chen, J.; Sun, Y.; Shen, X.S. Distributed sampling rate control for rechargeable sensor nodes with limited battery capacity. IEEE Trans. Wirel. Commun. 2013, 12, 3096–3106. [Google Scholar] [CrossRef]
  5. Tan, Y.K.; Panda, S.K. Energy harvesting from hybrid indoor ambient light and thermal energy sources for enhanced performance of wireless sensor nodes. IEEE Trans. Ind. Electron. 2011, 58, 4424–4435. [Google Scholar] [CrossRef]
  6. Wang, Y.; Liu, X.; Yin, J. Requirements of quality of service in wireless sensor network. In Proceedings of the International Conference on Systems and International Conference on Mobile Communications and Learning Technologies (ICN/ICONS/MCL 2006), Morne, Mauritius, 23–29 April 2006. [Google Scholar]
  7. Krishnarnurthy, R.; Alvandpour, A.; De, V.; Borkar, S. High-performance and low-power challenges for sub-70 nm microprocessor circuits. In Proceedings of the IEEE Custom Integrated Circuits Conference, Orlando, FL, USA, 15 May 2002. [Google Scholar]
  8. Venkatesh, G.; Sampson, J.; Goulding, N.; Garcia, S.; Bryksin, V.; Lugo-Martinez, J.; Swanson, S.; Taylor, M.B. Conservation cores: Reducing the energy of mature computations. In ACM SIGARCH Computer Architecture News; ACM: New York, NY, USA, 2010; pp. 205–218. [Google Scholar]
  9. Esmaeilzadeh, H.; Blem, E.; St Amant, R.; Sankaralingam, K.; Burger, D. Dark silicon and the end of multicore scaling. In ACM SIGARCH Computer Architecture News; ACM: New York, NY, USA, 2011; pp. 365–376. [Google Scholar]
  10. Smarr, L. Project Greenlight: Optimizing cyber-infrastructure for a carbon-constrained world. Computer 2010, 43, 22–27. [Google Scholar] [CrossRef]
  11. McLurkin, J.; Lynch, A.J.; Rixner, S.; Barr, T.W.; Chou, A.; Foster, K.; Bilstein, S. A low-cost multi-robot system for research, teaching, and outreach. In Distributed Autonomous Robotic Systems; Springer: New York, NY, USA, 2013; pp. 597–609. [Google Scholar]
  12. Mondada, F.; Bonani, M.; Raemy, X.; Pugh, J.; Cianci, C.; Klaptocz, A.; Magnenat, S.; Zufferey, J.C.; Floreano, D.; Martinoli, A. The e-puck, a robot designed for education in engineering. In Proceedings of the 9th Conference on Autonomous Robot Systems and Competitions, Castelo Branco, Portugal, 7 May 2009. [Google Scholar]
  13. Bonani, M.; Longchamp, V.; Magnenat, S.; Rétornaz, P.; Burnier, D.; Roulet, G.; Vaussard, F.; Bleuler, H.; Mondada, F. The marXbot, a miniature mobile robot opening new perspectives for the collective-robotic research. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2010), Taipei, Taiwan, 18–22 October 2010. [Google Scholar]
  14. Brutschy, A.; Pini, G.; Decugniere, A. Grippable Objects for the Foot-Bot; IRIDIA Technical Report, Technical Report TR/IRIDIA/2012-001; Université Libre de Bruxelles: Brussels, Belgium, 2012. [Google Scholar]
  15. Chen, P.; Ahammad, P.; Boyer, C.; Huang, S.I.; Lin, L.; Lobaton, E.; Meingast, M.; Oh, S.; Wang, S.; Yan, P.; et al. CITRIC: A low-bandwidth wireless camera network platform. In Proceedings of the Second ACM/IEEE International Conference on Distributed Smart Cameras (ICDSC 2008), Stanford, CA, USA, 7–11 September 2008. [Google Scholar]
  16. Betthauser, J.; Benavides, D.; Schornick, J.; O’Hara, N.; Patel, J.; Cole, J.; Lobaton, E. WolfBot: A distributed mobile sensing platform for research and education. In Proceedings of the 2014 Zone 1 Conference of the American Society for Engineering Education (ASEE Zone 1), Bridgeport, CT, USA, 3–5 April 2014. [Google Scholar]
  17. Chung, H.; Kang, M.; Cho, H.D. Heterogeneous Multi-Processing Solution of Exynos 5 Octa with ARM® big. LITTLE™ Technology. Samsung White Paper. 2012. Available online: https://www.arm.com/files/pdf/Heterogeneous_Multi_Processing_Solution_of_Exynos_5_Octa_with_ARM_bigLITTLE_Technology.pdf (accessed on 31 October 2017).
  18. Kamdar, S.; Kamdar, N. big. LITTLE architecture: Heterogeneous multicore processing. Int. J. Comput. Appl. 2015, 119, 1. [Google Scholar] [CrossRef]
  19. Youngmin, K.; Ki-Seong, L.; Byunghak, K.; Chan-Gun, L. TL Plane Based Real-Time Scheduling Using Dynamic Power Management. IEICE Trans. Inf. Syst. 2015, 98, 1596–1599. [Google Scholar]
  20. Kim, Y.; Lee, K.S.; Pham, N.S.; Lee, S.R.; Lee, C.G. TL Plane Abstraction-Based Energy-Efficient Real-Time Scheduling for Multi-Core Wireless Sensors. Sensors 2016, 16, 1054. [Google Scholar] [CrossRef] [PubMed]
  21. Jan, M.R.; Anantha, C.; Borivoje, N. Digital Integrated Circuits—A Design Perspective; Pearson Publishing: London, UK, 2003. [Google Scholar]
  22. Chen, G.; Huang, K.; Knoll, A. Energy optimization for real-time multiprocessor system-on-chip with optimal DVFS and DPM combination. ACM Trans. Embed. Comput. Syst. 2014, 13, 111. [Google Scholar] [CrossRef]
  23. Carpenter, J.; Funk, S.; Holman, P.; Srinivasan, A.; Anderson, J.H.; Baruah, S.K. A Categorization of Real-Time Multiprocessor Scheduling Problems and Algorithms In Handbook on Scheduling Algorithms, Methods, and Models; Chapman Hall/CRC: Boca Raton, FL, USA, 2004; Volume 19, pp. 1–30. [Google Scholar]
  24. Holman, P.; Anderson, J.H. Adapting Pfair scheduling for symmetric multiprocessors. J. Embed. Comput. 2005, 1, 543–564. [Google Scholar]
  25. Anderson, J.H.; Srinivasan, A. Pfair scheduling: Beyond periodic task systems. In Proceedings of the 2000 Seventh International Conference on Real-Time Computing Systems and Applications, Cheju Island, Korea, 12–14 December 2000. [Google Scholar]
  26. Levin, G.; Funk, S.; Sadowski, C.; Pye, I.; Brandt, S. DP-FAIR: A simple model for understanding optimal multiprocessor scheduling. In Proceedings of the 2010 22nd Euromicro Conference on Real-Time Systems (ECRTS), Brussels, Belgium, 6–9 July 2010. [Google Scholar]
  27. Cho, H.; Ravindran, B.; Jensen, E.D. An optimal real-time scheduling algorithm for multiprocessors. In Proceedings of the 27th IEEE International Real-Time Systems Symposium (RTSS’06), Rio de Janeiro, Brazil, 5–8 December 2006. [Google Scholar]
  28. Funk, S.H.; Meka, A. U-LLREF: An optimal scheduling algorithm for uniform multiprocessors. In Proceedings of the 9th Workshop on Models and Algorithms for Planning and Scheduling Problems, Abbey Rolduc, The Netherlands, 29 June–3 July 2009. [Google Scholar]
  29. Chen, S.Y.; Hsueh, C.W. Optimal dynamic-priority real-time scheduling algorithms for uniform multiprocessors. In Proceedings of the Real-Time Systems Symposium, Barcelona, Spain, 30 November–3 December 2008. [Google Scholar]
  30. Funaoka, K.; Kato, S.; Yamasaki, N. Energy-efficient optimal real-time scheduling on multiprocessors. In Proceedings of the 2008 11th IEEE International Symposium on Object Oriented Real-Time Distributed Computing (ISORC), Orlando, FL, USA, 5–7 May 2008. [Google Scholar]
  31. Moreno, G.A.; De Niz, D. An optimal real-time voltage and frequency scaling for uniform multiprocessors. In Proceedings of the 2012 IEEE 18th International Conference on Embedded and Real-Time Computing Systems and Applications (RTCSA), Seoul, Korea, 19–22 August 2012. [Google Scholar]
  32. Funaoka, K.; Takeda, A.; Kato, S.; Yamasaki, N. Dynamic voltage and frequency scaling for optimal real-time scheduling on multiprocessors. In Proceedings of the International Symposium on Industrial Embedded Systems (SIES 2008), Le Grande Motte, France, 11–13 June 2008. [Google Scholar]
  33. Horvath, E.C.; Lam, S.; Sethi, R. A level algorithm for preemptive scheduling. J. ACM 1977, 24, 32–43. [Google Scholar] [CrossRef]
  34. Funk, S.; Goossens, J.; Baruah, S. On-line scheduling on uniform multiprocessors. In Proceedings of the 22nd IEEE Real-Time Systems Symposium (RTSS 2001), London, UK, 3–6 December 2001. [Google Scholar]
Figure 1. Fluid schedule model.
Figure 1. Fluid schedule model.
Sensors 17 02906 g001
Figure 2. A scheduling example in the 1st T-Ler plane.
Figure 2. A scheduling example in the 1st T-Ler plane.
Sensors 17 02906 g002
Figure 3. Transistion of the T-L plane (frequency = α k ).
Figure 3. Transistion of the T-L plane (frequency = α k ).
Sensors 17 02906 g003
Figure 4. A scheduling in the first plane.
Figure 4. A scheduling in the first plane.
Sensors 17 02906 g004
Figure 5. The architecture of the simulator.
Figure 5. The architecture of the simulator.
Sensors 17 02906 g005
Figure 6. Comparing the energy consumption of an energy-efficient approach while varying the number of tasks: (a) 12; (b) 16; (c) 20; and (d) 24.
Figure 6. Comparing the energy consumption of an energy-efficient approach while varying the number of tasks: (a) 12; (b) 16; (c) 20; and (d) 24.
Sensors 17 02906 g006aSensors 17 02906 g006b
Table 1. Mobile sensing platforms.
Table 1. Mobile sensing platforms.
Platform NameProcessor TypeSensor TypeBattery Type
R-One [11]ARM Cortex-M3Accelerometer, gyroscope, bump, IR, ambient light3.7 V lithium-ploymer battery with 2000 mAh
E-puck [12]dsPIC 30F6014AIR, accelerometer, microphoneBattery swappable and rechargeable with 5 Wh
MarXBot [13]ARM11IR, camera, accelerometer, gyroscope, RFID, 2D force, microphoneHot-swappable battery with 38 Wh
Foot-Bot [14]i.MX31 ARM11IR, camera3.7 V lithium-polymer battery with 10-Ah
CITRIC [15]Xscale PXA-270Camera, microphoneFour AA batteries
WolfBot [16]ARM Cortex-A8IR, camera, microphone, ambient light, accelerometer, magnetometer7.4 V lithium-ion battery with 5200 mAh
Table 2. An example of the available processor sets.
Table 2. An example of the available processor sets.
S 1 S 2 S 3
p 1 (voltage = 1.4 v, freq. = 600 MHz, capacity = 1)OXO
XXX
p 5 (voltage = 1.2 v, freq. = 300 MHz, capacity = 0.5)XOO
p 6 (voltage = 1.2 v, freq. = 300 MHz, capacity = 0.5)XOX
XXX
p n 1 (voltage = 1 v, freq. = 150 MHz, capacity = 0.25)OOX
p n (voltage = 1 v, freq. = 150 MHz, capacity = 0.25)OOX
XXX
total capapcity1.51.51.5
Table 3. Task properties.
Table 3. Task properties.
TaskPeriodWCETUtilization
τ 1 5 ms2.5 ms0.5
τ 2 10 ms5 ms0.5
τ 3 10 ms1.25 ms0.25
τ 4 20 ms2.5 ms0.25
Table 4. Dynamic power consumption of some feasible processor sets.
Table 4. Dynamic power consumption of some feasible processor sets.
S 1 S 2 S 3
Dynamic power consumption2.46 α C1.94 α C2.68 α C
Table 5. Processor properties.
Table 5. Processor properties.
p 1 p 2 p 3 p 4
Supply voltage1.4 V1.2 V1.0 V1.0 V
Operating frequency600 MHz300 MHz150 MHz75 MHz
Processing capacity10.50.250.125
Table 6. Selecting processors for scheduling a task set.
Table 6. Selecting processors for scheduling a task set.
S 1 S 2 S 3 S 4 S 5 S 6 S 7 S 8
p 1 OOOOXOOX
p 2 OOOXOOXO
p 3 OOXOOXOO
p 4 OXOOOXXX
Total capacity1.8751.751.6251.3750.8751.51.250.75
Table 7. Task properties.
Table 7. Task properties.
TaskPeriodWCETUtilization
τ 1 5 ms4.5 ms0.9
τ 2 10 ms4.25 ms0.425
Table 8. Task properties.
Table 8. Task properties.
TaskPeriodWCETUtilization
τ 1 5 ms4 ms0.8
τ 2 5 ms2.5 ms0.5
τ 3 10 ms3 ms0.3
τ 4 10 ms2 ms0.2
τ 5 20 ms2 ms0.1
Table 9. Processor properties.
Table 9. Processor properties.
p 1 p 2 p 3 p 4
Supply voltage1.4 V1.2 V1.0 V1.0 V
Processing capacity10.50.250.25
Table 10. Example of sets at events in the plane.
Table 10. Example of sets at events in the plane.
SetElement
t 0 t 1 t 2 t 3
τ f i x e d τ 2 τ 2 , τ 5 τ 2 , τ 4 τ 2 , τ 4 , τ 5
P f i x e d p 2 p 2 , p 3 p 2 , p 3 p 1 , p 2 , p 3
τ m a x ....
P m a x ....
τ s l a c k τ 1 , τ 3 , τ 4 , τ 5 τ 1 , τ 3 , τ 4 τ 3 , τ 4 .
P s l a c k p 1 , p 3 , p 4 p 1 p 1 .
τ d o n e .. τ 1 τ 1 , τ 3
P s l e e p . p 4 p 4 p 4
Table 11. Frequency levels of the cortex-A7 core.
Table 11. Frequency levels of the cortex-A7 core.
ParameterLevel 1Level 2Level 3Level 4
Frequency (MHz)800106613331600
Run typical power (W)3.33.644.9
Table 12. Power states of the cortex-A7 core.
Table 12. Power states of the cortex-A7 core.
StatesPower (Watts)
Run Thermal5.9
Run Typical4.9
Idle2.4
Deep Idle0.07
Sleep0.07
Table 13. Summary of the energy-efficient scheduling algorithms.
Table 13. Summary of the energy-efficient scheduling algorithms.
Algorithm NamePlatform TypePower Management
PCGUniform-
Uniform-DPM (proposed)UniformDPM
GMFNon-uniformSVFS
Independent RT-SVFSNon-uniformSVFS
Uniform RT-SVFSUniformSVFS
Table 14. Summary of the experimental results by varying the number of tasks.
Table 14. Summary of the experimental results by varying the number of tasks.
Saved Norm. Power Consumption (%)
# of Processors# of Tasks (Total Untilization)Uniform-DPMGMFIndependent RT-SVFSUniform RT-SVFS
1212 (8)19.69.98.30.3
1216 (8)19.614110.8
1220 (8)19.616.414.61.7
1224 (8)19.618.6173.6
Table 15. Summary of the experimental results by varying the number of uniform processors.
Table 15. Summary of the experimental results by varying the number of uniform processors.
Saved Norm. Power Consumption (%)
# of Processors# of Tasks (Total Untilization)Uniform-DPMGMFIndependent RT-SVFSUniform RT-SVFS
1224 (8)19.618.417.13.6
1624 (8)32.817.616.43.3
2024 (8)42.21514.32.7
2424 (8)49.414.512.92.5
2824 (8)54.912.111.62.2
3224 (8)59.31110.62

Share and Cite

MDPI and ACS Style

Kim, Y.; Lee, K.-S.; Lee, C.-G. Energy Efficient Real-Time Scheduling Using DPM on Mobile Sensors with a Uniform Multi-Cores. Sensors 2017, 17, 2906. https://doi.org/10.3390/s17122906

AMA Style

Kim Y, Lee K-S, Lee C-G. Energy Efficient Real-Time Scheduling Using DPM on Mobile Sensors with a Uniform Multi-Cores. Sensors. 2017; 17(12):2906. https://doi.org/10.3390/s17122906

Chicago/Turabian Style

Kim, Youngmin, Ki-Seong Lee, and Chan-Gun Lee. 2017. "Energy Efficient Real-Time Scheduling Using DPM on Mobile Sensors with a Uniform Multi-Cores" Sensors 17, no. 12: 2906. https://doi.org/10.3390/s17122906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop