1. Introduction
Generational shifts in the world of mobile networks have been driven by the unprecedented growth in mobile data traffic. The exponentially growing number of connected devices and the unquenchable thirst for better mobile broadband experience is also a driving force behind this evolution. The story of 5G is no more different as it is envisioned to provide a gigabit experience and virtually zero latency while withstanding an expected 500-fold increase in mobile video traffic over the next ten years. It is predicted that 75% of global mobile data will be video content in which 6.7% will be machine-to-machine (M2M) communication. Such massive growth of multimedia traffic was further fueled by social media feeds, such as Facebook and Twitter (representing 15% of the traffic [
1]). This trend is undoubtedly going to stress the capacity of core networks, wireless links and mobile backhauls to their limits eventually leading to poor quality-of-experience (QoE).
In order to cope with this issue, a 5G network is gearing up with various technologies including device-to-device (D2D) communication [
2], Massive MIMO (Multiple-input multiple-output) [
3,
4], Device-centric architecture [
5], Small Cells, Caching, mmWave communication [
6] all orchestrated by self organizing networks (SON) [
7]. Among them, caching through D2D mode of communication is being considered a promising technology aimed for reducing backhaul load and minimizing latency [
5]. The key idea of D2D enabled caching is to store the popular content of the network at the user end. The key benefit here is that it brings likely to be accessed content as close as possible to the users. This approach can reduce latency substantially, a key challenging requirement in 5G [
8]. It can also increase the effective throughput and reduce the backhaul load [
9]. The three key research questions in designing an optimal caching enabled next generation Radio Access Network (RAN) are:
Where to cache? How to cache? and
What to cache? Where to cache focuses on the problem of caching the content over the network in such a way that it reduces the network traffic. How to cache focuses on multiple ways of content caching, and what to cache focuses on the problem of finding content having the highest probabilities to be used in the near future. In order to address these questions, various attempts have been made by the researchers such as [
9,
10,
11,
12,
13]. For a detailed review of caching in RAN, the reader is referred to [
14,
15]. However, to the best of the authors’ knowledge, this is a first study that investigates the potential of the leveraging detection of influential users via centrality measure in cellular networks and then uses that detection for identifying what to cache and pro-actively caching those likely to be accessing content using D2D communications.
Online Social Networks, one of the key sources of mobile data explosion, are known to be scale-free networks, which follow power-law distribution. For example, the Twitter network has millions of users and an enormous flow of information is published daily. There are small number of Twitter users who have a large number of followers while there is a huge number of users who have a small number of followers. In this paper, our key idea to design and evaluate a novel proactive yet low complexity caching solution is as follows. It is highly probable that content accessed or generated by influential users will become popular due to their relatively higher following and connectivity. Furthermore, popular content is expected to be in demand in the same community i.e., the community of the influential user first and then spread slowly in the remaining network. The mapping structure, information diffusion and influence phenomena of the social networks can thus be exploited for proactively caching those content which are expected to be highly demanded in the near future, for reducing the backhaul traffic load [
16]. Therefore, we represent a mobile cellular network in the form of a graph and use graph theory and Social Network Analysis (SNA) techniques to improve QoE and reduce the backhaul traffic load.
This study considers social networks aware D2D caching by focusing on influential users, their communities, and content. Initially, we use a preferential attachment model [
17] for representing the mobile cellular network. We then perform clustering using a CC-GA algorithm [
18], and find influential users using an eigenvector centrality measure [
16]. Finally, based on the users’ previous history, we proactively cache the content of the influential users with respect to the available cache size. We perform extensive system level experiments to evaluate the gain of the proposed approach. The results reveal that 48% of backhaul traffic load can be reduced (48% satisfied requests) when users generating requests are 100% using the proposed approach.
The rest of the paper is organized as follows.
Section 2 presents a brief background to caching methodologies in future cellular networks.
Section 3 gives the most important related work.
Section 4 discusses the methodology of the proposed approach.
Section 5 presents the experimental setup and numerical results. Finally,
Section 6 summarizes and concludes the paper.
2. Caching in Future Cellular Networks
We have seen a tremendous growth in cellular networks in the last decade. This growth is mainly driven by the exponentially growing demands of a high data rate by the end users. Various techniques have been employed to achieve a high data rate as well as spectral efficiency. Spectral efficiency can be achieved by efficiently managing cellular spectrum. These techniques range from control data split architecture (CDSA) [
19], a cloud-radio access network (CRAN) to ultra-dense cellular networks (UDCN) [
20] and D2D communication [
21]. Almost all of these technologies are considered as enabling technologies for an upcoming fifth-generation (5G) cellular network. Among these, multi-tier architecture and D2D communication are well thought out as key enablers for providing access to a massive number of users as well as high data rate connectivity. In multi-tier architecture, there is an overlaid macro-cell (MC) comprising multiple micro-cells (mCs) operating in its coverage area. MC provides an overall coverage service and mobility management, whereas mCs can provide local area services and high data rate connectivity.
On the other hand, caching in cellular networks has been introduced to exploit storage capacity of diverse network devices. These devices range from user side (user equipment (UE), edge devices) to the network side (mCs, MC and extended packet core). In the following, we briefly describe how caching can be done on these diverse network devices.
Edge Caching: When an UE generates a request for specific content, first it will search for that content in its own memory and if that content is cached locally (edge), the UE will access that without any delay. This kind of caching can drastically reduce the backhaul traffic as well as access delay at the cost of large UE memory requirements.
Cluster-Head Caching: If the requested content is not available at UE itself, it will then ask its peers for that content through D2D communication. One way is to search for the content sequentially in every peer device. The other way could be to search for an influential user where it is highly probable that content accessed or generated by this influential user will become popular due to its relativity higher connectivity [
22]. D2D communication will enable this kind of caching in the network. This mechanism also reduces backhaul traffic and access delay.
Micro-Cell Caching: If the requested content is not found using D2D communication, then mC will provide that content to the user if it is cached there. mC will utilize a radio access network (RAN) for delivering that content and not affect the backhaul channel.
Macro-Cell Caching: If the content can not be accessed using the above-mentioned ways, then MC will provide that content by downloading it from EPC (Evolved Packet Core) or from the cloud itself. It will ensure that content is delivered successfully. This mechanism can cause large access delays and an increase in backhaul traffic.
In the above, we answer the question where to cache?; however, in order to understand how to cache?, we might have two different approaches such as reactive and proactive. In the following, we describe these approaches in the context of cellular communication.
Reactive Caching: In the reactive caching approach, files are stored in cache if they were repeatedly requested in the past based on the history of the files accessed. In this caching, any content that remains in high demand for a specific period of time can be cached—for example, viral videos, popular tweets and other highly requested social media content.
Proactive Caching: It is a mechanism to predict future content that might be requested by the users. Based on previous content and user histories, certain content might become popular in the near future, which will be cached before they are actually requested, e.g., popular entities in various societies might trigger similar kinds of trends such as top trends on Twitter, etc. Due to the exponential increase of the demand of cellular data, it is very difficult to maintain the user satisfaction rate. Therefore, the importance of backhaul is increasing dramatically. Various studies [
11,
12] have shown that proactive caching in D2D communication can reduce a significant amount of backhaul traffic load, and is a smarter way to exploit D2D cellular networks. Using this approach in D2D communication in small cellular networks, the popular content is cached at the UE.
In the next section, we focus on the related work that deals with proactive caching using D2D communication.
3. Related Work
In recent years, a large number of studies have investigated caching on the edge in small cell networks. In [
9], a proactive caching approach has been proposed. The authors studied two cases and utilized the spatial and social structure of the network. First, based on correlations among users and file popularity, files are cached during off-peak time. Secondly, influential users are detected and strategic content is cached using D2D communication and social networks. The authors used an eigenvector centrality measure for detecting influential users of the social network and model the content dissemination as a Dirichlet Process (CRP). In the experimental setup, a preferential attachment model is used to map the D2D small cellular network. The numerical results show a significant improvement in gain ratio; however, the proposed technique does not consider content of influential users in a specific community, which makes it less suitable for D2D communication in future cellular networks. The study in [
11] introduced a new QoE metric for satisfying a given file request using proactive caching and then they proposed an optimization algorithm (
propCaching) to maximize QoE. This algorithm is based on the popularity statistics of the requested files and cache files with highest popularity. However, unlike our proposed work, the algorithm has a limitation in that it chooses the popular files for caching based on their statistics or previous history. The work in [
23] considers UE devices as cooperating nodes on which distributed cache is implemented for efficient downloading. However, this approach has certain drawbacks. Firstly, it does not consider the distance between D2D devices for cooperation and, secondly, it stores random content without prioritizing high in demand content. Golrezaei et al. [
10] introduces a collaborative architecture which uses the distributed storage of popular content. In order to choose and cache random files at the user end, authors compute an average number of D2D links that can coexist without interference. However, this work also does not consider content of influential users to cache.
A probabilistic approach for optimizing scheduling policies and D2D caching is considered in [
24]. The authors first derived approximated uploading probability for uploading gain and then they optimize the scheduling factor and caching distribution to optimize the successful offloading probability. In [
25], authors investigate an optimal caching strategy by formulating an optimization problem and showing the relationship between D2D caching distribution and demand distribution for homogeneous Poisson Point Process models with different noise levels. With the exponential increase in multimedia traffic, content caching at every edge node might become a challenging task. To overcome this problem, authors in [
26] proposed an idea of peer-to-peer content delivery networks. The authors have exploited the benefits of distrusted fog nodes for content delivery among the networks. In order to efficiently deploy enabling 5G technologies, authors in [
27] proposed an algorithm to reduce total installation costs. The proposed algorithm considers both the hardware and cloud enabled softwarized components (reusable functional blocks) to ensure the users’ good performance as well as reduction in computation times. Kennington et al. [
28] gives a comprehensive overview of modelling and solving optimization problems arising in wireless networks. Moreover, for a detailed review of applications of big data and caching techniques in mobile computing, the reader is referred to [
9,
29].
Table 1 summarizes caching strategies in comparison with the one proposed in this paper.
Finding social hubs in social networks remains an active research area of Social Network Analysis (SNA). Various measures including Eigenvector Centrality measure, Betweenness measure, Closeness measure, Degree centrality measure and PageRank algorithm have been used extensively for finding the nodes importance in social graphs. Among these measures, Eigenvector centrality measure is the most widely used measure for detecting social hubs in social graphs. This measure has been used by three popular methods: PageRank [
34], HITS (Hypertext Induced Topic Search) [
35], and SALSA (Stochastic Approach for Link Structure Analysis) [
36] for retrieving web information.
It is worth mentioning that several prior studies exist that propose various design approaches for proactive caching schemes [
9,
16,
37]. However, the novelty of the proposed work stems from the simple yet under-explored idea of considering the content of influential users as popular content. Furthermore, our work uses a new community detection algorithm for detecting communities in the network and then
k influential users are found, one for each community. In the next section, we will explain our proposed methodology for proactive caching in socially aware D2D networks.
4. Proposed Methodology
We exploit the social and spatial structure of social networks and use SNA approaches to find the content of the network that is expected to become viral in the future and proactively cache them in order to reduce the backhaul load. As the number of active users increases in the network, load on Small Base Station (SBS) increases. Therefore, D2D communication can play a vital role in reducing the load and increasing the user satisfaction rate. The D2D communication paradigm can be exploited to perform proactive caching and store highly probable content in the users’ caches. In this paper, we introduce a new proactive caching approach, which uses SNA techniques to address this problem. In the following subsection, after describing the system model and the essentials, we shall present our proposed algorithm.
4.1. System Model
Assume that there are
n number of users who generate number of requests
, number of files
with lengths
, and bit rate
. Note that bit rate of the file represents the transferring rate of the file per unit time. The requests are generated by the users in different time frames
. A request
generated by user
n is said to be satisfied iff D2D link capacity is higher than the bit rate of the file. The user satisfaction can be formulated as follows:
where
is the size(length) of the file,
and
is the end and starting time of delivery of file
i.
represents the bit rate of file
f. Based on Equation (
1), the user satisfaction rate
can be defined as:
where
In order to reduce backhaul traffic load, our main goal is to maximize .
The users are considered to be connected with other users of their community in a D2D mode of communication. Thus, the request r of the user i for a file f is directed to its nearest user to search the requested file and will be served if the file exists in the nearby cache. Thus, a good proactive caching strategy can satisfy more requests at the user end, which results in reducing the backhaul traffic load.
The architecture considered in this work assumes the D2D communication enabled small cellular network. Every community in the network finds a central node that can serve a maximum number of D2D users with a minimum distance as shown in
Figure 1. There are a total of
n number of users that are connected with each other via D2D communication with a link capacity
. The users are also connected with the small cell base station with a link capacity
. Assume that the
, then the average backhaul traffic load due to each file of length
can be formulated as follows [
13]:
where
is the backhaul data rate during the content delivery for request
r.
We consider the network as a graph
where
V represents set of users
and
E is set of social links
between the users. For generating the social network for our experimental setup, we used a well known model called preferential attachment proposed by Albert Barabasi [
17] to form the small cellular network. This model basically represents degree distribution of the random generated network i.e., how users connect to each other. The core idea behind the preferential attachment model is that new users prefer to connect to well-connected users over less-connected users proportional to the probability of their degree. The generated network using this model follows power law distribution with exponent
. This process is also known as rich-get-richer or cumulative advantage or Matthew effect. The process starts with some initial subgraph. Each new user comes with some initial
m links. The probability
of connecting user
u with other users
j in the network is:
where
represents the probability of connecting user
u with other users
j in the network.
m represents the number of links to which the new user joins the network and
k represents the degree of each user in the network. This process results in a network with a power-law distribution with exponent = 3.
Figure 2 shows a network of 256 users generated using a preferential attachment model where nodes represent users while links represent the D2D communication. We see that the network contains very few users having a large number of links while the rest of the users have very few links. The degree distribution of the generated network on a log–log scale is plotted in
Figure 3. We can see that the network degree distribution follows power-law distribution.
4.2. Cluster Formation in a Social Network
The first step of our proposed approach is to detect the cluster of each user in the social network as discussed in the previous subsection. For a given graph , there exist partitions of V, where 0. Our goal is to identify a set of clusters that are densely intra-connected and sparsely interconnected. The process of identifying clusters in social networks is known as community detection.
Various kinds of community detection algorithms have been proposed. In this paper, we use the CC-GA algorithm [
18]. The main reason for using CC-GA is its promising results when compared to other state-of-the-art algorithms such as the information theoretic algorithm (Infomap) [
38] and the label prorogation algorithm (LPA) [
39] on various kinds of networks including power-law and Forest fire models. This algorithm uses a well-known SNA measure called a Clustering Coefficient (CC) for generating an initial population. CC gives a better population for a genetic algorithm and results in better clustering of the network. Additionally, the algorithm using CC has shown that it converges very quickly [
18]. In the following equation,
represents the value of clustering coefficient of node
:
where
represents the total number of links between the
neighbors. The higher value of
represents the dense neighborhood of a node. The main intuition behind using CC is that if the neighborhood of a node is densely populated, then it is more likely that all of the neighborhood nodes belong to the same cluster. Furthermore, CC-GA uses Modularity measure [
40] as an objective function that is a quality index for clustering of social networks. It can be formally defined as follows:
where
represents the number internal links within the cluster while
represents the links between clusters.
m represents the total number of links in the whole network. Equation (
6) can be rewritten into a precise mathematical formulation as given below:
In Equation (
7), the first fraction represents the true intra-connectivity of the users, while the second fraction represents the expected number of links within the cluster. The value of modularity ranges between
and 1. Greater value of modularity indicates better clustering of the network while a negative value represents an absence of true clustering in the network.
CC-GA tries to maximize the value of modularity in order to get nearly optimal clustering of the network. It is to be noted that modularity optimization is an NP-hard problem, which can not be solved with a guaranteed optimal solution. Algorithm 1 presents the pseudocode of CC-GA. Initially, the algorithm requires a graph consisting of vertices (V) and edges (E) as inputs. It also requires a set of input parameters including: the termination criteria
(r), population size (
), crossover rate
, mutation rate
, mutation extension rate
and percentage of population
to be selected for the next iteration. Initially, the algorithm computes CC of all nodes (steps 2–4) and generates the initial population (step 5). Afterwards, the genetic operators (crossover and mutation) are performed iteratively until the termination criterion is satisfied (steps 5-25). The algorithm also uses extension of mutation to merge small communities (step 19), and uses modularity measure as an objective function (steps 9, 15, 21). In the end, the algorithm returns the community structure of the input network. For more details about this algorithm, readers are referred to [
18]. The interested readers are referred to [
41,
42] for more about genetic algorithms.
Algorithm 1 CC-GA (G) |
Inputs: Crossover rate; Mutation rate; Population size; Rate of mutation extension; Termination criteria; Selection criteria; Output: = ; - 1:
procedure CC-GA( ) - 2:
for do - 3:
- 4:
end for - 5:
Initialize population - 6:
repeat - 7:
Apply Crossover (); - 8:
Q = Compute fitness (P); - 9:
Compute modularity (O); - 10:
if then - 11:
P ← update population (O); - 12:
end if - 13:
O← Apply mutation (O, ); - 14:
Q = Compute fitness (P); - 15:
Compute modularity (O); - 16:
if then - 17:
P ← update population (O); - 18:
end if - 19:
O ← Apply mutation extension (O, ); - 20:
Q = Compute fitness (P); - 21:
Compute modularity (O); - 22:
if then - 23:
P ← update population (O); - 24:
end if - 25:
until (termination-criterion(r)) - 26:
Chromosome having highest fitness value - 27:
Return - 28:
end procedure
|
4.3. Finding Influential Users and Popular Content
After clustering of the social network, our task is to find the Influential Users (IUs) of the network and then find their content. An influential user can be defined as the most connected and active user of the network [
43]. In order to find the IUs, we use the notion of centrality [
44], which is a well-known concept in SNA. The centrality measure provides a way to quantify the importance of nodes with respect to their positions and neighborhood connectivity within a social network. While various centrality measures exist, eigenvalue centrality is the most successful measure for detecting the social hubs or IUs within a social network. It is a widely used measure for finding the nodes centrality in the networks [
16,
45,
46].
In order to find the IUs, we exploit the EigenVector Centrality measure. This measure uses the notion of eigenvector and eigenvalue of the adjacency matrix. It quantifies centrality of nodes based on centrality of its neighbors. Based on this measure, a node having more central neighbors will have greater centrality value. It can be formally defined as follows:
where
is a normalization constant and
N represents neighbors of node
.
is 1 if nodes
i and
j are directly connected to each other, and 0, otherwise. The greatest value of
represents the highest influence of the user in the network. Equation (
8) can be rewritten in a vector form as follows:
where A represents the adjacency matrix of the graph, and
x is the vector of the eigenvalue scores. Based on eigenvector centrality measure, if our clustering algorithm returns
K clusters of the network, then we find
K influential users, one for each cluster in our social network. As we are considering D2D communication, it is necessary to identify at least one influential user in each cluster in order to communicate with other users directly. In our social network model shown in
Figure 2, different clusters of the network have been shown in different colors after applying our clustering algorithm. There is one central influential user (nodes with greater sizes) found by using an eigenvector centrality measure. For example, user 11 is the influential user of its community represented with the green color, and there are 16 members of its community. Overall, the network is divided into 15 different communities and each community has one influential user. Furthermore, users who join the network earlier have more of a chance to become more connected and are hence more popular and influential (users 0–20).
Once we know the influential users and their communities, our next step is to determine the popular content of each community. For this purpose, inspired by the pervasive use of online social media, we assume that the content requested, generated, or accessed by the influential users is going to be the popular content for the same community, and we proactively cache this content based on the available cache size.
Algorithm 2 shows a pseudocode for computing the centrality values. We initialize an empty matrix for storing the centrality value of each node. Then, the algorithm recomputes the centrality score of each node as a weighted sum for centralities of all nodes in a node neighborhood. In order to normalize the centrality score, is used. We set the value of to the largest value of . The procedure repeats until the values of each node converge.
After describing all the essentials in the following, we shall now present the proposed algorithm.
Algorithm 2 Compute Centrality |
- 1:
procedureCompute Centrality(A) - 2:
Initialize matrix by 1 - 3:
repeat - 4:
for do - 5:
- 6:
Update - 7:
end for - 8:
until values of converge - 9:
end procedure - 10:
Return M
|
4.4. Proposed Algorithm
Algorithm 3 presents our proposed approach for centrality based D2D proactive caching. Initially, we model the system based on the procedure presented in
Section 4.1. After designing the model (step 2), we apply the CC-GA algorithm (step 3) for identifying clusters in the network that returns a set
C depicting the communities of the network (communities and clustering have the same meaning and we use them interchangeably). Afterwards, we initialize a matrix
M (step 4) having the dimensions
, where
C represents the total number of clusters (
row contains the centrality values of cluster
) and
U equals
. This matrix stores the centrality value of each node with respect to its community. Note that
has
values and the remaining indexes which are
will have 0 entries. We also initialize an
matrix to find and store the influential users. The dimensions of this matrix are
. Then, we call the compute centrality algorithm (step 5), presented in Algorithm 2, which returns a vector of centrality values (
) indexes by node labels that are in the form of integers. As we need matrix
M for computing the influential users, we therefore compute it from
(step 6). We compute the primary influential users and secondary influential users from the matrix in each cluster and store them in the matrix (steps 7–10). The reason for finding secondary influential users is to ensure reliability of the system. In the end, we compute the popular content of influential users as described in
Section 4.3 and load it into the cache (step 12).
We see that the proposed Algorithm 3 relies on one primary influential user from one cluster, which raises important questions such as how many clusters the network has and how many users each cluster contains. To answer these questions, we present the distribution of community sizes in power-law networks in
Figure 4. The figure shows the community sizes distribution of the network shown in
Figure 2. We can see that the community size also follows a power law distribution that implies that there are no such communities who have either a large number of users that cover a great part of the network and neither there are such communities who have very few numbers of nodes. Taking this into consideration, we also follow network community size distribution to load the content in the cache of influential users. The user that has a greater community size has more content in the cache and vice versa.
The cellular networks suffer from a network outage problem, which can be in the form of power failure, signal loss or link break down. For maintaining the reliability of the network, we also maintain a list of secondary influential users as shown in Algorithm 3. In case of network outage problems, the secondary influential users will be used for D2D communication.
Algorithm 3 Proactive Caching |
- 1:
procedure D2D Proactive Caching - 2:
Model the system - 3:
CC-GA (G) - 4:
initialize and matrix by 0 - 5:
Compute Centrality(G) - 6:
compute matrix M from - 7:
for do - 8:
- 9:
remove() - 10:
= - 11:
end for - 12:
Compute the popular content of influential users and load into cache - 13:
end procedure
|