Download as pdf or txt
Download as pdf or txt
You are on page 1of 19

Wireless Caching: Technical Misconceptions

and Business Barriers


Georgios Paschos , Ejder Baştuğ? , Ingmar Land , Giuseppe Caire† , and
Mérouane Debbah

Huawei Technologies, France Research Center, Mathematical and Algorithmic Sciences Lab,
arXiv:1602.00173v2 [cs.IT] 3 Jun 2016

20 Quai du Point du Jour, Boulogne-Billancourt, 92100, France


?
Large Networks and Systems Group (LANEAS), CentraleSupélec, Université Paris-Saclay,
3 rue Joliot-Curie, 91192 Gif-sur-Yvette, France

Technische Universität Berlin, Electrical Engineering and Computer Science
Einsteinufer 25, 10587, Berlin, Germany
{georgios.paschos, ingmar.land, merouane.debbah}@huawei.com
[email protected], [email protected]

This research has been partly supported by the ERC Starting Grant 305123 MORE (Advanced Mathematical Tools for Complex
Network Engineering), and the project BESTCOM.

June 7, 2016 DRAFT


1

Wireless Caching: Technical Misconceptions


and Business Barriers
Abstract

Caching is a hot research topic and poised to develop into a key technology for the upcoming 5G
wireless networks. The successful implementation of caching techniques however, crucially depends on
joint research developments in different scientific domains such as networking, information theory,
machine learning, and wireless communications. Moreover, there exist business barriers related to
the complex interactions between the involved stakeholders, the users, the cellular operators, and the
Internet content providers. In this article we discuss several technical misconceptions with the aim to
uncover enabling research directions for caching in wireless systems. Ultimately we make a speculative
stakeholder analysis for wireless caching in 5G.

Index Terms

Edge caching, 5G, Wireless networks.

I. I NTRODUCTION

Caching is a mature idea from the domains of web caching, content delivery networks, and
memory optimization in operating systems. Why is caching still an active topic of discussion? In
the 90s, the traffic in the web exploded, leading its inventor Sir Tim Berners-Lee to declare the
network congestion as one of the main challenges for the Internet of the future. The congestion
was caused by the dotcom boom and specifically due to the client-server model of connectivity,
whereby a webpage was downloaded from the same network server by every Internet user in the
world. The challenge was ultimately resolved by the invention of Content Delivery Networks
(CDNs), and the exploitation of web caching. The latter replicates popular content in many
geographical areas and saves bandwidth by avoiding unnecessary multihop retransmissions. As
a byproduct, it also decreases access time (latency) by decreasing the distance between the two
communicating entities.
Today, 30 years later, we are reviving the same challenge in the wireless domain. The latest
report of Cisco [1] predicts a massive increase of Internet devices connected through the wireless

June 7, 2016 DRAFT


2

access, and warns of a steep increase in mobile traffic which is expected to reach by 2018
roughly 60% of total network traffic, the majority of which will be video. The wireless system
designers strive to fortify 5G wireless networks with higher access rates on the one hand and
with increased densification of network infrastructure on the other. Over the last three decades,
these two approaches are responsible for the majority of network capacity upgrade per unit area,
successfully absorbing the wireless traffic growth. However, with the explosion of access rates
and number of base stations, the backhaul of wireless networks will also become congested [2],
[3] which motivates further the use of caching: store popular reusable information at the base
stations to reduce the load at the backhaul. Furthermore a recent technique [4] combined caching
with coding and revolutionized how goodput scales in bandwidth-limited networks. Therefore,
caching has the potential to become the third key technology for wireless systems sustainability.
The research community is converging to an enabling architecture as the one described in
Figure 1. In the network of the future, memory units can be installed in gateway routers between
the wireless network and the Internet (e.g. in 4G this is called S-GW), in base stations of different
sizes (small or regular size cells), and in end-user devices (e.g. mobile phones, laptops, routers,
etc). In this article, we discuss important topics such as (a) the characteristics of cacheable
content and how this affects caching technologies in wireless, (b) where to install memory, and
(c) the differences between wireless caching and legacy caching techniques. Last, we focus on
business barriers that must be overcome for the successful adoption of wireless caching by the
industry.

II. D EALING WITH M ASSIVE C ONTENT

Not all network traffic is cacheable. Interactive applications, gaming, voice calls, and remote
control signals are examples of information objects that are not reusable and hence cannot
be cached. Nevertheless, most network traffic today (estimated 60% [1]) is deemed cacheable.
We refer to cacheable information objects as content in this article. Since the performance of
caching is inherently connected to the specifics of contents, this section is dedicated to the
understanding of these specifics. In particular we focus on the following misconceptions: (i)
The static IRM model is sufficient for experimentation, (ii) user information cannot be used for
popularity estimation due to the vast number of users, and (iii) security issues precludes caching
at the edge.

June 7, 2016 DRAFT


3

Caching at the
central Edge?
router

core router

base station (or fog)

? ?
controller

origin content base


server/website backhaul link
content delivery station
server storage
small unit
nk
base station link/upli mobile user
down terminal link
D
D2

Figure 1: An illustration of caching in future wireless networks. Contents available in the origin
server are cached at the base stations and user devices, for offloading the backhaul and the
wireless links.

Insufficiency of Static Popularity Models

The standard approach to design and analyze caching systems involves a model for generating
content requests to replace the actual request traces–this approach is often several orders of
magnitude faster. The de facto model for performance analysis of web caching is the Inde-
pendence Reference Model (IRM): content n is requested according to an independent Poisson
process with rate λpn , where pn refers to the content popularity, modelled by a power law,
i.e., pn ∝ n−α , α > 0. This well-established model thrives due to its simplicity; it only has
two parameters, namely λ to control the rate of requests, and α to control the skewness of the
popularity.
Numerous studies fit IRM to real traffic with satisfactory results [5], so why do we need
to change it? The IRM assumes that the content popularity is static, which is of course not
true. Trending tweets, breaking news, and the next episode of Game of Thrones, are examples
of ephemeral content with rapidly changing popularity; they appear, they become increasingly
popular, and they gradually become unpopular again.
In fact, [6] considers large YouTube and Video on Demand (VoD) datasets and discovers:
time-varying models are more accurate than IRM with respect to caching performance analysis–
Figure 2 reproduces the comparison when fitting YouTube data and shows the superiority of

June 7, 2016 DRAFT


4

modeling the popularity as time-varying. In the inhomogeneous Poisson model proposed in


[6] each content is associated to a “pulse” whose duration reflects the content lifespan and
whose height denotes its instantaneous popularity–the model is called the Shot Noise Model
(SNM), mirroring the Poisson noise from electronics. While the shape of the pulse is not
important, the study observes strong correlations between popularity and duration; apparently
popular contents prosper longer. Finally, a class-based model [6] can conveniently capture spatio-
temporal correlations while allowing analytical tractability. Mobile users are especially keen
on downloading ephemeral content, thus it is expected that in case of wireless content the
improvement in modeling accuracy will be even greater.
Cache size [# of videos]

104

103
IRM fit
SNM fit
Original trace
102
0 0.05 0.1 0.15 0.2
Hit probability

Figure 2: (from [6]) Hit probability comparison between best fit of IRM, SNM and the YouTube
traces.

To optimize a cache one needs to track the changes in content popularity. For example, the
classical web caching systems adopt dynamic eviction policies like Least-Recently-Used (LRU)
in order to combat time-varying content popularity in a heuristic manner. However, the joint
consideration of popularity variations with wireless systems reveals a new challenge that renders
LRU policies inefficient. While a typical CDN cache normally receives 50 requests/content/day,
the corresponding figure for base station cache may be as low as 0.1 requests/content/day. With
such a small number of requests, fast variations of popularity become very difficult to be tracked

June 7, 2016 DRAFT


5

and classical LRU schemes fail.


This development motivates novel caching techniques that employ learning methodologies to
accurately track the evolution of content popularity over time. A recent study [7] analyzes
the SNM model and gives the optimal1 policy for joint caching and popularity estimation.
Additionally [8] proposes as an alternative solution the use of LRU with prefilters.

How to Track Popularity Variations

Since content popularity is time-varying, the caching operations can only be optimized if a
fresh view of the system is maintained. This requires massive data collection and processing,
and statistical inference from this data, which by itself is a complex task to handle. Additionally,
user privacy is a concern that can limit the potential of collecting such information. So can we
timely gather all this information in a wireless network?
Consider K users subscribed to the telecom operator and L caches placed in the network (e.g.,
in base stations or other locations). Each of these caches has capability of storing M contents
out of N contents in the catalog. Then, let the matrix P ∈ RK×N model the content access
statistics where rows are users and columns are contents. In other words, each entry (or rating)
in this matrix quantifies how popular is content n to user k. The popularity matrix P is large,
sparse, and only partially known in practice, and has to be continuously estimated in order to
enable the correct cache decisions at the base stations. At first, this seems to be an impossible
feat.
To deal with the complexity of handling matrix P, it is possible to use machine learning
tools to estimate the unknown entries. Such an estimation is particularly efficient when the
matrix has low spectral dimension, and the system can be described by a small number of
“features”; fortunately popularity correlation induces such behaviour in matrices obtained from
real data. For instance, low-rank matrix factorization methods, i.e., P ≈ KT N where K ∈ Rr×K
and N ∈ Rr×N are factor matrices, can be employed to construct the r-rank version of the
matrix, using the fact that users’ interests are correlated and predictable when r is small. This
additionally allows to store the collected statistics in a more compact way. As a result, a big

1
The optimality is established for the restricted case of homogeneous rectangularly-shaped pulses and asymptotically large
content catalogs.

June 7, 2016 DRAFT


6

data platform installed in the operator network can provide efficient collection and processing of
user access patterns from several locations, as evidenced in [9]. Further development of novel
machine learning tools, such as clustering techniques, are needed to improve the estimation of
the time-evolving content popularity matrix, i.e., Pl (t) for base station l, which may differ from
base station to base station.
It is worth noting that a caching system has requirements similar to those of a recommendation
system. For example, the well-known Netflix movie recommendation system exploits information
of user’s past activity in order to predict which movie is likely to be scored high by the user.
Similarly, a caching system exploits the request sequence to predict what contents are popular
enough to be cached. In this context, user privacy regulations may affect the collection of these
valuable data.
A key topic of research in this direction is privacy-preserving mechanisms that can enable
sufficient sampling of the time-evolving and location dependent popularity matrix Pl (t) with-
out compromising user privacy. Moreover, it is interesting to consider learning enhancement
approaches that transfer information to caching domain from other domains such as "transfer
learning" [10].

Security Is A Kind of Death

A common anti-caching argument relates to the operation of caching in a secure environment.


The secure counterpart of HTTP protocol, called HTTPS, was originally used to provide end-
to-end (e2e) encryption for securing sensitive information like online banking transactions and
authentication. Owing to the recent adoption from traffic giants Netflix and YouTube, the HTTPS
protocol is growing in numbers to soon exceed 50% of the total network traffic. Content
encryption poses an unsurmountable obstacle to in-network operations, including caching. Since
encrypting the data makes them unique and not reusable, caching, or even statistically process-
ing encrypted content is impossible. Ironically, the statement “security is a kind of death” of
Tennessee Williams seems to squarely apply to wireless caching.
Security is definitely a precious good everyone welcomes. Although securing a video stream
might seem as an excessive measure, in some cases it may be well justified. Unfortunately, e2e
encryption is clearly not in the Berners-Lee spirit since it prevents operators from optimizing
their networks and reanimates the server-client ghost of congestion, a reality that equally no

June 7, 2016 DRAFT


7

one can overlook. In fact, modern CDN systems resolve this issue by having “representatives”
of the content provider at the edge of the Internet. These representatives are trusted entities
which hold the user keys and are able to decrypt the requests and perform standard caching
operations. Ultimately, this methodology is neither entirely secure for the user [11], nor efficient
for the network [12], [13]. The need to make the system sustainable finally overrules the need
for e2e encryption, which is an argument against HTTPS for video delivery. Given however this
situation, how can we realistically push caching deeper into the wireless access?
Currently, content providers install their own caching boxes in the operator network and
intercept the related encrypted content requests deeper in the wireless access network. Examples
include the Google Global Cache [14], the Saguna solution [15], and the CacheBOX of Appli-
ansys [16]. In this approach, the boxes are not controlled by the operator which leads to several
limitations: (a) The caching boxes cannot perform complex tasks. (b) It is difficult to apply
learning techniques without context information from the operator. (c) The caching approach is
similar to the CDNs and therefore does not exploit the performance opportunities specific to
wireless caching, as we will discuss below.
New security protocols have been proposed to enable the operators to perform caching on
encrypted requests [13]. This leads to an interesting research direction, to combine the user
security and privacy with facilitation of the network management operations, which are crucial
for the sustainability of the future wireless systems.

III. T OWARDS A U NIFIED N ETWORK M EMORY

The proposition of Information Centric Networking (ICN) as a candidate for the future Internet
has also risen the subject of where to install network memory [17]. The ICN approach proposed to
equip routers with caches, and to allow content replication everywhere in the network. A recent
work [18] came up with striking conclusions about the ICN approach: most of the caching
benefits of ICN can be obtained by caching at the edges of the network using the existing
CDNs, and that any extra caching in the core network brings only negligible improvements at
very high costs. In the wireless domain, however, the question remains relevant: does it make
sense to cache even closer to the user than CDN?
It is commonly believed that caching is very inefficient near the user, and thus it should be
done at CDN. Below we explain the main reasons of inefficiency and argue that they can be

June 7, 2016 DRAFT


8

overcome.

Caching Deeper than CDN

Mitigating backhaul and wireless link overload requires going beyond CDN and caching at
the base stations and the mobile users. However, the efficient operation of such caches is very
challenging. In particular, there are two main challenges: (a) caches used in wireless networks
are typically small as compared to CDN caches, and (b) the popularity profile of traffic is highly
unpredictable when non-aggregated.
To understand point (a), consider that the effectiveness of a cache is measured with the hit
probability, i.e., the fraction of requests found in the cache. This can be upper bounded by the
popularity sum M
P
n=1 pn , where pn is the ordered popularity distribution with p1 denoting the

probability to request the most popular file. For power-law popularity the sum can further be
approximated by (M/N )1−α , where α < 1 is the power-law exponent. A very small ratio M/N
means that the hit probability becomes vanishingly small. For example, if we are caching Netflix
(12.5PB) in a mobile phone (10GB), M/N ∼ 10−6 , α ∼ 0.8 and the hit probability is less than
10%. However, base stations equipped with a disk array (40TB) can be extremely effective when
caching contents for a mobile VoD application. Table I provides some indicative numbers for
the memory types available and the catalogue sizes of reasonable applications.
In this context there are three promising research directions: (i) choose a part of the catalog to
cache while maintaining network neutrality [19], (ii) store only parts of the content using partial
caching techniques [20], and (iii) install massive memory at the edge in the form of small-sized
datacenters [21]. The third option will be realized by the fog computing paradigm [22].
To understand the unpredictable nature of sparse requests (formulated as challenge (b) above),
consider as an example the delivery of breaking e-news in a city served by a single CDN node.
Most users will download the news only once. The CDN system can quickly detect the rising
popularity of the news, since it will receive many requests in a short time frame. From the point
of view of a mobile user, however, the detection of the popularity of the trending news becomes
very difficult because the news is requested only once by the given user. This example shows that

2
The entire catalogue was anecdotally measured to contain 3.14PB of content in 2013, which however we multiply by 4 since
the same video is available in multiple formats.

June 7, 2016 DRAFT


9

Netflix catalogue (12.5PB)2 Torrents (1.5PB) Wireless VoD catalogue (1TB)

Disk 2TB ∼ 0.01% ∼ 0.1% 100%

Disk Array 40TB ∼ 0.3% ∼ 2% 100%

Data Center 150PB 50% 100% 100%

Table I: Typical data size values for normalized cache size M/N taken from the study of [8].
In practice it is anticipated that the wireless traffic is a 80-20 mix of torrent-like traffic and live
VoD traffic tailored to wireless device capabilities.

the detection efficiency depends on the number of requests aggregated at the popularity learner.
To illustrate this, Figure 3 shows the optimal hit probability in a hierarchy of L base stations.
Learning at the global CDN cache is shown to detect variations that are L times faster than
those at local caches. To remedy the situation it is possible to use an architecture that combines
information obtained at different aggregation layers [7].

cache with 0.7


global learning
0.6
content delivery
network
Hit probability
distance

0.5
l
ba
glo

central 0.4
router base station
controller
0.3
cache with
local learning
al
base
0.2 loc
station
0.1
10-5 10-3 1 101 103 105
Catalog refresh time

Figure 3: Optimal hit probability comparison between observing the aggregate request process at
the CDN-level (global), and observing the individual request process at each base station cache
(local), when refreshing the catalogue. The hit probability performance depends on how fast the
time-varying popularities can be learnt: global is faster than local.

June 7, 2016 DRAFT


10

Memory Is Cheap But Not Free

Although the cost of a small cache is dwarfed by the base station cost, the total amount
of installed memory in a mobile network can be considerable, therefore deciding to install
wireless caching requires a careful cost analysis [8]. To compute the optimal size of memory to
install at each location, one needs to know (a) the cost coefficients, (b) the skewness of content
popularity, and (c) the local traffic distribution in cells. Predicting how (a)-(b) will evolve is
quite challenging, but as in [8] a survey may determine a good set of parameters at any given
time.
For (c), the literature is extensively based on grid models, which in the case of future
wireless networks might be off for a significant factor. More accurate models have been recently
introduced from the field of stochastic geometry, where the cache-enabled base stations are
distributed according to a spatial point process (often chosen to be 2D Poisson), thus enabling
to handle the problem analytically. The validity of such modelling compared to regular cellular
models has been verified using extensive simulations. Additional insights for the deployment
of cache-enabled base stations can be obtained by analytically characterizing the performance
metrics, such as the outage probability and average delivery rate, for a given set of parameters
such as given number of base stations, storage size, skewness of the distribution, transmit power
and target Signal-to-Interference-plus-Noise Ratio (SINR) [23]. In addition to initially studied
single-tier network [23], the more detailed modeling and analysis of heterogeneous networks
[24], [25], online caching policies [26], uniform caching [27], power consumption [28], and a
Markovian mobility [29] are some recent examples in this direction. Therefore, although storage
units become increasingly cheaper, the question of how much storage we should place at each
location should be studied jointly with topological models from stochastic geometry, which are
both tractable and realistic. An example for such modelling and analysis of a network with
cache-enabled base stations is provided in Figure 4.

IV. W IRELESS 6= W IRED

Web caching has been traditionally studied by the networking community. A common miscon-
ception says that caching is a network layer technique, and hence the web caching approaches
are sufficient for wireless caching as well.

June 7, 2016 DRAFT


11

1
mobile 1.0 Target bitrate [nats/sec/hz]
user
0.8 0.4 0.2 0.1

Outage probability
0.8 Simulation
Unit distance

Theory
0.6
base
0.6
station
0.4
0.4
0.2
Voronoi
tesselation 0.2
0
0 0.2 0.4 0.6 0.8 1 0 2 4 4 8 10
Unit distance Cache size [nats]
(a) (b)

Figure 4: An example of base station deployment with caching capabilities [23]: (a) illustrates a
snapshot of 2D topology in which users and cache-enabled base stations are distributed according
to Poisson Point Processes (PPPs), (b) shows theoretical performance gains of such a deployment,
where the validation of results is done via simulations.

However, following the fundamental work of Maddah-Ali and Niesen [4], the idea of caching
has penetrated the information theory community with a new twist termed coded caching, which
promises unprecedented gains. In the following, we discuss the differences between wired and
wireless caching.

Wireless Caching Lies Both at Network and PHY Layers

Suppose that a base station wants to deliver information to K users at a rate of 1Mbps each,
i.e., for streaming a video. If the video is the same for all users (broadcast video) then this might
be possible for an arbitrarily large number of users. For example the base station could use an
omni-directional antenna, exploit the broadcast characteristic of wireless medium, and transmit
at 1Mbps to all users simultaneously. When the videos are different, this is clearly not possible:
the base station needs to multiplex the users over frequency, time, or codes, where each such
resource block is then associated to a single user. Since the resource blocks are finitely many,
ultimately the base station can serve 1Mbps videos up to a maximum number of users Kmax ,
after which the resources are exhausted. To increase Kmax , physical layer researchers propose
ways to increase the resource blocks of a given spectrum, i.e., increase spectral efficiency, or

June 7, 2016 DRAFT


12

install more base stations so there are more resource blocks per unit area, referred to as network
densification.
The novel paradigm of [4] shows a surprising fact: exploiting caching in a smart way, an
unbounded number of users watching different videos can be accomodated. How is this made
possible? During off-peak operation of the network, the users can cheaply populate their caches
with parts of popular contents. This is a perfectly reasonable assumption since the question of
sustainability that caching is trying to tackle refers to the hours of the day that the network
experiences the traffic peak. The content parts are appropriately chosen according to a caching
code which ensures symmetric properties. Then at request time, a coding technique called
“index coding” is employed, to minimize the number of transmissions to satisfy all users.3
The combination of these schemes is shown in [4] to yield required resource blocks equal to
K(1 − M/N )/(1 + KM/N ), where K is the number of users, M the cache size, and N the
catalog size. Hence, if the cacheable fraction of the catalog M/N is kept fixed, then the required
number of resource blocks does not increase with the number of users K, this can be verified
by taking the limit K → ∞ whereby the above quantity converges to a constant. The result is
pictorially summarized in Figure 5.
This promising idea has sprung a wealth of research efforts, some of them in parallel or right
after, such as device-to-device (D2D) networks [30], [31], non-uniform content popularities [32],
online caching policies [33], multi-servers [34], multi-library [35] and multihop wireless networks
with cooperative caching [36]. More recently, it has been shown that in order to achieve such
“order of K” gain over conventional unicast (with possible legacy uncoded caching) systems,
the content objects must be split into an O(exp(K)) number of subpackets; for networks of
practical size this gain is not achievable. The optimal tradeoff between coded caching gain and
content object size is a very interesting topic of current research.
From the implementation point of view, promising research directions include extensions to
capture system aspects such as (i) popularity skewness, (ii) asynchronous requests, (iii) content
objects of finite size and (iv) cache sizes that scale slower than N . Assuming that these practical
challenges are resolved, caching for wireless systems will become intertwined with physical

3
In fact, finding the optimal index code is a very difficult problem, and hence the proposed approach resorts to efficient
heuristics.

June 7, 2016 DRAFT


13

hing
coded cac
legacy un layer)

Required resource blocks


(network

[log scaling]
coded caching (PHY layer)

Users [K]

Figure 5: Required resource blocks for K mobile users with unicast demands, when the caches
fit 30% of the catalog. Coded caching can serve an arbitrarily large population of users with a
fixed number of resource blocks.

layer techniques employed at the base station and the handheld.

One Cache Analysis Is Not Sufficient

A contemporary mobile receives the signal of more than 10 base stations simultaneously. In
future densified cellular networks, the mobile will be connected to several femto-, pico-, or nano-
cells. The phenomenon of wireless multi-access opens a new horizon in caching exploitation [37].
Since a user can retrieve the requested content from many network endpoints, neighboring caches
should cooperate and avoid storing the same objects multiple times.
Content placement optimizations of wireless caching typically boil down to a set cover problem
in a bipartite graph connecting the users to the reachable caches. Therefore, finding what contents
to store at each cache is a difficult problem even if the popularities are assumed known [37].
It is possible to relax the problem to convex optimization by the use of distributed storage
codes, where each cache stores coded combinations of contents [37], or by obtaining a fractional
placement by time sharing different integral placements. These ideas lead to several interesting
algorithms in the literature of cooperative caching [21], [38]–[41].

June 7, 2016 DRAFT


14

What is the gain from these approaches? Cooperative caching typically saves space in the
cache by avoiding caching the same popular contents in neighboring caches. Equivalently we
may think of multiplying the cache size M by a small number, at best say a gain of 3-5. With
respect to hit probability, this can correspond to very different levels of gain, depending on the
value of M/N . Due to the skewness of the popularity distribution, marginal hit probability gain4
is high when M/N is small, and very small when M/N is large. Since in wireless we expect
the former, high gains are expected from cooperative wireless caching.
The current proposals on cooperative caching assume static popularity, and therefore a promis-
ing direction of research along these lines is to design caching schemes that combine cooperation
with learning of the time-varying popularity. The time to search and retrieve the content from a
nearby cache may also be significant, hence intelligent hash-based filtering and routing schemes
are required [42].

V. A S TAKEHOLDER A NALYSIS FOR W IRELESS C ACHING

The business of wireless caching involves three key stakeholders that together form a complex
ecosystem.
The users of telecommunication services are primarily the customers and the consumers of
the content, but in the case of wireless caching they are also active stakeholders. The users might
be requested to help in the form of contributing with their own resource, (for example in the case
of coded caching it will be memory and processing, or in D2D caching it will be also relaying
transmissions) and they will end up spending energy in the benefit of better performance. On the
other hand, one could envision the users employing D2D technology to enable caching without
the participation of other stakeholders. Due to the complexities mentioned above however,
efficient wireless caching will require heavy coordination and extensive monitoring/processing,
hence D2D approaches will be limited to restricted environments.
The operators of telecommunication networks are well placed for wireless caching. Due to the
particularities of coded caching and multi-access caching, the operators are in a unique position
to implement new protocols in base stations, affect the standards for new mobile devices, and
develop big data processing infrastructure that can realize wireless caching. Nevertheless, for

4
Gain obtained in hit probability when increasing M slightly.

June 7, 2016 DRAFT


15

reasons related to encryption, privacy, and global popularity estimation, the operators might not
be able to install these technologies without the cooperation of the other two stakeholders.
The providers of Internet content are champions of trust from the user community. Apart
from the security keys, they also hold extensive expertise in implementing caching techniques
in the core networks. From this advantageous position, they can positively affect the progressive
evolution of caching in wireless networks. On the other hand, content provider-only solutions
cannot unleash the full potential of wireless caching, since they are limited to alienated boxes
in the operator network that can perform caching only with legacy CDN techniques. The deeper
the caches go into the wireless network, the less efficient they will be if they stick to legacy
CDN techniques.

Users privacy
resources/energy
performance

access to
data Content
load Providers
reduction cheap
transport
hosting
data trust
collection standards content delivery
networks

Telecom What they need


Operators What they offer

Figure 6: A stakeholder analysis.

We summarize what each stakeholder offers and needs in Figure 6. What are the prospects of
the required collaboration among the stakeholders? Operators and content providers are seeking
the “best friends forever” union [43], [44] in order to mutually harvest benefits in the digital

June 7, 2016 DRAFT


16

value chain while keeping their users happy. This is a favorable environment for the scenario
of wireless caching. In fact, if the telecom operators enable caching capabilities at the edge of
their network, their infrastructure will become sustainable while they will gain access to new
business models. On the other hand, content providers can benefit from a caching collaboration
since (a) traffic will be intercepted earlier and the content transport cost will be reduced, (b) the
user demand will not be held back by sustainability issues, (c) costs associated to deployment of
large memory units will be avoided, (d) they will be able to reach closer to their users and extend
computing infrastructures to the fog paradigm. Last, it is foreseeable that in some situations the
role of the content provider and the wireless operator may converge.

R EFERENCES

[1] Cisco, “Cisco visual networking index: Global mobile data traffic forecast update 2014–2019,” White Paper, 2015.
[Online]. Available: http://goo.gl/tZ6QMk
[2] E. Baştuğ, M. Bennis, and M. Debbah, “Living on the edge: The role of proactive caching in 5G wireless networks,” IEEE
Communications Magazine, vol. 52, no. 8, pp. 82 – 89, August 2014.
[3] X. Wang, M. Chen, T. Taleb, A. Ksentini, and V. Leung, “Cache in the air: Exploiting content caching and delivery
techniques for 5G systems,” IEEE Communications Magazine, vol. 52, no. 2, pp. 131–139, February 2014.
[4] M. Maddah-Ali and U. Niesen, “Fundamental limits of caching,” IEEE Transactions on Information Theory, vol. 60, no. 5,
pp. 2856–2867, May 2014.
[5] L. Breslau, P. Cao, L. Fan, G. Phillips, and S. Shenker, “Web caching and Zipf-like distributions: evidence and implications,”
in IEEE INFOCOM, Mar 1999.
[6] S. Traverso, M. Ahmed, M. Garetto, P. Giaccone, E. Leonardi, and S. Niccolini, “Temporal locality in today’s content
caching: Why it matters and how to model it,” ACM SIGCOMM, 2013.
[7] M. Leconte, G. S. Paschos, L. Gkatzikis, M. Draief, S. Vassilaras, and S. Chouvardas, “Placing dynamic content in caches
with small population,” in IEEE INFOCOM, 2016.
[8] S.-E. Elayoubi and J. Roberts, “Performance and cost effectiveness of caching in mobile access networks,” in Proceedings
of the 2Nd International Conference on Information-Centric Networking, ser. ICN ’15. ACM, 2015, pp. 79–88.
[9] E. Baştuğ, M. Bennis, E. Zeydan, M. A. Kader, A. Karatepe, A. S. Er, and M. Debbah, “Big data meets telcos: A proactive
caching perspective,” Journal of Communications and Networks, Special Issue on Big Data Networking-Challenges and
Applications, vol. 17, no. 6, pp. 549–558, December 2015.
[10] E. Baştuğ, M. Bennis, and M. Debbah, “A transfer learning approach for cache-enabled wireless networks,” in International
Symposium on Modeling and Optimization in Mobile, Ad Hoc and Wireless Networks (WiOpt’15) - D2D Workshop, Mumbai,
India, May 2015.
[11] X. d. C. de Carnavalet and M. Mannan, “Killed by proxy: Analyzing client-end TLS interception software,” 2016.
[12] J. Maisonneuve, V. K. Gurbani, and T. Fossati, “The security pendulum,” in Managing Radio Networks in an Encrypted
World (MaRNEW) Workshop, 2015. [Online]. Available: https://goo.gl/jP2TYv

June 7, 2016 DRAFT


17

[13] D. Naylor, K. Schomp, M. Varvello, I. Leontiadis, J. Blackburn, D. R. López, K. Papagiannaki, P. Rodriguez, and
P. Steenkiste, “Multi-context TLS (mcTLS): Enabling secure in-network functionality in TLS,” in ACM SIGCOMM, 2015.
[14] Google global cache (ggc). [Online]. Available: https://goo.gl/VHUINF
[15] Saguna. [Online]. Available: http://goo.gl/0wkgrJ
[16] The appliansys cachebox. [Online]. Available: http://goo.gl/jbW82d
[17] J. Roberts and N. Sbihi, “Exploring the memory-bandwidth tradeoff in an information-centric network,” in ITC, 2013.
[18] S. K. Fayazbakhsh, Y. Lin, A. Tootoonchian, A. Ghodsi, T. Koponen, B. Maggs, K. Ng, V. Sekar, and S. Shenker, “Less
pain, most of the gain: Incrementally deployable ICN,” in ACM SIGCOMM, 2013.
[19] F. Kocak, G. Kesidis, and S. Fdida, “Network neutrality with content caching and its effect on access pricing,” Smart Data
Pricing, pp. 47–66, 2013.
[20] L. Maggi, L. Gkatzikis, G. Paschos, and J. Leguay, “Adapting caching to audience retention rate: Which video chunk to
store?” arXiv preprint arXiv:1512.03274, 2015.
[21] V. Bioglio, F. Gabry, and I. Land, “Optimizing MDS codes for caching at the edge,” in IEEE Global Communications
Conference (GLOBECOM’15), San Diego, USA, December 2015.
[22] F. Bonomi, R. Milito, J. Zhu, and S. Addepalli, “Fog computing and its role in the internet of things,” in Proceedings of
the first edition of the MCC workshop on Mobile cloud computing. ACM, 2012, pp. 13–16.
[23] E. Baştuğ, M. Bennis, M. Kountouris, and M. Debbah, “Cache-enabled small cell networks: Modeling and tradeoffs,”
EURASIP Journal on Wireless Communications and Networking, no. 1, p. 41, February 2015.
[24] C. Yang, Y. Yao, Z. Chen, and B. Xia, “Analysis on cache-enabled wireless heterogeneous networks,” IEEE Transactions
on Wireless Communications, vol. PP, no. 99, pp. 1–1, 2015.
[25] Z. Chen, J. Lee, T. Q. Quek, and M. Kountouris, “Cooperative caching and transmission design in cluster-centric small
cell networks,” arXiv preprint arXiv:1601.00321, 2016.
[26] S. A. R. Zaidi, M. Ghogho, and D. C. McLernon, “Information centric modeling for two-tier cache enabled cellular
networks,” in IEEE International Conference on Communications Workshop (ICC Workshop), 2015, pp. 80–86.
[27] B. Blaszczyszyn and A. Giovanidis, “Optimal geographic caching in cellular networks,” arXiv preprint arXiv:1409.7626,
2014.
[28] B. Perabathini, E. Baştuğ, M. Kountouris, M. Debbah, and A. Conte, “Caching at the edge: a green perspective for 5G
networks,” in IEEE International Conference on Communications Workshop (ICC’15 Workshop), London, UK, June 2015.
[29] K. Poularakis and L. Tassiulas, “Exploiting user mobility for wireless content delivery,” in IEEE International Symposium
on Information Theory Proceedings (ISIT). IEEE, July 2013, pp. 1017–1021.
[30] M. Ji, G. Caire, and A. F. Molisch, “Fundamental limits of caching in wireless D2D networks,” arXiv preprint arXiv:
1405.5336, 2014.
[31] S.-W. Jeon, S.-N. Hong, M. Ji, and G. Caire, “On the capacity of multihop device-to-device caching networks,” in IEEE
Information Theory Workshop (ITW). IEEE, 2015, pp. 1–5.
[32] M. Ji, A. M. Tulino, J. Llorca, and G. Caire, “Order-optimal rate of caching and coded multicasting with random demands,”
arXiv preprint arXiv: 1502.03124, 2015.
[33] R. Pedarsani, M. A. Maddah-Ali, and U. Niesen, “Online coded caching,” arXiv preprint arXiv: 1311.3646, 2013.
[34] S. P. Shariatpanahi, S. A. Motahari, and B. H. Khalaj, “Multi-server coded caching,” arXiv preprint arXiv: 1503.00265,
2015.
[35] S. Sahraei and M. Gastpar, “Multi-library coded caching,” arXiv preprint arXiv:1601.06016, 2016.

June 7, 2016 DRAFT


18

[36] S. Gitzenis, G. Paschos, and L. Tassiulas, “Asymptotic laws for joint content replication and delivery in wireless networks,”
IEEE Transactions on Information Theory, vol. 59, no. 5, pp. 2760–2776, May 2013.
[37] N. Golrezaei, K. Shanmugam, A. Dimakis, A. Molisch, and G. Caire, “FemtoCaching: Wireless video content delivery
through distributed caching helpers,” IEEE Transactions on Information Theory, vol. 59, no. 12, pp. 8402–8413, 2013.
[38] K. Poularakis, G. Iosifidis, and L. Tassiulas, “Approximation algorithms for mobile data caching in small cell networks,”
IEEE Transactions on Communications, vol. 62, no. 10, pp. 3665–3677, 2014.
[39] K. Naveen, L. Massoulie, E. Baccelli, A. Carneiro Viana, and D. Towsley, “On the interaction between content caching and
request assignment in cellular cache networks,” in Proceedings of the 5th Workshop on All Things Cellular: Operations,
Applications and Challenges. ACM, 2015, pp. 37–42.
[40] J. Zhang and P. Elia, “Fundamental limits of cache-aided wireless BC: Interplay of coded-caching and CSIT feedback,”
arXiv preprint arXiv: 1511.03961, 2015.
[41] M. Dehghan, A. Seetharam, B. Jiang, T. He, T. Salonidis, J. Kurose, D. Towsley, and R. Sitaraman, “On the complexity
of optimal routing and content caching in heterogeneous networks,” in IEEE INFOCOM, 2015.
[42] M. Tao, E. Chen, H. Zhou, and W. Yu, “Content-centric sparse multicast beamforming for cache-enabled cloud RAN,”
arXiv preprint arXiv: 1512.06938, 2015.
[43] D. Thomas and P. Taylor, “Content providers and operators at peace,” Financial Times, 2012. [Online]. Available:
http://goo.gl/6StCVk
[44] Openet, “Content and OTT partnerships – the key to unlocking new business models,” White Paper, 2014. [Online].
Available: http://goo.gl/BVZra6

June 7, 2016 DRAFT

You might also like