Emergent Mind

Wireless Multihop Device-to-Device Caching Networks

(1511.02574)
Published Nov 9, 2015 in cs.IT and math.IT

Abstract

We consider a wireless device-to-device (D2D) network where $n$ nodes are uniformly distributed at random over the network area. We let each node with storage capacity $M$ cache files from a library of size $m \geq M$. Each node in the network requests a file from the library independently at random, according to a popularity distribution, and is served by other nodes having the requested file in their local cache via (possibly) multihop transmissions. Under the classical "protocol model" of wireless networks, we characterize the optimal per-node capacity scaling law for a broad class of heavy-tailed popularity distributions including Zipf distributions with exponent less than one. In the parameter regimes of interest, we show that a decentralized random caching strategy with uniform probability over the library yields the optimal per-node capacity scaling of $\Theta(\sqrt{M/m})$, which is constant with $n$, thus yielding throughput scalability with the network size. Furthermore, the multihop capacity scaling can be significantly better than for the case of single-hop caching networks, for which the per-node capacity is $\Theta(M/m)$. The multihop capacity scaling law can be further improved for a Zipf distribution with exponent larger than some threshold $> 1$, by using a decentralized random caching uniformly across a subset of most popular files in the library. Namely, ignoring a subset of less popular files (i.e., effectively reducing the size of the library) can significantly improve the throughput scaling while guaranteeing that all nodes will be served with high probability as $n$ increases.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.