PA-Cache: Evolving Learning-Based Popularity-Aware Content Caching in
Edge Networks
- URL: http://arxiv.org/abs/2002.08805v2
- Date: Thu, 10 Dec 2020 02:54:51 GMT
- Title: PA-Cache: Evolving Learning-Based Popularity-Aware Content Caching in
Edge Networks
- Authors: Qilin Fan, Xiuhua Li, Jian Li, Qiang He, Kai Wang, Junhao Wen
- Abstract summary: We propose an evolving learning-based content caching policy, named PA-Cache in edge networks.
It adaptively learns time-varying content popularity and determines which contents should be replaced when the cache is full.
We extensively evaluate the performance of our proposed PA-Cache on real-world traces from a large online video-on-demand service provider.
- Score: 14.939950326112045
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As ubiquitous and personalized services are growing boomingly, an
increasingly large amount of traffic is generated over the network by massive
mobile devices. As a result, content caching is gradually extending to network
edges to provide low-latency services, improve quality of service, and reduce
redundant data traffic. Compared to the conventional content delivery networks,
caches in edge networks with smaller sizes usually have to accommodate more
bursty requests. In this paper, we propose an evolving learning-based content
caching policy, named PA-Cache in edge networks. It adaptively learns
time-varying content popularity and determines which contents should be
replaced when the cache is full. Unlike conventional deep neural networks
(DNNs), which learn a fine-tuned but possibly outdated or biased prediction
model using the entire training dataset with high computational complexity,
PA-Cache weighs a large set of content features and trains the multi-layer
recurrent neural network from shallow to deeper when more requests arrive over
time. We extensively evaluate the performance of our proposed PA-Cache on
real-world traces from a large online video-on-demand service provider. \rb{The
results show that PA-Cache outperforms existing popular caching algorithms and
approximates the optimal algorithm with only a 3.8\% performance gap when the
cache percentage is 1.0\%}. PA-Cache also significantly reduces the
computational cost compared to conventional DNN-based approaches.
Related papers
- Efficient Inference of Vision Instruction-Following Models with Elastic Cache [76.44955111634545]
We introduce Elastic Cache, a novel strategy for efficient deployment of instruction-following large vision-language models.
We propose an importance-driven cache merging strategy to prune redundancy caches.
For instruction encoding, we utilize the frequency to evaluate the importance of caches.
Results on a range of LVLMs demonstrate that Elastic Cache not only boosts efficiency but also notably outperforms existing pruning methods in language generation.
arXiv Detail & Related papers (2024-07-25T15:29:05Z) - A Learning-Based Caching Mechanism for Edge Content Delivery [2.412158290827225]
5G networks and the rise of the Internet of Things (IoT) are increasingly extending into the network edge.
This shift introduces unique challenges, particularly due to the limited cache storage and the diverse request patterns at the edge.
We introduce HR-Cache, a learning-based caching framework grounded in the principles of Hazard Rate (HR) ordering.
arXiv Detail & Related papers (2024-02-05T08:06:03Z) - Predictive Edge Caching through Deep Mining of Sequential Patterns in
User Content Retrievals [34.716416311132946]
We propose a novel Predictive Edge Caching (PEC) system that predicts the future content popularity using fine-grained learning models.
PEC can adapt to highly dynamic content popularity, and significantly improve cache hit ratio and reduce user content retrieval latency.
arXiv Detail & Related papers (2022-10-06T03:24:19Z) - Accelerating Deep Learning Classification with Error-controlled
Approximate-key Caching [72.50506500576746]
We propose a novel caching paradigm, that we named approximate-key caching.
While approximate cache hits alleviate DL inference workload and increase the system throughput, they however introduce an approximation error.
We analytically model our caching system performance for classic LRU and ideal caches, we perform a trace-driven evaluation of the expected performance, and we compare the benefits of our proposed approach with the state-of-the-art similarity caching.
arXiv Detail & Related papers (2021-12-13T13:49:11Z) - TEDGE-Caching: Transformer-based Edge Caching Towards 6G Networks [30.160404936777947]
Mobile Edge Caching (MEC) in the 6G networks has been evolved as an efficient solution to meet the phenomenal growth of the global mobile data traffic.
Recent advancements in Deep Neural Networks (DNNs) have drawn much research attention to predict the content popularity in proactive caching schemes.
We propose an edge caching framework incorporated with the attention-based Vision Transformer (ViT) neural network, referred to as the Transformer-based Edge (TEDGE) caching.
arXiv Detail & Related papers (2021-12-01T16:38:18Z) - Compact representations of convolutional neural networks via weight
pruning and quantization [63.417651529192014]
We propose a novel storage format for convolutional neural networks (CNNs) based on source coding and leveraging both weight pruning and quantization.
We achieve a reduction of space occupancy up to 0.6% on fully connected layers and 5.44% on the whole network, while performing at least as competitive as the baseline.
arXiv Detail & Related papers (2021-08-28T20:39:54Z) - Learning from Images: Proactive Caching with Parallel Convolutional
Neural Networks [94.85780721466816]
A novel framework for proactive caching is proposed in this paper.
It combines model-based optimization with data-driven techniques by transforming an optimization problem into a grayscale image.
Numerical results show that the proposed scheme can reduce 71.6% computation time with only 0.8% additional performance cost.
arXiv Detail & Related papers (2021-08-15T21:32:47Z) - A Survey of Deep Learning for Data Caching in Edge Network [1.9798034349981157]
This paper summarizes the utilization of deep learning for data caching in edge network.
We first outline the typical research topics in content caching and formulate a taxonomy based on network hierarchical structure.
Then, a number of key types of deep learning algorithms are presented, ranging from supervised learning to unsupervised learning as well as reinforcement learning.
arXiv Detail & Related papers (2020-08-17T12:02:32Z) - Caching Placement and Resource Allocation for Cache-Enabling UAV NOMA
Networks [87.6031308969681]
This article investigates the cache-enabling unmanned aerial vehicle (UAV) cellular networks with massive access capability supported by non-orthogonal multiple access (NOMA)
We formulate the long-term caching placement and resource allocation optimization problem for content delivery delay minimization as a Markov decision process (MDP)
We propose a Q-learning based caching placement and resource allocation algorithm, where the UAV learns and selects action with emphsoft $varepsilon$-greedy strategy to search for the optimal match between actions and states.
arXiv Detail & Related papers (2020-08-12T08:33:51Z) - Reinforcement Learning for Caching with Space-Time Popularity Dynamics [61.55827760294755]
caching is envisioned to play a critical role in next-generation networks.
To intelligently prefetch and store contents, a cache node should be able to learn what and when to cache.
This chapter presents a versatile reinforcement learning based approach for near-optimal caching policy design.
arXiv Detail & Related papers (2020-05-19T01:23:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.