Learning Centric Power Allocation for Edge Intelligence
- URL: http://arxiv.org/abs/2007.11399v1
- Date: Tue, 21 Jul 2020 07:02:07 GMT
- Title: Learning Centric Power Allocation for Edge Intelligence
- Authors: Shuai Wang, Rui Wang, Qi Hao, Yik-Chung Wu, and H. Vincent Poor
- Abstract summary: Edge intelligence has been proposed, which collects distributed data and performs machine learning at the edge.
This paper proposes a learning centric power allocation (LCPA) method, which allocates radio resources based on an empirical classification error model.
Experimental results show that the proposed LCPA algorithm significantly outperforms other power allocation algorithms.
- Score: 84.16832516799289
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While machine-type communication (MTC) devices generate massive data, they
often cannot process this data due to limited energy and computation power. To
this end, edge intelligence has been proposed, which collects distributed data
and performs machine learning at the edge. However, this paradigm needs to
maximize the learning performance instead of the communication throughput, for
which the celebrated water-filling and max-min fairness algorithms become
inefficient since they allocate resources merely according to the quality of
wireless channels. This paper proposes a learning centric power allocation
(LCPA) method, which allocates radio resources based on an empirical
classification error model. To get insights into LCPA, an asymptotic optimal
solution is derived. The solution shows that the transmit powers are inversely
proportional to the channel gain, and scale exponentially with the learning
parameters. Experimental results show that the proposed LCPA algorithm
significantly outperforms other power allocation algorithms.
Related papers
- RESIST: Resilient Decentralized Learning Using Consensus Gradient Descent [11.22833419439317]
Empirical robustness risk (ERM) is a cornerstone of modern machine learning (ML)
This paper focuses on the man-in-the-middle (MITM) attack, which can cause models to deviate significantly from their intended ERM solutions.
We propose RESIST, an algorithm designed to be robust against adversarially compromised communication links.
arXiv Detail & Related papers (2025-02-11T21:48:10Z) - Learning for Cross-Layer Resource Allocation in MEC-Aided Cell-Free Networks [71.30914500714262]
Cross-layer resource allocation over mobile edge computing (MEC)-aided cell-free networks can sufficiently exploit the transmitting and computing resources to promote the data rate.
Joint subcarrier allocation and beamforming optimization are investigated for the MEC-aided cell-free network from the perspective of deep learning.
arXiv Detail & Related papers (2024-12-21T10:18:55Z) - Sampling Strategies for Static Powergrid Models [0.0]
Power flow calculation is an iterative method to compute the voltage magnitudes of the power grid's buses from power values.
Machine learning and, especially, artificial neural networks were successfully used as surrogates for the power flow calculation.
arXiv Detail & Related papers (2022-04-19T11:38:07Z) - Resource-constrained Federated Edge Learning with Heterogeneous Data:
Formulation and Analysis [8.863089484787835]
We propose a distributed approximate Newton-type Newton-type training scheme, namely FedOVA, to solve the heterogeneous statistical challenge brought by heterogeneous data.
FedOVA decomposes a multi-class classification problem into more straightforward binary classification problems and then combines their respective outputs using ensemble learning.
arXiv Detail & Related papers (2021-10-14T17:35:24Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Reconfigurable Intelligent Surface Assisted Mobile Edge Computing with
Heterogeneous Learning Tasks [53.1636151439562]
Mobile edge computing (MEC) provides a natural platform for AI applications.
We present an infrastructure to perform machine learning tasks at an MEC with the assistance of a reconfigurable intelligent surface (RIS)
Specifically, we minimize the learning error of all participating users by jointly optimizing transmit power of mobile users, beamforming vectors of the base station, and the phase-shift matrix of the RIS.
arXiv Detail & Related papers (2020-12-25T07:08:50Z) - Edge Intelligence for Energy-efficient Computation Offloading and
Resource Allocation in 5G Beyond [7.953533529450216]
5G beyond is an end-edge-cloud orchestrated network that can exploit heterogeneous capabilities of the end devices, edge servers, and the cloud.
In multi user wireless networks, diverse application requirements and the possibility of various radio access modes for communication among devices make it challenging to design an optimal computation offloading scheme.
Deep Reinforcement Learning (DRL) is an emerging technique to address such an issue with limited and less accurate network information.
arXiv Detail & Related papers (2020-11-17T05:51:03Z) - Learning Centric Wireless Resource Allocation for Edge Computing:
Algorithm and Experiment [15.577056429740951]
Edge intelligence is an emerging network architecture that integrates sensing, communication, computing components, and supports various machine learning applications.
Existing methods ignore two important facts: 1) different models have heterogeneous demands on training data; 2) there is a mismatch between the simulated environment and the real-world environment.
This paper proposes the learning centric wireless resource allocation scheme that maximizes the worst learning performance of multiple tasks.
arXiv Detail & Related papers (2020-10-29T06:20:40Z) - Resource Allocation via Model-Free Deep Learning in Free Space Optical
Communications [119.81868223344173]
The paper investigates the general problem of resource allocation for mitigating channel fading effects in Free Space Optical (FSO) communications.
Under this framework, we propose two algorithms that solve FSO resource allocation problems.
arXiv Detail & Related papers (2020-07-27T17:38:51Z) - A Machine Learning Approach for Task and Resource Allocation in Mobile
Edge Computing Based Networks [108.57859531628264]
A joint task, spectrum, and transmit power allocation problem is investigated for a wireless network.
The proposed algorithm can reduce the number of iterations needed for convergence and the maximal delay among all users by up to 18% and 11.1% compared to the standard Q-learning algorithm.
arXiv Detail & Related papers (2020-07-20T13:46:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.