Solving Large-scale Spatial Problems with Convolutional Neural Networks
- URL: http://arxiv.org/abs/2306.08191v2
- Date: Wed, 7 Feb 2024 18:18:54 GMT
- Title: Solving Large-scale Spatial Problems with Convolutional Neural Networks
- Authors: Damian Owerko, Charilaos I. Kanatsoulis, Alejandro Ribeiro
- Abstract summary: We employ transfer learning to improve training efficiency for large-scale spatial problems.
We propose that a convolutional neural network (CNN) can be trained on small windows of signals, but evaluated on arbitrarily large signals with little to no performance degradation.
- Score: 88.31876586547848
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the past decade, deep learning research has been accelerated by
increasingly powerful hardware, which facilitated rapid growth in the model
complexity and the amount of data ingested. This is becoming unsustainable and
therefore refocusing on efficiency is necessary. In this paper, we employ
transfer learning to improve training efficiency for large-scale spatial
problems. We propose that a convolutional neural network (CNN) can be trained
on small windows of signals, but evaluated on arbitrarily large signals with
little to no performance degradation, and provide a theoretical bound on the
resulting generalization error. Our proof leverages shift-equivariance of CNNs,
a property that is underexploited in transfer learning. The theoretical results
are experimentally supported in the context of mobile infrastructure on demand
(MID). The proposed approach is able to tackle MID at large scales with
hundreds of agents, which was computationally intractable prior to this work.
Related papers
- Apply Distributed CNN on Genomics to accelerate Transcription-Factor TAL1 Motif Prediction [0.0]
We highlight the potential of deep learning in the field of genomics and its challenges such as the training time that takes hours, weeks, and in some cases months.
We propose to apply a distributed deep learning implementation based on Convolutional Neural Networks (CNN) that showed good results in decreasing the training time.
We proved the efficiency of using a distributed strategy based on data-parallelism in predicting the transcription-factor TAL1 motif faster.
arXiv Detail & Related papers (2024-05-25T07:09:44Z) - Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - Decouple Graph Neural Networks: Train Multiple Simple GNNs Simultaneously Instead of One [60.5818387068983]
Graph neural networks (GNN) suffer from severe inefficiency.
We propose to decouple a multi-layer GNN as multiple simple modules for more efficient training.
We show that the proposed framework is highly efficient with reasonable performance.
arXiv Detail & Related papers (2023-04-20T07:21:32Z) - A Principle of Least Action for the Training of Neural Networks [10.342408668490975]
We show the presence of a low kinetic energy displacement bias in the transport map of the network, and link this bias with generalization performance.
We propose a new learning algorithm, which automatically adapts to the complexity of the given task, and leads to networks with a high generalization ability even in low data regimes.
arXiv Detail & Related papers (2020-09-17T15:37:34Z) - Deep Learning for Ultra-Reliable and Low-Latency Communications in 6G
Networks [84.2155885234293]
We first summarize how to apply data-driven supervised deep learning and deep reinforcement learning in URLLC.
To address these open problems, we develop a multi-level architecture that enables device intelligence, edge intelligence, and cloud intelligence for URLLC.
arXiv Detail & Related papers (2020-02-22T14:38:11Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z) - Model-Driven Beamforming Neural Networks [47.754731555563836]
This article introduces general data- and model-driven beamforming neural networks (BNNs)
It presents various possible learning strategies, and also discusses complexity reduction for the DL-based BNNs.
We also offer enhancement methods such as training-set augmentation and transfer learning in order to improve the generality of BNNs.
arXiv Detail & Related papers (2020-01-15T12:50:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.