A sparse code increases the speed and efficiency of neuro-dynamic
programming for optimal control tasks with correlated inputs
- URL: http://arxiv.org/abs/2006.11968v3
- Date: Wed, 18 Nov 2020 01:10:23 GMT
- Title: A sparse code increases the speed and efficiency of neuro-dynamic
programming for optimal control tasks with correlated inputs
- Authors: Peter N. Loxley
- Abstract summary: A sparse code is used to represent natural images in an optimal control task solved with neuro-dynamic programming.
A 2.25 times over-complete sparse code is shown to at least double memory capacity compared with a complete sparse code using the same input.
This is used in sequential learning to store a potentially large number of optimal control tasks in the network.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sparse codes in neuroscience have been suggested to offer certain
computational advantages over other neural representations of sensory data. To
explore this viewpoint, a sparse code is used to represent natural images in an
optimal control task solved with neuro-dynamic programming, and its
computational properties are investigated. The central finding is that when
feature inputs to a linear network are correlated, an over-complete sparse code
increases the memory capacity of the network in an efficient manner beyond that
possible for any complete code with the same-sized input, and also increases
the speed of learning the network weights. A complete sparse code is found to
maximise the memory capacity of a linear network by decorrelating its feature
inputs to transform the design matrix of the least-squares problem to one of
full rank. It also conditions the Hessian matrix of the least-squares problem,
thereby increasing the rate of convergence to the optimal network weights.
Other types of decorrelating codes would also achieve this. However, an
over-complete sparse code is found to be approximately decorrelated, extracting
a larger number of approximately decorrelated features from the same-sized
input, allowing it to efficiently increase memory capacity beyond that possible
for any complete code: a 2.25 times over-complete sparse code is shown to at
least double memory capacity compared with a complete sparse code using the
same input. This is used in sequential learning to store a potentially large
number of optimal control tasks in the network, while catastrophic forgetting
is avoided using a partitioned representation, yielding a cost-to-go function
approximator that generalizes over the states in each partition. Sparse code
advantages over dense codes and local codes are also discussed.
Related papers
- Erasure Coded Neural Network Inference via Fisher Averaging [28.243239815823205]
Erasure-coded computing has been successfully used in cloud systems to reduce tail latency caused by factors such as straggling servers and heterogeneous traffic variations.
We design a method to code over neural networks, that is, given two or more neural network models, how to construct a coded model whose output is a linear combination of the outputs of the given neural networks.
We conduct experiments to perform erasure coding over neural networks trained on real-world vision datasets and show that the accuracy of the decoded outputs using COIN is significantly higher than other baselines.
arXiv Detail & Related papers (2024-09-02T18:46:26Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - Instant Neural Graphics Primitives with a Multiresolution Hash Encoding [67.33850633281803]
We present a versatile new input encoding that permits the use of a smaller network without sacrificing quality.
A small neural network is augmented by a multiresolution hash table of trainable feature vectors whose values are optimized through a gradient descent.
We achieve a combined speed of several orders of magnitude, enabling training of high-quality neural graphics primitives in a matter of seconds.
arXiv Detail & Related papers (2022-01-16T07:22:47Z) - Efficient and Robust Mixed-Integer Optimization Methods for Training
Binarized Deep Neural Networks [0.07614628596146598]
We study deep neural networks with binary activation functions and continuous or integer weights (BDNN)
We show that the BDNN can be reformulated as a mixed-integer linear program with bounded weight space which can be solved to global optimality by classical mixed-integer programming solvers.
For the first time a robust model is presented which enforces robustness of the BDNN during training.
arXiv Detail & Related papers (2021-10-21T18:02:58Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - Efficient bit encoding of neural networks for Fock states [77.34726150561087]
The complexity of the neural network scales only with the number of bit-encoded neurons rather than the maximum boson number.
In the high occupation regime its information compression efficiency is shown to surpass even maximally optimized density matrix implementations.
arXiv Detail & Related papers (2021-03-15T11:24:40Z) - Neural network approaches to point lattice decoding [6.025026882312586]
Voronoi-reduced basis is introduced to restrict the space of solutions to a binary set.
We count the number of affine pieces in the CPWL decoding function to characterize the complexity of the decoding problem.
arXiv Detail & Related papers (2020-12-13T10:53:34Z) - Pruning Neural Belief Propagation Decoders [77.237958592189]
We introduce a method to tailor an overcomplete parity-check matrix to (neural) BP decoding using machine learning.
We achieve performance within 0.27 dB and 1.5 dB of the ML performance while reducing the complexity of the decoder.
arXiv Detail & Related papers (2020-01-21T12:05:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.