Metalearning: Sparse Variable-Structure Automata
- URL: http://arxiv.org/abs/2102.00315v1
- Date: Sat, 30 Jan 2021 21:32:23 GMT
- Title: Metalearning: Sparse Variable-Structure Automata
- Authors: Pedram Fekri, Ali Akbar Safavi, Mehrdad Hosseini Zadeh, and Peyman
Setoodeh
- Abstract summary: We propose a metalearning approach to increase the number of basis vectors used in dynamic sparse coding vectors on the fly.
An actor-critic algorithm is deployed to automatically choose an appropriate dimension for feature regarding the required level of accuracy.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dimension of the encoder output (i.e., the code layer) in an autoencoder is a
key hyper-parameter for representing the input data in a proper space. This
dimension must be carefully selected in order to guarantee the desired
reconstruction accuracy. Although overcomplete representation can address this
dimension issue, the computational complexity will increase with dimension.
Inspired by non-parametric methods, here, we propose a metalearning approach to
increase the number of basis vectors used in dynamic sparse coding on the fly.
An actor-critic algorithm is deployed to automatically choose an appropriate
dimension for feature vectors regarding the required level of accuracy. The
proposed method benefits from online dictionary learning and fast iterative
shrinkage-thresholding algorithm (FISTA) as the optimizer in the inference
phase. It aims at choosing the minimum number of bases for the overcomplete
representation regarding the reconstruction error threshold. This method allows
for online controlling of both the representation dimension and the
reconstruction error in a dynamic framework.
Related papers
- An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - Optimal Projections for Discriminative Dictionary Learning using the JL-lemma [0.5461938536945723]
Dimensionality reduction-based dictionary learning methods have often used iterative random projections.
This paper proposes a constructive approach to derandomize the projection matrix using the Johnson-Lindenstrauss lemma.
arXiv Detail & Related papers (2023-08-27T02:59:59Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Deep Identification of Nonlinear Systems in Koopman Form [0.0]
The present paper treats the identification of nonlinear dynamical systems using Koopman-based deep state-space encoders.
An input-affine formulation is considered for the lifted model structure and we address both full and partial state availability.
arXiv Detail & Related papers (2021-10-06T08:50:56Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - Dynamic Convolution for 3D Point Cloud Instance Segmentation [146.7971476424351]
We propose an approach to instance segmentation from 3D point clouds based on dynamic convolution.
We gather homogeneous points that have identical semantic categories and close votes for the geometric centroids.
The proposed approach is proposal-free, and instead exploits a convolution process that adapts to the spatial and semantic characteristics of each instance.
arXiv Detail & Related papers (2021-07-18T09:05:16Z) - Generalized Learning Vector Quantization for Classification in
Randomized Neural Networks and Hyperdimensional Computing [4.4886210896619945]
We propose a modified RVFL network that avoids computationally expensive matrix operations during training.
The proposed approach achieved state-of-the-art accuracy on a collection of datasets from the UCI Machine Learning Repository.
arXiv Detail & Related papers (2021-06-17T21:17:17Z) - Non-linear, Sparse Dimensionality Reduction via Path Lasso Penalized
Autoencoders [0.0]
We present path lasso penalized autoencoders for complex data structures.
Our algorithm uses a group lasso penalty and non-negative matrix factorization to construct a sparse, non-linear latent representation.
We show that the algorithm exhibits much lower reconstruction errors than sparse PCA and parameter-wise lasso regularized autoencoders for low-dimensional representations.
arXiv Detail & Related papers (2021-02-22T10:14:46Z) - Kaleidoscope: An Efficient, Learnable Representation For All Structured
Linear Maps [20.151950843660973]
We introduce kaleidoscope matrices (K-matrices) that provably capture any structured matrix with near-optimal space.
K-matrices can be automatically learned within end-to-end pipelines to replace hand-crafted procedures.
We use K-matrices in a Transformer network to attain 36% faster end-to-end inference speed on a language translation task.
arXiv Detail & Related papers (2020-12-29T22:51:29Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - MetaSDF: Meta-learning Signed Distance Functions [85.81290552559817]
Generalizing across shapes with neural implicit representations amounts to learning priors over the respective function space.
We formalize learning of a shape space as a meta-learning problem and leverage gradient-based meta-learning algorithms to solve this task.
arXiv Detail & Related papers (2020-06-17T05:14:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.