Multi-Model Least Squares-Based Recomputation Framework for Large Data
Analysis
- URL: http://arxiv.org/abs/2101.01271v4
- Date: Wed, 3 Mar 2021 16:31:59 GMT
- Title: Multi-Model Least Squares-Based Recomputation Framework for Large Data
Analysis
- Authors: Wandong Zhang (1 and 2), QM Jonathan Wu (1), Yimin Yang (2 and 3), WG
Will Zhao (2 and 4), Tianlei Wang (5), and Hui Zhang (6) ((1) University of
Windsor, (2) Lakehead University, (3) Vector Institute for Artificial
Intelligence, (4) CEGEP de Ste Foy, (5) Hangzhou Dianzi University, (6) Hunan
University)
- Abstract summary: In complex tasks such as handling the ImageNet dataset, there are often many more clues that can be directly encoded.
This serves as the motivation to retrain the latent space representations to learn some clues that unsupervised learning has not yet learned.
In this paper, a recomputation-based multilayer network using MP inverse (RML-MP) is developed.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most multilayer least squares (LS)-based neural networks are structured with
two separate stages: unsupervised feature encoding and supervised pattern
classification. Once the unsupervised learning is finished, the latent encoding
would be fixed without supervised fine-tuning. However, in complex tasks such
as handling the ImageNet dataset, there are often many more clues that can be
directly encoded, while the unsupervised learning, by definition cannot know
exactly what is useful for a certain task. This serves as the motivation to
retrain the latent space representations to learn some clues that unsupervised
learning has not yet learned. In particular, the error matrix from the output
layer is pulled back to each hidden layer, and the parameters of the hidden
layer are recalculated with Moore-Penrose (MP) inverse for more generalized
representations. In this paper, a recomputation-based multilayer network using
MP inverse (RML-MP) is developed. A sparse RML-MP (SRML-MP) model to boost the
performance of RML-MP is then proposed. The experimental results with varying
training samples (from 3 K to 1.8 M) show that the proposed models provide
better generalization performance than most representation learning algorithms.
Related papers
- Multimodal Learned Sparse Retrieval with Probabilistic Expansion Control [66.78146440275093]
Learned retrieval (LSR) is a family of neural methods that encode queries and documents into sparse lexical vectors.
We explore the application of LSR to the multi-modal domain, with a focus on text-image retrieval.
Current approaches like LexLIP and STAIR require complex multi-step training on massive datasets.
Our proposed approach efficiently transforms dense vectors from a frozen dense model into sparse lexical vectors.
arXiv Detail & Related papers (2024-02-27T14:21:56Z) - Automated Sizing and Training of Efficient Deep Autoencoders using
Second Order Algorithms [0.46040036610482665]
We propose a multi-step training method for generalized linear classifiers.
validation error is minimized by pruning of unnecessary inputs.
desired outputs are improved via a method similar to the Ho-Kashyap rule.
arXiv Detail & Related papers (2023-08-11T16:48:31Z) - MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Masked Image Modeling with Local Multi-Scale Reconstruction [54.91442074100597]
Masked Image Modeling (MIM) achieves outstanding success in self-supervised representation learning.
Existing MIM models conduct reconstruction task only at the top layer of encoder.
We design local multi-scale reconstruction, where the lower and upper layers reconstruct fine-scale and coarse-scale supervision signals respectively.
arXiv Detail & Related papers (2023-03-09T13:42:04Z) - Decoupled Multi-task Learning with Cyclical Self-Regulation for Face
Parsing [71.19528222206088]
We propose a novel Decoupled Multi-task Learning with Cyclical Self-Regulation for face parsing.
Specifically, DML-CSR designs a multi-task model which comprises face parsing, binary edge, and category edge detection.
Our method achieves the new state-of-the-art performance on the Helen, CelebA-HQ, and LapaMask datasets.
arXiv Detail & Related papers (2022-03-28T02:12:30Z) - Understanding Self-supervised Learning with Dual Deep Networks [74.92916579635336]
We propose a novel framework to understand contrastive self-supervised learning (SSL) methods that employ dual pairs of deep ReLU networks.
We prove that in each SGD update of SimCLR with various loss functions, the weights at each layer are updated by a emphcovariance operator.
To further study what role the covariance operator plays and which features are learned in such a process, we model data generation and augmentation processes through a emphhierarchical latent tree model (HLTM)
arXiv Detail & Related papers (2020-10-01T17:51:49Z) - Evolving Losses for Unsupervised Video Representation Learning [91.2683362199263]
We present a new method to learn video representations from large-scale unlabeled video data.
The proposed unsupervised representation learning results in a single RGB network and outperforms previous methods.
arXiv Detail & Related papers (2020-02-26T16:56:07Z) - Inference in Multi-Layer Networks with Matrix-Valued Unknowns [32.635971570510755]
We consider the problem of inferring the input and hidden variables of a multi-layer neural network from an observation of the output.
A unified approximation algorithm for both MAP and MMSE inference is proposed.
It is shown that the performance of the proposed Multi-Layer Matrix VAMP (ML-Mat-VAMP) algorithm can be exactly predicted in a certain random large-system limit.
arXiv Detail & Related papers (2020-01-26T04:00:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.