ParsNets: A Parsimonious Orthogonal and Low-Rank Linear Networks for
Zero-Shot Learning
- URL: http://arxiv.org/abs/2312.09709v2
- Date: Thu, 21 Dec 2023 09:40:00 GMT
- Title: ParsNets: A Parsimonious Orthogonal and Low-Rank Linear Networks for
Zero-Shot Learning
- Authors: Jingcai Guo, Qihua Zhou, Ruibing Li, Xiaocheng Lu, Ziming Liu, Junyang
Chen, Xin Xie, Jie Zhang
- Abstract summary: This paper provides a novel parsimonious yet efficient design for zero-shot learning (ZSL), dubbed ParsNets, to achieve equivalent or even better performance against existing deep models.
To facilitate the generalization of local linearities, we construct a maximal margin geometry on the learned features by enforcing low-rank constraints on intra-class samples and high-rank constraints on inter-class samples.
To enhance the model's adaptability and counterbalance over/under-fittings in ZSL, a set of sample-wise indicators is employed to select a sparse subset from these base linear networks to form a composite
- Score: 22.823915322924304
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper provides a novel parsimonious yet efficient design for zero-shot
learning (ZSL), dubbed ParsNets, where we are interested in learning a
composition of on-device friendly linear networks, each with orthogonality and
low-rankness properties, to achieve equivalent or even better performance
against existing deep models. Concretely, we first refactor the core module of
ZSL, i.e., visual-semantics mapping function, into several base linear networks
that correspond to diverse components of the semantic space, where the complex
nonlinearity can be collapsed into simple local linearities. Then, to
facilitate the generalization of local linearities, we construct a maximal
margin geometry on the learned features by enforcing low-rank constraints on
intra-class samples and high-rank constraints on inter-class samples, resulting
in orthogonal subspaces for different classes and each subspace lies on a
compact manifold. To enhance the model's adaptability and counterbalance
over/under-fittings in ZSL, a set of sample-wise indicators is employed to
select a sparse subset from these base linear networks to form a composite
semantic predictor for each sample. Notably, maximal margin geometry can
guarantee the diversity of features, and meanwhile, local linearities guarantee
efficiency. Thus, our ParsNets can generalize better to unseen classes and can
be deployed flexibly on resource-constrained devices. Theoretical explanations
and extensive experiments are conducted to verify the effectiveness of the
proposed method.
Related papers
- ReLU Neural Networks with Linear Layers are Biased Towards Single- and Multi-Index Models [9.96121040675476]
This manuscript explores how properties of functions learned by neural networks of depth greater than two layers affect predictions.
Our framework considers a family of networks of varying depths that all have the same capacity but different representation costs.
arXiv Detail & Related papers (2023-05-24T22:10:12Z) - Learning an Invertible Output Mapping Can Mitigate Simplicity Bias in
Neural Networks [66.76034024335833]
We investigate why diverse/ complex features are learned by the backbone, and their brittleness is due to the linear classification head relying primarily on the simplest features.
We propose Feature Reconstruction Regularizer (FRR) to ensure that the learned features can be reconstructed back from the logits.
We demonstrate up to 15% gains in OOD accuracy on the recently introduced semi-synthetic datasets with extreme distribution shifts.
arXiv Detail & Related papers (2022-10-04T04:01:15Z) - Neural Manifold Clustering and Embedding [13.08270828061924]
Non-linear subspace clustering or manifold clustering aims to cluster data points based on manifold structures and learn to parameterize each manifold as a linear subspace in a feature space.
Deep neural networks have the potential to achieve this goal under highly non-linear settings given their large capacity and flexibility.
We argue that achieving manifold clustering with neural networks requires two essential ingredients: a domain-specific constraint that ensures the identification of the manifold, and a learning algorithm for embedding each manifold to a linear subspace in the feature space.
arXiv Detail & Related papers (2022-01-24T23:13:37Z) - Clustered Federated Learning via Generalized Total Variation
Minimization [83.26141667853057]
We study optimization methods to train local (or personalized) models for local datasets with a decentralized network structure.
Our main conceptual contribution is to formulate federated learning as total variation minimization (GTV)
Our main algorithmic contribution is a fully decentralized federated learning algorithm.
arXiv Detail & Related papers (2021-05-26T18:07:19Z) - Sample-Efficient Reinforcement Learning Is Feasible for Linearly
Realizable MDPs with Limited Revisiting [60.98700344526674]
Low-complexity models such as linear function representation play a pivotal role in enabling sample-efficient reinforcement learning.
In this paper, we investigate a new sampling protocol, which draws samples in an online/exploratory fashion but allows one to backtrack and revisit previous states in a controlled and infrequent manner.
We develop an algorithm tailored to this setting, achieving a sample complexity that scales practicallyly with the feature dimension, the horizon, and the inverse sub-optimality gap, but not the size of the state/action space.
arXiv Detail & Related papers (2021-05-17T17:22:07Z) - Linear Iterative Feature Embedding: An Ensemble Framework for
Interpretable Model [6.383006473302968]
A new ensemble framework for interpretable model called Linear Iterative Feature Embedding (LIFE) has been developed.
LIFE is able to fit a wide single-hidden-layer neural network (NN) accurately with three steps.
LIFE consistently outperforms directly trained single-hidden-layer NNs and also outperforms many other benchmark models.
arXiv Detail & Related papers (2021-03-18T02:01:17Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - DessiLBI: Exploring Structural Sparsity of Deep Networks via
Differential Inclusion Paths [45.947140164621096]
We propose a new approach based on differential inclusions of inverse scale spaces.
We show that DessiLBI unveils "winning tickets" in early epochs.
arXiv Detail & Related papers (2020-07-04T04:40:16Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z) - OSLNet: Deep Small-Sample Classification with an Orthogonal Softmax
Layer [77.90012156266324]
This paper aims to find a subspace of neural networks that can facilitate a large decision margin.
We propose the Orthogonal Softmax Layer (OSL), which makes the weight vectors in the classification layer remain during both the training and test processes.
Experimental results demonstrate that the proposed OSL has better performance than the methods used for comparison on four small-sample benchmark datasets.
arXiv Detail & Related papers (2020-04-20T02:41:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.