Toward Learning Robust and Invariant Representations with Alignment
Regularization and Data Augmentation
- URL: http://arxiv.org/abs/2206.01909v1
- Date: Sat, 4 Jun 2022 04:29:19 GMT
- Title: Toward Learning Robust and Invariant Representations with Alignment
Regularization and Data Augmentation
- Authors: Haohan Wang, Zeyi Huang, Xindi Wu, Eric P. Xing
- Abstract summary: This paper is motivated by a proliferation of options of alignment regularizations.
We evaluate the performances of several popular design choices along the dimensions of robustness and invariance.
We also formally analyze the behavior of alignment regularization to complement our empirical study under assumptions we consider realistic.
- Score: 76.85274970052762
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data augmentation has been proven to be an effective technique for developing
machine learning models that are robust to known classes of distributional
shifts (e.g., rotations of images), and alignment regularization is a technique
often used together with data augmentation to further help the model learn
representations invariant to the shifts used to augment the data. In this
paper, motivated by a proliferation of options of alignment regularizations, we
seek to evaluate the performances of several popular design choices along the
dimensions of robustness and invariance, for which we introduce a new test
procedure. Our synthetic experiment results speak to the benefits of squared l2
norm regularization. Further, we also formally analyze the behavior of
alignment regularization to complement our empirical study under assumptions we
consider realistic. Finally, we test this simple technique we identify
(worst-case data augmentation with squared l2 norm alignment regularization)
and show that the benefits of this method outrun those of the specially
designed methods. We also release a software package in both TensorFlow and
PyTorch for users to use the method with a couple of lines at
https://github.com/jyanln/AlignReg.
Related papers
- Transducer Consistency Regularization for Speech to Text Applications [4.510630624936377]
We present Transducer Consistency Regularization (TCR), a consistency regularization method for transducer models.
We utilize occupational probabilities to give different weights on transducer output distributions, thus only alignments close to oracle alignments would contribute to the model learning.
Our experiments show the proposed method is superior to other consistency regularization implementations and could effectively reduce word error rate (WER) by 4.3% relatively comparing with a strong baseline on the textscLibrispeech dataset.
arXiv Detail & Related papers (2024-10-09T23:53:13Z) - Joint Distributional Learning via Cramer-Wold Distance [0.7614628596146602]
We introduce the Cramer-Wold distance regularization, which can be computed in a closed-form, to facilitate joint distributional learning for high-dimensional datasets.
We also introduce a two-step learning method to enable flexible prior modeling and improve the alignment between the aggregated posterior and the prior distribution.
arXiv Detail & Related papers (2023-10-25T05:24:23Z) - Single Domain Generalization via Normalised Cross-correlation Based
Convolutions [14.306250516592304]
Single Domain Generalization aims to train robust models using data from a single source.
We propose a novel operator called XCNorm that computes the normalized cross-correlation between weights and an input feature patch.
We show that deep neural networks composed of this operator are robust to common semantic distribution shifts.
arXiv Detail & Related papers (2023-07-12T04:15:36Z) - Automatic Data Augmentation via Invariance-Constrained Learning [94.27081585149836]
Underlying data structures are often exploited to improve the solution of learning tasks.
Data augmentation induces these symmetries during training by applying multiple transformations to the input data.
This work tackles these issues by automatically adapting the data augmentation while solving the learning task.
arXiv Detail & Related papers (2022-09-29T18:11:01Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - Squared $\ell_2$ Norm as Consistency Loss for Leveraging Augmented Data
to Learn Robust and Invariant Representations [76.85274970052762]
Regularizing distance between embeddings/representations of original samples and augmented counterparts is a popular technique for improving robustness of neural networks.
In this paper, we explore these various regularization choices, seeking to provide a general understanding of how we should regularize the embeddings.
We show that the generic approach we identified (squared $ell$ regularized augmentation) outperforms several recent methods, which are each specially designed for one task.
arXiv Detail & Related papers (2020-11-25T22:40:09Z) - AdaS: Adaptive Scheduling of Stochastic Gradients [50.80697760166045]
We introduce the notions of textit"knowledge gain" and textit"mapping condition" and propose a new algorithm called Adaptive Scheduling (AdaS)
Experimentation reveals that, using the derived metrics, AdaS exhibits: (a) faster convergence and superior generalization over existing adaptive learning methods; and (b) lack of dependence on a validation set to determine when to stop training.
arXiv Detail & Related papers (2020-06-11T16:36:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.