Maximum Batch Frobenius Norm for Multi-Domain Text Classification
- URL: http://arxiv.org/abs/2202.00537v1
- Date: Sat, 29 Jan 2022 14:37:56 GMT
- Title: Maximum Batch Frobenius Norm for Multi-Domain Text Classification
- Authors: Yuan Wu, Diana Inkpen, Ahmed El-Roby
- Abstract summary: We propose a maximum batch Frobenius norm (MBF) method to boost the feature discriminability for multi-domain text classification.
Experiments on two MDTC benchmarks show that our MBF approach can effectively advance the performance of the state-of-the-art.
- Score: 19.393393465837377
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-domain text classification (MDTC) has obtained remarkable achievements
due to the advent of deep learning. Recently, many endeavors are devoted to
applying adversarial learning to extract domain-invariant features to yield
state-of-the-art results. However, these methods still face one challenge:
transforming original features to be domain-invariant distorts the
distributions of the original features, degrading the discriminability of the
learned features. To address this issue, we first investigate the structure of
the batch classification output matrix and theoretically justify that the
discriminability of the learned features has a positive correlation with the
Frobenius norm of the batch output matrix. Based on this finding, we propose a
maximum batch Frobenius norm (MBF) method to boost the feature discriminability
for MDTC. Experiments on two MDTC benchmarks show that our MBF approach can
effectively advance the performance of the state-of-the-art.
Related papers
- Regularized Multi-output Gaussian Convolution Process with Domain Adaptation [0.0]
Multi-output Gaussian process (MGP) has been attracting increasing attention as a transfer learning method to model multiple outputs.
Despite its high flexibility and generality, MGP still faces two critical challenges when applied to transfer learning.
The first one is negative transfer, which occurs when there exists no shared information among the outputs.
The second challenge is the input domain inconsistency, which is commonly studied in transfer learning yet not explored in MGP.
arXiv Detail & Related papers (2024-09-04T14:56:28Z) - Stochastic Adversarial Networks for Multi-Domain Text Classification [9.359945319927675]
We introduce the Adversarial Network (SAN), which innovatively models the parameters of the domain-specific feature extractor.
Our approach integrates domain label smoothing and robust pseudo-label regularization to fortify the stability of adversarial training.
The performance of our SAN, evaluated on two leading MDTC benchmarks, demonstrates its competitive edge against the current state-of-the-art methodologies.
arXiv Detail & Related papers (2024-05-28T00:02:38Z) - SOFARI: High-Dimensional Manifold-Based Inference [8.860162863559163]
We introduce two SOFARI variants to handle strongly and weakly latent factors, where the latter covers a broader range of applications.
We show that SOFARI provides bias-corrected estimators for both latent left factor vectors and singular values, for which we show to enjoy the mean-zero normal distributions with sparse estimable variances.
We illustrate the effectiveness of SOFARI and justify our theoretical results through simulation examples and a real data application in economic forecasting.
arXiv Detail & Related papers (2023-09-26T16:01:54Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - Synergies between Disentanglement and Sparsity: Generalization and
Identifiability in Multi-Task Learning [79.83792914684985]
We prove a new identifiability result that provides conditions under which maximally sparse base-predictors yield disentangled representations.
Motivated by this theoretical result, we propose a practical approach to learn disentangled representations based on a sparsity-promoting bi-level optimization problem.
arXiv Detail & Related papers (2022-11-26T21:02:09Z) - Unitary Approximate Message Passing for Matrix Factorization [90.84906091118084]
We consider matrix factorization (MF) with certain constraints, which finds wide applications in various areas.
We develop a Bayesian approach to MF with an efficient message passing implementation, called UAMPMF.
We show that UAMPMF significantly outperforms state-of-the-art algorithms in terms of recovery accuracy, robustness and computational complexity.
arXiv Detail & Related papers (2022-07-31T12:09:32Z) - Learning Multiresolution Matrix Factorization and its Wavelet Networks
on Graphs [11.256959274636724]
Multiresolution Matrix Factorization (MMF) is unusual amongst fast matrix factorization algorithms.
We propose a learnable version of MMF that carfully optimize the factorization with a combination of reinforcement learning and Stiefel manifold optimization.
We show that the resulting wavelet basis far outperforms prior MMF algorithms and provides the first version of this type of factorization that can be robustly deployed on standard learning tasks.
arXiv Detail & Related papers (2021-11-02T23:14:17Z) - Learning with Multiclass AUC: Theory and Algorithms [141.63211412386283]
Area under the ROC curve (AUC) is a well-known ranking metric for problems such as imbalanced learning and recommender systems.
In this paper, we start an early trial to consider the problem of learning multiclass scoring functions via optimizing multiclass AUC metrics.
arXiv Detail & Related papers (2021-07-28T05:18:10Z) - MCDAL: Maximum Classifier Discrepancy for Active Learning [74.73133545019877]
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition.
We propose in this paper a novel active learning framework that we call Maximum Discrepancy for Active Learning (MCDAL)
In particular, we utilize two auxiliary classification layers that learn tighter decision boundaries by maximizing the discrepancies among them.
arXiv Detail & Related papers (2021-07-23T06:57:08Z) - Feature Weighted Non-negative Matrix Factorization [92.45013716097753]
We propose the Feature weighted Non-negative Matrix Factorization (FNMF) in this paper.
FNMF learns the weights of features adaptively according to their importances.
It can be solved efficiently with the suggested optimization algorithm.
arXiv Detail & Related papers (2021-03-24T21:17:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.