Double-Bounded Optimal Transport for Advanced Clustering and
Classification
- URL: http://arxiv.org/abs/2401.11418v1
- Date: Sun, 21 Jan 2024 07:43:01 GMT
- Title: Double-Bounded Optimal Transport for Advanced Clustering and
Classification
- Authors: Liangliang Shi, Zhaoqi Shen, Junchi Yan
- Abstract summary: We propose Doubly Bounded Optimal Transport (DB-OT), which assumes that the target distribution is restricted within two boundaries instead of a fixed one.
We show that our method can achieve good results with our improved inference scheme in the testing stage.
- Score: 58.237576976486544
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optimal transport (OT) is attracting increasing attention in machine
learning. It aims to transport a source distribution to a target one at minimal
cost. In its vanilla form, the source and target distributions are
predetermined, which contracts to the real-world case involving undetermined
targets. In this paper, we propose Doubly Bounded Optimal Transport (DB-OT),
which assumes that the target distribution is restricted within two boundaries
instead of a fixed one, thus giving more freedom for the transport to find
solutions. Based on the entropic regularization of DB-OT, three scaling-based
algorithms are devised for calculating the optimal solution. We also show that
our DB-OT is helpful for barycenter-based clustering, which can avoid the
excessive concentration of samples in a single cluster. Then we further develop
DB-OT techniques for long-tailed classification which is an emerging and open
problem. We first propose a connection between OT and classification, that is,
in the classification task, training involves optimizing the Inverse OT to
learn the representations, while testing involves optimizing the OT for
predictions. With this OT perspective, we first apply DB-OT to improve the
loss, and the Balanced Softmax is shown as a special case. Then we apply DB-OT
for inference in the testing process. Even with vanilla Softmax trained
features, our extensive experimental results show that our method can achieve
good results with our improved inference scheme in the testing stage.
Related papers
- A Bayesian Approach to Data Point Selection [24.98069363998565]
Data point selection (DPS) is becoming a critical topic in deep learning.
Existing approaches to DPS are predominantly based on a bi-level optimisation (BLO) formulation.
We propose a novel Bayesian approach to DPS.
arXiv Detail & Related papers (2024-11-06T09:04:13Z) - Estimating Barycenters of Distributions with Neural Optimal Transport [93.28746685008093]
We propose a new scalable approach for solving the Wasserstein barycenter problem.
Our methodology is based on the recent Neural OT solver.
We also establish theoretical error bounds for our proposed approach.
arXiv Detail & Related papers (2024-02-06T09:17:07Z) - A Class-aware Optimal Transport Approach with Higher-Order Moment
Matching for Unsupervised Domain Adaptation [33.712557972990034]
Unsupervised domain adaptation (UDA) aims to transfer knowledge from a labeled source domain to an unlabeled target domain.
We introduce a novel approach called class-aware optimal transport (OT), which measures the OT distance between a distribution over the source class-conditional distributions.
arXiv Detail & Related papers (2024-01-29T08:27:31Z) - Analyzing and Improving Optimal-Transport-based Adversarial Networks [9.980822222343921]
Optimal Transport (OT) problem aims to find a transport plan that bridges two distributions while minimizing a given cost function.
OT theory has been widely utilized in generative modeling.
Our approach achieves a FID score of 2.51 on CIFAR-10 and 5.99 on CelebA-HQ-256, outperforming unified OT-based adversarial approaches.
arXiv Detail & Related papers (2023-10-04T06:52:03Z) - Building the Bridge of Schr\"odinger: A Continuous Entropic Optimal
Transport Benchmark [96.06787302688595]
We propose a novel way to create pairs of probability distributions for which the ground truth OT solution is known by the construction.
We use these benchmark pairs to test how well existing neural EOT/SB solvers actually compute the EOT solution.
arXiv Detail & Related papers (2023-06-16T20:03:36Z) - Generative Modeling through the Semi-dual Formulation of Unbalanced
Optimal Transport [9.980822222343921]
We propose a novel generative model based on the semi-dual formulation of Unbalanced Optimal Transport (UOT)
Unlike OT, UOT relaxes the hard constraint on distribution matching. This approach provides better robustness against outliers, stability during training, and faster convergence.
Our model outperforms existing OT-based generative models, achieving FID scores of 2.97 on CIFAR-10 and 6.36 on CelebA-HQ-256.
arXiv Detail & Related papers (2023-05-24T06:31:05Z) - Active Finetuning: Exploiting Annotation Budget in the
Pretraining-Finetuning Paradigm [132.9949120482274]
This paper focuses on the selection of samples for annotation in the pretraining-finetuning paradigm.
We propose a novel method called ActiveFT for active finetuning task to select a subset of data distributing similarly with the entire unlabeled pool.
Extensive experiments show the leading performance and high efficiency of ActiveFT superior to baselines on both image classification and semantic segmentation.
arXiv Detail & Related papers (2023-03-25T07:17:03Z) - On Second-order Optimization Methods for Federated Learning [59.787198516188425]
We evaluate the performance of several second-order distributed methods with local steps in the federated learning setting.
We propose a novel variant that uses second-order local information for updates and a global line search to counteract the resulting local specificity.
arXiv Detail & Related papers (2021-09-06T12:04:08Z) - Robust Optimal Transport with Applications in Generative Modeling and
Domain Adaptation [120.69747175899421]
Optimal Transport (OT) distances such as Wasserstein have been used in several areas such as GANs and domain adaptation.
We propose a computationally-efficient dual form of the robust OT optimization that is amenable to modern deep learning applications.
Our approach can train state-of-the-art GAN models on noisy datasets corrupted with outlier distributions.
arXiv Detail & Related papers (2020-10-12T17:13:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.