Regularized Frank-Wolfe for Dense CRFs: Generalizing Mean Field and
Beyond
- URL: http://arxiv.org/abs/2110.14759v1
- Date: Wed, 27 Oct 2021 20:44:47 GMT
- Title: Regularized Frank-Wolfe for Dense CRFs: Generalizing Mean Field and
Beyond
- Authors: {\DJ}.Khu\^e L\^e-Huu and Karteek Alahari
- Abstract summary: We introduce regularized Frank-Wolfe, a general and effective CNN baseline inference for dense conditional fields.
We show that our new algorithms, with our new algorithms, with our new datasets, with significant improvements in strong strong neural networks.
- Score: 19.544213396776268
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce regularized Frank-Wolfe, a general and effective algorithm for
inference and learning of dense conditional random fields (CRFs). The algorithm
optimizes a nonconvex continuous relaxation of the CRF inference problem using
vanilla Frank-Wolfe with approximate updates, which are equivalent to
minimizing a regularized energy function. Our proposed method is a
generalization of existing algorithms such as mean field or concave-convex
procedure. This perspective not only offers a unified analysis of these
algorithms, but also allows an easy way of exploring different variants that
potentially yield better performance. We illustrate this in our empirical
results on standard semantic segmentation datasets, where several
instantiations of our regularized Frank-Wolfe outperform mean field inference,
both as a standalone component and as an end-to-end trainable layer in a neural
network. We also show that dense CRFs, coupled with our new algorithms, produce
significant improvements over strong CNN baselines.
Related papers
- DFWLayer: Differentiable Frank-Wolfe Optimization Layer [33.20550212188619]
Differentiable optimization has received a significant amount of attention due to its foundational role in the domain of machine learning based on neural networks.
This paper proposes a differentiable layer, named Differentiable Frank-Wolfe Layer (DFWLayer), by rolling out the Frank-Wolfe method.
Experimental results demonstrate that the DFWLayer not only attains competitive accuracy in solutions and gradients but also consistently adheres to constraints.
arXiv Detail & Related papers (2023-08-21T15:53:38Z) - Stochastic Unrolled Federated Learning [85.6993263983062]
We introduce UnRolled Federated learning (SURF), a method that expands algorithm unrolling to federated learning.
Our proposed method tackles two challenges of this expansion, namely the need to feed whole datasets to the unrolleds and the decentralized nature of federated learning.
arXiv Detail & Related papers (2023-05-24T17:26:22Z) - Sarah Frank-Wolfe: Methods for Constrained Optimization with Best Rates and Practical Features [65.64276393443346]
The Frank-Wolfe (FW) method is a popular approach for solving optimization problems with structured constraints.
We present two new variants of the algorithms for minimization of the finite-sum gradient.
arXiv Detail & Related papers (2023-04-23T20:05:09Z) - Large-scale Optimization of Partial AUC in a Range of False Positive
Rates [51.12047280149546]
The area under the ROC curve (AUC) is one of the most widely used performance measures for classification models in machine learning.
We develop an efficient approximated gradient descent method based on recent practical envelope smoothing technique.
Our proposed algorithm can also be used to minimize the sum of some ranked range loss, which also lacks efficient solvers.
arXiv Detail & Related papers (2022-03-03T03:46:18Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - RNN Training along Locally Optimal Trajectories via Frank-Wolfe
Algorithm [50.76576946099215]
We propose a novel and efficient training method for RNNs by iteratively seeking a local minima on the loss surface within a small region.
We develop a novel RNN training method that, surprisingly, even with the additional cost, the overall training cost is empirically observed to be lower than back-propagation.
arXiv Detail & Related papers (2020-10-12T01:59:18Z) - AIN: Fast and Accurate Sequence Labeling with Approximate Inference
Network [75.44925576268052]
The linear-chain Conditional Random Field (CRF) model is one of the most widely-used neural sequence labeling approaches.
Exact probabilistic inference algorithms are typically applied in training and prediction stages of the CRF model.
We propose to employ a parallelizable approximate variational inference algorithm for the CRF model.
arXiv Detail & Related papers (2020-09-17T12:18:43Z) - Efficient Optimization of Dominant Set Clustering with Frank-Wolfe
Algorithms [4.873362301533825]
We study Frank-Wolfe algorithms -- standard, pairwise, and away-steps -- for efficient optimization of Dominant Set Clustering.
We present a unified and computationally efficient framework to employ the different variants of Frank-Wolfe methods.
arXiv Detail & Related papers (2020-07-22T20:08:41Z) - Stochastic Frank-Wolfe for Constrained Finite-Sum Minimization [30.980431848476925]
We propose an algorithm for constrained finite smooth-sum minimization with a generalized linear prediction/structure.
The proposed method is simple to implement, does not require step-size tuning, and has a constant per-iteration independent of the dataset size.
We provide an implementation of all considered methods in an open-source package.
arXiv Detail & Related papers (2020-02-27T00:47:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.