f-BRS: Rethinking Backpropagating Refinement for Interactive
Segmentation
- URL: http://arxiv.org/abs/2001.10331v3
- Date: Tue, 25 Aug 2020 11:52:26 GMT
- Title: f-BRS: Rethinking Backpropagating Refinement for Interactive
Segmentation
- Authors: Konstantin Sofiiuk, Ilia Petrov, Olga Barinova and Anton Konushin
- Abstract summary: We propose f-BRS (feature backpropagating refinement scheme) to solve an optimization problem with respect to auxiliary variables.
Experiments on GrabCut, Berkeley, DAVIS and SBD datasets set new state-of-the-art at an order of magnitude lower time per click compared to original BRS.
- Score: 8.304331351572277
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks have become a mainstream approach to interactive
segmentation. As we show in our experiments, while for some images a trained
network provides accurate segmentation result with just a few clicks, for some
unknown objects it cannot achieve satisfactory result even with a large amount
of user input. Recently proposed backpropagating refinement (BRS) scheme
introduces an optimization problem for interactive segmentation that results in
significantly better performance for the hard cases. At the same time, BRS
requires running forward and backward pass through a deep network several times
that leads to significantly increased computational budget per click compared
to other methods. We propose f-BRS (feature backpropagating refinement scheme)
that solves an optimization problem with respect to auxiliary variables instead
of the network inputs, and requires running forward and backward pass just for
a small part of a network. Experiments on GrabCut, Berkeley, DAVIS and SBD
datasets set new state-of-the-art at an order of magnitude lower time per click
compared to original BRS. The code and trained models are available at
https://github.com/saic-vul/fbrs_interactive_segmentation .
Related papers
- Convolutional Networks as Extremely Small Foundation Models: Visual Prompting and Theoretical Perspective [1.79487674052027]
In this paper, we design a prompting module which performs few-shot adaptation of generic deep networks to new tasks.
Driven by learning theory, we derive prompting modules that are as simple as possible, as they generalize better under the same training error.
In practice, SDForest has extremely low cost and achieves real-time even on CPU.
arXiv Detail & Related papers (2024-09-03T12:34:23Z) - SPEAR:Exact Gradient Inversion of Batches in Federated Learning [11.799563040751591]
Federated learning is a framework for machine learning where clients only share gradient updates and not their private data with a server.
We propose SPEAR, the first algorithm reconstructing whole batches with $b >1$ exactly.
We show that it recovers high-dimensional ImageNet inputs in batches of up to $b lesssim 25$ exactly while scaling to large networks.
arXiv Detail & Related papers (2024-03-06T18:52:39Z) - Receptive Field-based Segmentation for Distributed CNN Inference
Acceleration in Collaborative Edge Computing [93.67044879636093]
We study inference acceleration using distributed convolutional neural networks (CNNs) in collaborative edge computing network.
We propose a novel collaborative edge computing using fused-layer parallelization to partition a CNN model into multiple blocks of convolutional layers.
arXiv Detail & Related papers (2022-07-22T18:38:11Z) - Reviving Iterative Training with Mask Guidance for Interactive
Segmentation [8.271859911016719]
Recent works on click-based interactive segmentation have demonstrated state-of-the-art results by using various inference-time optimization schemes.
We propose a simple feedforward model for click-based interactive segmentation that employs the segmentation masks from previous steps.
We find that the models trained on a combination of COCO and LVIS with diverse and high-quality annotations show performance superior to all existing models.
arXiv Detail & Related papers (2021-02-12T15:44:31Z) - Solving Mixed Integer Programs Using Neural Networks [57.683491412480635]
This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one.
Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP.
We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each.
arXiv Detail & Related papers (2020-12-23T09:33:11Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z) - A Deep-Unfolded Reference-Based RPCA Network For Video
Foreground-Background Separation [86.35434065681925]
This paper proposes a new deep-unfolding-based network design for the problem of Robust Principal Component Analysis (RPCA)
Unlike existing designs, our approach focuses on modeling the temporal correlation between the sparse representations of consecutive video frames.
Experimentation using the moving MNIST dataset shows that the proposed network outperforms a recently proposed state-of-the-art RPCA network in the task of video foreground-background separation.
arXiv Detail & Related papers (2020-10-02T11:40:09Z) - A Partial Regularization Method for Network Compression [0.0]
We propose an approach of partial regularization rather than the original form of penalizing all parameters, which is said to be full regularization, to conduct model compression at a higher speed.
Experimental results show that as we expected, the computational complexity is reduced by observing less running time in almost all situations.
Surprisingly, it helps to improve some important metrics such as regression fitting results and classification accuracy in both training and test phases on multiple datasets.
arXiv Detail & Related papers (2020-09-03T00:38:27Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z) - Conditional Convolutions for Instance Segmentation [109.2706837177222]
We propose a simple yet effective instance segmentation framework, termed CondInst.
We employ dynamic instance-aware networks, conditioned on instances.
We demonstrate a simpler instance segmentation method that can achieve improved performance in both accuracy and inference speed.
arXiv Detail & Related papers (2020-03-12T08:42:36Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.