BoundaryNet: An Attentive Deep Network with Fast Marching Distance Maps
for Semi-automatic Layout Annotation
- URL: http://arxiv.org/abs/2108.09433v1
- Date: Sat, 21 Aug 2021 04:24:00 GMT
- Title: BoundaryNet: An Attentive Deep Network with Fast Marching Distance Maps
for Semi-automatic Layout Annotation
- Authors: Abhishek Trivedi and Ravi Kiran Sarvadevabhatla
- Abstract summary: BoundaryNet is a novel resizing-free approach for high-precision semi-automatic layout annotation.
Results on a challenging image manuscript dataset demonstrate that BoundaryNet outperforms strong baselines.
- Score: 10.990447273771592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Precise boundary annotations of image regions can be crucial for downstream
applications which rely on region-class semantics. Some document collections
contain densely laid out, highly irregular and overlapping multi-class region
instances with large range in aspect ratio. Fully automatic boundary estimation
approaches tend to be data intensive, cannot handle variable-sized images and
produce sub-optimal results for aforementioned images. To address these issues,
we propose BoundaryNet, a novel resizing-free approach for high-precision
semi-automatic layout annotation. The variable-sized user selected region of
interest is first processed by an attention-guided skip network. The network
optimization is guided via Fast Marching distance maps to obtain a good quality
initial boundary estimate and an associated feature representation. These
outputs are processed by a Residual Graph Convolution Network optimized using
Hausdorff loss to obtain the final region boundary. Results on a challenging
image manuscript dataset demonstrate that BoundaryNet outperforms strong
baselines and produces high-quality semantic region boundaries. Qualitatively,
our approach generalizes across multiple document image datasets containing
different script systems and layouts, all without additional fine-tuning. We
integrate BoundaryNet into a document annotation system and show that it
provides high annotation throughput compared to manual and fully automatic
alternatives.
Related papers
- Adaptive Region Selection for Active Learning in Whole Slide Image
Semantic Segmentation [3.1392713791311766]
Region-based active learning (AL) involves training the model on a limited number of annotated image regions.
We introduce a novel technique for selecting annotation regions adaptively, mitigating the reliance on this AL hyper parameter.
We evaluate our method using the task of breast cancer segmentation on the public CAMELYON16 dataset.
arXiv Detail & Related papers (2023-07-14T05:34:13Z) - ReFit: A Framework for Refinement of Weakly Supervised Semantic
Segmentation using Object Border Fitting for Medical Images [4.945138408504987]
Weakly Supervised Semantic (WSSS) relying only on image-level supervision is a promising approach to deal with the need for networks.
We propose our novel ReFit framework, which deploys state-of-the-art class activation maps combined with various post-processing techniques.
By applying our method to WSSS predictions, we achieved up to 10% improvement over the current state-of-the-art WSSS methods for medical imaging.
arXiv Detail & Related papers (2023-03-14T12:46:52Z) - Push-the-Boundary: Boundary-aware Feature Propagation for Semantic
Segmentation of 3D Point Clouds [0.5249805590164901]
We propose a boundary-aware feature propagation mechanism to improve semantic segmentation near object boundaries.
With one shared encoder, our network outputs (i) boundary localization, (ii) prediction of directions pointing to the object's interior, and (iii) semantic segmentation, in three parallel streams.
Our proposed approach yields consistent improvements by reducing boundary errors.
arXiv Detail & Related papers (2022-12-23T15:42:01Z) - SpaText: Spatio-Textual Representation for Controllable Image Generation [61.89548017729586]
SpaText is a new method for text-to-image generation using open-vocabulary scene control.
In addition to a global text prompt that describes the entire scene, the user provides a segmentation map.
We show its effectiveness on two state-of-the-art diffusion models: pixel-based and latent-conditional-based.
arXiv Detail & Related papers (2022-11-25T18:59:10Z) - Grasp-Oriented Fine-grained Cloth Segmentation without Real Supervision [66.56535902642085]
This paper tackles the problem of fine-grained region detection in deformed clothes using only a depth image.
We define up to 6 semantic regions of varying extent, including edges on the neckline, sleeve cuffs, and hem, plus top and bottom grasping points.
We introduce a U-net based network to segment and label these parts.
We show that training our network solely with synthetic data and the proposed DA yields results competitive with models trained on real data.
arXiv Detail & Related papers (2021-10-06T16:31:20Z) - Improving Semantic Segmentation via Decoupled Body and Edge Supervision [89.57847958016981]
Existing semantic segmentation approaches either aim to improve the object's inner consistency by modeling the global context, or refine objects detail along their boundaries by multi-scale feature fusion.
In this paper, a new paradigm for semantic segmentation is proposed.
Our insight is that appealing performance of semantic segmentation requires textitexplicitly modeling the object textitbody and textitedge, which correspond to the high and low frequency of the image.
We show that the proposed framework with various baselines or backbone networks leads to better object inner consistency and object boundaries.
arXiv Detail & Related papers (2020-07-20T12:11:22Z) - Split-Merge Pooling [36.2980225204665]
Split-Merge pooling is introduced to preserve spatial information without subsampling.
We evaluate our approach for dense semantic segmentation of large image sizes taken from the Cityscapes and GTA-5 datasets.
arXiv Detail & Related papers (2020-06-13T23:20:30Z) - DeepStrip: High Resolution Boundary Refinement [60.00241966809684]
We propose to convert regions of interest into strip images and compute a boundary prediction in the strip domain.
To detect the target boundary, we present a framework with two prediction layers.
We enforce a matching consistency and C0 continuity regularization to the network to reduce false alarms.
arXiv Detail & Related papers (2020-03-25T22:44:48Z) - Saliency Enhancement using Gradient Domain Edges Merging [65.90255950853674]
We develop a method to merge the edges with the saliency maps to improve the performance of the saliency.
This leads to our proposed saliency enhancement using edges (SEE) with an average improvement of at least 3.4 times higher on the DUT-OMRON dataset.
The SEE algorithm is split into 2 parts, SEE-Pre for preprocessing and SEE-Post pour postprocessing.
arXiv Detail & Related papers (2020-02-11T14:04:56Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.