Polygonal Building Segmentation by Frame Field Learning
- URL: http://arxiv.org/abs/2004.14875v2
- Date: Wed, 31 Mar 2021 13:30:18 GMT
- Title: Polygonal Building Segmentation by Frame Field Learning
- Authors: Nicolas Girard, Dmitriy Smirnov, Justin Solomon, Yuliya Tarabalka
- Abstract summary: We bridge the gap between deep network output and the format used in downstream tasks by adding a frame field output to a deep segmentation model.
We train a deep neural network that aligns a predicted frame field to ground truth contours.
- Score: 37.86051935654666
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While state of the art image segmentation models typically output
segmentations in raster format, applications in geographic information systems
often require vector polygons. To help bridge the gap between deep network
output and the format used in downstream tasks, we add a frame field output to
a deep segmentation model for extracting buildings from remote sensing images.
We train a deep neural network that aligns a predicted frame field to ground
truth contours. This additional objective improves segmentation quality by
leveraging multi-task learning and provides structural information that later
facilitates polygonization; we also introduce a polygonization algorithm that
utilizes the frame field along with the raster segmentation. Our code is
available at https://github.com/Lydorn/Polygonization-by-Frame-Field-Learning.
Related papers
- Enhancing Polygonal Building Segmentation via Oriented Corners [0.3749861135832072]
This paper introduces a novel deep convolutional neural network named OriCornerNet, which directly extracts delineated building polygons from input images.
Our approach involves a deep model that predicts building footprint masks, corners, and orientation vectors that indicate directions toward adjacent corners.
Performance evaluations conducted on SpaceNet Vegas and CrowdAI-small datasets demonstrate the competitive efficacy of our approach.
arXiv Detail & Related papers (2024-07-17T01:59:06Z) - View-Consistent Hierarchical 3D Segmentation Using Ultrametric Feature Fields [52.08335264414515]
We learn a novel feature field within a Neural Radiance Field (NeRF) representing a 3D scene.
Our method takes view-inconsistent multi-granularity 2D segmentations as input and produces a hierarchy of 3D-consistent segmentations as output.
We evaluate our method and several baselines on synthetic datasets with multi-view images and multi-granular segmentation, showcasing improved accuracy and viewpoint-consistency.
arXiv Detail & Related papers (2024-05-30T04:14:58Z) - Real-Time Semantic Segmentation using Hyperspectral Images for Mapping
Unstructured and Unknown Environments [2.408714894793063]
We propose the use of hyperspectral images for real-time pixel-wise semantic classification and segmentation.
The resulting segmented image is processed to extract, filter, and approximate objects as polygons.
The resulting polygons are then used to generate a semantic map of the environment.
arXiv Detail & Related papers (2023-03-27T22:33:55Z) - Towards General-Purpose Representation Learning of Polygonal Geometries [62.34832826705641]
We develop a general-purpose polygon encoding model, which can encode a polygonal geometry into an embedding space.
We conduct experiments on two tasks: 1) shape classification based on MNIST; 2) spatial relation prediction based on two new datasets - DBSR-46K and DBSR-cplx46K.
Our results show that NUFTspec and ResNet1D outperform multiple existing baselines with significant margins.
arXiv Detail & Related papers (2022-09-29T15:59:23Z) - PolyWorld: Polygonal Building Extraction with Graph Neural Networks in
Satellite Images [10.661430927191205]
This paper introduces PolyWorld, a neural network that directly extracts building vertices from an image and connects them correctly to create precise polygons.
PolyWorld significantly outperforms the state-of-the-art in building polygonization.
arXiv Detail & Related papers (2021-11-30T15:23:17Z) - CenterPoly: real-time instance segmentation using bounding polygons [11.365829102707014]
We present a novel method, called CenterPoly, for real-time instance segmentation using bounding polygons.
We apply it to detect road users in dense urban environments, making it suitable for applications in intelligent transportation systems like automated vehicles.
Most of the network parameters are shared by the network heads, making it fast and lightweight enough to run at real-time speed.
arXiv Detail & Related papers (2021-08-19T21:31:30Z) - Polygonal Point Set Tracking [50.445151155209246]
We propose a novel learning-based polygonal point set tracking method.
Our goal is to track corresponding points on the target contour.
We present visual-effects applications of our method on part distortion and text mapping.
arXiv Detail & Related papers (2021-05-30T17:12:36Z) - Quantization in Relative Gradient Angle Domain For Building Polygon
Estimation [88.80146152060888]
CNN approaches often generate imprecise building morphologies including noisy edges and round corners.
We propose a module that uses prior knowledge of building corners to create angular and concise building polygons from CNN segmentation outputs.
Experimental results demonstrate that our method refines CNN output from a rounded approximation to a more clear-cut angular shape of the building footprint.
arXiv Detail & Related papers (2020-07-10T21:33:06Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z) - Learning light field synthesis with Multi-Plane Images: scene encoding
as a recurrent segmentation task [30.058283056074426]
This paper addresses the problem of view synthesis from large baseline light fields by turning a sparse set of input views into a Multi-plane Image (MPI)
Because available datasets are scarce, we propose a lightweight network that does not require extensive training.
Our model does not learn to estimate RGB layers but only encodes the scene geometry within MPI alpha layers, which comes down to a segmentation task.
arXiv Detail & Related papers (2020-02-12T14:35:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.