SAPA: Similarity-Aware Point Affiliation for Feature Upsampling
- URL: http://arxiv.org/abs/2209.12866v1
- Date: Mon, 26 Sep 2022 17:32:25 GMT
- Title: SAPA: Similarity-Aware Point Affiliation for Feature Upsampling
- Authors: Hao Lu, Wenze Liu, Zixuan Ye, Hongtao Fu, Yuliang Liu, Zhiguo Cao
- Abstract summary: We introduce point affiliation into feature upsampling, a notion that describes the affiliation of each upsampled point to a semantic cluster formed by local decoder feature points with semantic similarity.
We present a generic formulation for generating upsampling kernels that encourage semantic smoothness and boundary sharpness.
The key idea of our formulation is to generate similarity-aware kernels by comparing the similarity between each encoder feature point and the spatially associated local region of decoder features.
- Score: 27.546863377935118
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce point affiliation into feature upsampling, a notion that
describes the affiliation of each upsampled point to a semantic cluster formed
by local decoder feature points with semantic similarity. By rethinking point
affiliation, we present a generic formulation for generating upsampling
kernels. The kernels encourage not only semantic smoothness but also boundary
sharpness in the upsampled feature maps. Such properties are particularly
useful for some dense prediction tasks such as semantic segmentation. The key
idea of our formulation is to generate similarity-aware kernels by comparing
the similarity between each encoder feature point and the spatially associated
local region of decoder features. In this way, the encoder feature point can
function as a cue to inform the semantic cluster of upsampled feature points.
To embody the formulation, we further instantiate a lightweight upsampling
operator, termed Similarity-Aware Point Affiliation (SAPA), and investigate its
variants. SAPA invites consistent performance improvements on a number of dense
prediction tasks, including semantic segmentation, object detection, depth
estimation, and image matting. Code is available at:
https://github.com/poppinace/sapa
Related papers
- Semantics Meets Temporal Correspondence: Self-supervised Object-centric Learning in Videos [63.94040814459116]
Self-supervised methods have shown remarkable progress in learning high-level semantics and low-level temporal correspondence.
We propose a novel semantic-aware masked slot attention on top of the fused semantic features and correspondence maps.
We adopt semantic- and instance-level temporal consistency as self-supervision to encourage temporally coherent object-centric representations.
arXiv Detail & Related papers (2023-08-19T09:12:13Z) - On Point Affiliation in Feature Upsampling [32.28512034705838]
We introduce the notion of point affiliation into feature upsampling.
We show that an upsampled point can resort to its low-res decoder neighbors and high-res encoder point to reason the affiliation.
This formulation constitutes a novel, lightweight, and universal upsampling solution.
arXiv Detail & Related papers (2023-07-17T01:59:14Z) - Learning Implicit Feature Alignment Function for Semantic Segmentation [51.36809814890326]
Implicit Feature Alignment function (IFA) is inspired by the rapidly expanding topic of implicit neural representations.
We show that IFA implicitly aligns the feature maps at different levels and is capable of producing segmentation maps in arbitrary resolutions.
Our method can be combined with improvement on various architectures, and it achieves state-of-the-art accuracy trade-off on common benchmarks.
arXiv Detail & Related papers (2022-06-17T09:40:14Z) - SemAffiNet: Semantic-Affine Transformation for Point Cloud Segmentation [94.11915008006483]
We propose SemAffiNet for point cloud semantic segmentation.
We conduct extensive experiments on the ScanNetV2 and NYUv2 datasets.
arXiv Detail & Related papers (2022-05-26T17:00:23Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - CFNet: Learning Correlation Functions for One-Stage Panoptic
Segmentation [46.252118473248316]
We propose to first predict semantic-level and instance-level correlations among different locations that are utilized to enhance the backbone features.
We then feed the improved discriminative features into the corresponding segmentation heads, respectively.
We achieve state-of-the-art performance on MS with $45.1$% PQ and ADE20k with $32.6$% PQ.
arXiv Detail & Related papers (2022-01-13T05:31:14Z) - CPSeg: Cluster-free Panoptic Segmentation of 3D LiDAR Point Clouds [2.891413712995641]
We propose a novel real-time end-to-end panoptic segmentation network for LiDAR point clouds, called CPSeg.
CPSeg comprises a shared encoder, a dual decoder, a task-aware attention module (TAM) and a cluster-free instance segmentation head.
arXiv Detail & Related papers (2021-11-02T16:44:06Z) - SPU-Net: Self-Supervised Point Cloud Upsampling by Coarse-to-Fine
Reconstruction with Self-Projection Optimization [52.20602782690776]
It is expensive and tedious to obtain large scale paired sparse-canned point sets for training from real scanned sparse data.
We propose a self-supervised point cloud upsampling network, named SPU-Net, to capture the inherent upsampling patterns of points lying on the underlying object surface.
We conduct various experiments on both synthetic and real-scanned datasets, and the results demonstrate that we achieve comparable performance to the state-of-the-art supervised methods.
arXiv Detail & Related papers (2020-12-08T14:14:09Z) - Few-shot 3D Point Cloud Semantic Segmentation [138.80825169240302]
We propose a novel attention-aware multi-prototype transductive few-shot point cloud semantic segmentation method.
Our proposed method shows significant and consistent improvements compared to baselines in different few-shot point cloud semantic segmentation settings.
arXiv Detail & Related papers (2020-06-22T08:05:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.