A generalised feature for low level vision
- URL: http://arxiv.org/abs/2102.02000v1
- Date: Wed, 3 Feb 2021 11:02:03 GMT
- Title: A generalised feature for low level vision
- Authors: Dr David Sinclair and Dr Christopher Town
- Abstract summary: The Sinclair-Town transform subsumes the rolls of both edge-detector, MSER style region detector and corner detector.
The difference from the local mean is quantised to 3 values (dark-neutral-light)
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This papers presents a novel quantised transform (the Sinclair-Town or ST
transform for short) that subsumes the rolls of both edge-detector, MSER style
region detector and corner detector. The transform is similar to the $unsharp$
transform but the difference from the local mean is quantised to 3 values
(dark-neutral-light). The transform naturally leads to the definition of an
appropriate local scale. A range of methods for extracting shape features form
the transformed image are presented. The generalized feature provides a robust
basis for establishing correspondence between images. The transform readily
admits more complicated kernel behaviour including multi-scale and asymmetric
elements to prefer shorter scale or oriented local features.
Related papers
- Learning SO(3)-Invariant Semantic Correspondence via Local Shape Transform [62.27337227010514]
We introduce a novel self-supervised Rotation-Invariant 3D correspondence learner with Local Shape Transform, dubbed RIST.
RIST learns to establish dense correspondences between shapes even under challenging intra-class variations and arbitrary orientations.
RIST demonstrates state-of-the-art performances on 3D part label transfer and semantic keypoint transfer given arbitrarily rotated point cloud pairs.
arXiv Detail & Related papers (2024-04-17T08:09:25Z) - Region Feature Descriptor Adapted to High Affine Transformations [11.300618381337777]
This paper proposes a region feature descriptor based on simulating affine transformations using classification.
It exhibits higher precision and robustness compared to existing classical descriptors.
arXiv Detail & Related papers (2024-02-15T05:33:58Z) - LGFCTR: Local and Global Feature Convolutional Transformer for Image
Matching [8.503217766507584]
A novel convolutional transformer is proposed to capture both local contexts and global structures.
A universal FPN-like framework captures global structures in self-encoder as well as cross-decoder by transformers.
A novel regression-based sub-pixel refinement module exploits the whole fine-grained window features for fine-level positional deviation regression.
arXiv Detail & Related papers (2023-11-29T12:06:19Z) - Learning Local Implicit Fourier Representation for Image Warping [11.526109213908091]
We propose a local texture estimator for image warping (LTEW) followed by an implicit neural representation to deform images into continuous shapes.
Our LTEW-based neural function outperforms existing warping methods for asymmetric-scale SR and homography transform.
arXiv Detail & Related papers (2022-07-05T06:30:17Z) - 3D Equivariant Graph Implicit Functions [51.5559264447605]
We introduce a novel family of graph implicit functions with equivariant layers that facilitates modeling fine local details.
Our method improves over the existing rotation-equivariant implicit function from 0.69 to 0.89 on the ShapeNet reconstruction task.
arXiv Detail & Related papers (2022-03-31T16:51:25Z) - LocalViT: Bringing Locality to Vision Transformers [132.42018183859483]
locality is essential for images since it pertains to structures like lines, edges, shapes, and even objects.
We add locality to vision transformers by introducing depth-wise convolution into the feed-forward network.
This seemingly simple solution is inspired by the comparison between feed-forward networks and inverted residual blocks.
arXiv Detail & Related papers (2021-04-12T17:59:22Z) - Feature Pyramid Transformer [121.50066435635118]
We propose a fully active feature interaction across both space and scales, called Feature Pyramid Transformer (FPT)
FPT transforms any feature pyramid into another feature pyramid of the same size but with richer contexts.
We conduct extensive experiments in both instance-level (i.e., object detection and instance segmentation) and pixel-level segmentation tasks.
arXiv Detail & Related papers (2020-07-18T15:16:32Z) - Inability of spatial transformations of CNN feature maps to support
invariant recognition [0.0]
We show that spatial transformations of CNN feature maps cannot align the feature maps of a transformed image to match those of its original.
For rotations and reflections, spatially transforming feature maps or filters can enable invariance but only for networks with learnt or hardcoded rotation- or reflection-invariant features.
arXiv Detail & Related papers (2020-04-30T12:12:58Z) - Understanding when spatial transformer networks do not support
invariance, and what to do about it [0.0]
spatial transformer networks (STNs) were designed to enable convolutional neural networks (CNNs) to learn invariance to image transformations.
We show that STNs do not have the ability to align the feature maps of a transformed image with those of its original.
We investigate alternative STN architectures that make use of complex features.
arXiv Detail & Related papers (2020-04-24T12:20:35Z) - Steering Self-Supervised Feature Learning Beyond Local Pixel Statistics [60.92229707497999]
We introduce a novel principle for self-supervised feature learning based on the discrimination of specific transformations of an image.
We demonstrate experimentally that learning to discriminate transformations such as LCI, image warping and rotations, yields features with state of the art generalization capabilities.
arXiv Detail & Related papers (2020-04-05T22:09:08Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.