FisheyeSuperPoint: Keypoint Detection and Description Network for
Fisheye Images
- URL: http://arxiv.org/abs/2103.00191v1
- Date: Sat, 27 Feb 2021 11:26:34 GMT
- Title: FisheyeSuperPoint: Keypoint Detection and Description Network for
Fisheye Images
- Authors: Anna Konrad, Ciar\'an Eising, Ganesh Sistu, John McDonald, Rudi
Villing, Senthil Yogamani
- Abstract summary: Keypoint detection and description is a commonly used building block in computer vision systems.
SuperPoint is a self-supervised keypoint detector and descriptor that has achieved state-of-the-art results on homography estimation.
We introduce a fisheye adaptation pipeline to enable training on undistorted fisheye images.
- Score: 2.187613144178315
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Keypoint detection and description is a commonly used building block in
computer vision systems particularly for robotics and autonomous driving.
Recently CNN based approaches have surpassed classical methods in a number of
perception tasks. However, the majority of techniques to date have focused on
standard cameras with little consideration given to fisheye cameras which are
commonly used in autonomous driving. In this paper, we propose a novel training
and evaluation pipeline for fisheye images. We make use of SuperPoint as our
baseline which is a self-supervised keypoint detector and descriptor that has
achieved state-of-the-art results on homography estimation. We introduce a
fisheye adaptation pipeline to enable training on undistorted fisheye images.
We evaluate the performance on the HPatches benchmark, and, by introducing a
fisheye based evaluation methods for detection repeatability and descriptor
matching correctness on the Oxford RobotCar datasets.
Related papers
- RoFIR: Robust Fisheye Image Rectification Framework Impervious to Optical Center Deviation [88.54817424560056]
We propose a distortion vector map (DVM) that measures the degree and direction of local distortion.
By learning the DVM, the model can independently identify local distortions at each pixel without relying on global distortion patterns.
In the pre-training stage, it predicts the distortion vector map and perceives the local distortion features of each pixel.
In the fine-tuning stage, it predicts a pixel-wise flow map for deviated fisheye image rectification.
arXiv Detail & Related papers (2024-06-27T06:38:56Z) - SimFIR: A Simple Framework for Fisheye Image Rectification with
Self-supervised Representation Learning [105.01294305972037]
We introduce SimFIR, a framework for fisheye image rectification based on self-supervised representation learning.
To learn fine-grained distortion representations, we first split a fisheye image into multiple patches and extract their representations with a Vision Transformer.
The transfer performance on the downstream rectification task is remarkably boosted, which verifies the effectiveness of the learned representations.
arXiv Detail & Related papers (2023-08-17T15:20:17Z) - Streaming Object Detection on Fisheye Cameras for Automatic Parking [0.0]
We propose a real-time detection framework equipped with a dualflow perception module that can predict the future and alleviate the time-lag problem.
The standard bounding box is unsuitable for the object in fisheye camera images due to the strong radial distortion of the fisheye camera.
We propose a new periodic angle loss function to regress the angle of the box, which is the simple and accurate representation method of objects.
arXiv Detail & Related papers (2023-05-24T04:30:25Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - Pixel-level Correspondence for Self-Supervised Learning from Video [56.24439897867531]
Pixel-level Correspondence (PiCo) is a method for dense contrastive learning from video.
We validate PiCo on standard benchmarks, outperforming self-supervised baselines on multiple dense prediction tasks.
arXiv Detail & Related papers (2022-07-08T12:50:13Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - Fisheye Lens Camera based Autonomous Valet Parking System [3.461121828373003]
This paper proposes an efficient autonomous valet parking system utilizing only cameras which are the most widely used sensor.
Fisheye cameras which have a wider angle of view compared to pinhole cameras are used.
arXiv Detail & Related papers (2021-04-27T11:36:03Z) - SVDistNet: Self-Supervised Near-Field Distance Estimation on Surround
View Fisheye Cameras [30.480562747903186]
A 360deg perception of scene geometry is essential for automated driving, notably for parking and urban driving scenarios.
We present novel camera-geometry adaptive multi-scale convolutions which utilize the camera parameters as a conditional input.
We evaluate our approach on the Fisheye WoodScape surround-view dataset, significantly improving over previous approaches.
arXiv Detail & Related papers (2021-04-09T15:20:20Z) - Where to drive: free space detection with one fisheye camera [1.7499351967216341]
We propose to use synthetic training data based on Unity3D.
A five-pass algorithm is used to create a virtual fisheye camera.
The results indicate that synthetic fisheye images can be used in deep learning context.
arXiv Detail & Related papers (2020-11-11T14:36:45Z) - Neural Ray Surfaces for Self-Supervised Learning of Depth and Ego-motion [51.19260542887099]
We show that self-supervision can be used to learn accurate depth and ego-motion estimation without prior knowledge of the camera model.
Inspired by the geometric model of Grossberg and Nayar, we introduce Neural Ray Surfaces (NRS), convolutional networks that represent pixel-wise projection rays.
We demonstrate the use of NRS for self-supervised learning of visual odometry and depth estimation from raw videos obtained using a wide variety of camera systems.
arXiv Detail & Related papers (2020-08-15T02:29:13Z) - SilhoNet-Fisheye: Adaptation of A ROI Based Object Pose Estimation
Network to Monocular Fisheye Images [15.573003283204958]
We present a novel framework for adapting a ROI-based 6D object pose estimation method to work on full fisheye images.
We also contribute a fisheye image dataset, called UWHandles, with 6D object pose and 2D bounding box annotations.
arXiv Detail & Related papers (2020-02-27T19:57:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.