D-VPnet: A Network for Real-time Dominant Vanishing Point Detection in
Natural Scenes
- URL: http://arxiv.org/abs/2006.05407v1
- Date: Tue, 9 Jun 2020 17:12:27 GMT
- Title: D-VPnet: A Network for Real-time Dominant Vanishing Point Detection in
Natural Scenes
- Authors: Yin-Bo Liu, Ming Zeng, Qing-Hao Meng
- Abstract summary: Vanishing points (VPs) provide useful clues for mapping objects from 2D photos to 3D space.
We present a new convolutional neural network (CNN) to detect dominant VPs in natural scenes.
- Score: 3.8170259685864165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As an important part of linear perspective, vanishing points (VPs) provide
useful clues for mapping objects from 2D photos to 3D space. Existing methods
are mainly focused on extracting structural features such as lines or contours
and then clustering these features to detect VPs. However, these techniques
suffer from ambiguous information due to the large number of line segments and
contours detected in outdoor environments. In this paper, we present a new
convolutional neural network (CNN) to detect dominant VPs in natural scenes,
i.e., the Dominant Vanishing Point detection Network (D-VPnet). The key
component of our method is the feature line-segment proposal unit (FLPU), which
can be directly utilized to predict the location of the dominant VP. Moreover,
the model also uses the two main parallel lines as an assistant to determine
the position of the dominant VP. The proposed method was tested using a public
dataset and a Parallel Line based Vanishing Point (PLVP) dataset. The
experimental results suggest that the detection accuracy of our approach
outperforms those of state-of-the-art methods under various conditions in
real-time, achieving rates of 115fps.
Related papers
- Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - PointOcc: Cylindrical Tri-Perspective View for Point-based 3D Semantic
Occupancy Prediction [72.75478398447396]
We propose a cylindrical tri-perspective view to represent point clouds effectively and comprehensively.
Considering the distance distribution of LiDAR point clouds, we construct the tri-perspective view in the cylindrical coordinate system.
We employ spatial group pooling to maintain structural details during projection and adopt 2D backbones to efficiently process each TPV plane.
arXiv Detail & Related papers (2023-08-31T17:57:17Z) - DPA-P2PNet: Deformable Proposal-aware P2PNet for Accurate Point-based
Cell Detection [5.994317314012678]
Point-based cell detection (PCD) has garnered increased attention in computational pathology community.
Unlike mainstream PCD methods that rely on intermediate density map representations, the Point-to-Point network (P2PNet) has recently emerged as an end-to-end solution for PCD.
P2PNet is limited to decoding from a single-level feature map due to the scale-agnostic property of point proposals.
The proposed method directly extracts multi-scale features for decoding according to the coordinates of point proposals on hierarchical feature maps.
arXiv Detail & Related papers (2023-03-05T08:02:42Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - Learning point embedding for 3D data processing [2.12121796606941]
Current point-based methods are essentially spatial relationship processing networks.
Our architecture, PE-Net, learns the representation of point clouds in high-dimensional space.
Experiments show that PE-Net achieves the state-of-the-art performance in multiple challenging datasets.
arXiv Detail & Related papers (2021-07-19T00:25:28Z) - PV-RCNN++: Point-Voxel Feature Set Abstraction With Local Vector
Representation for 3D Object Detection [100.60209139039472]
We propose the PointVoxel Region based Convolution Neural Networks (PVRCNNs) for accurate 3D detection from point clouds.
Our proposed PV-RCNNs significantly outperform previous state-of-the-art 3D detection methods on both the Open dataset and the highly-competitive KITTI benchmark.
arXiv Detail & Related papers (2021-01-31T14:51:49Z) - ODFNet: Using orientation distribution functions to characterize 3D
point clouds [0.0]
We leverage on point orientation distributions around a point in order to obtain an expressive local neighborhood representation for point clouds.
New ODFNet model achieves state-of-the-art accuracy for object classification on ModelNet40 and ScanObjectNN datasets.
arXiv Detail & Related papers (2020-12-08T19:54:20Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z) - Cylindrical Convolutional Networks for Joint Object Detection and
Viewpoint Estimation [76.21696417873311]
We introduce a learnable module, cylindrical convolutional networks (CCNs), that exploit cylindrical representation of a convolutional kernel defined in the 3D space.
CCNs extract a view-specific feature through a view-specific convolutional kernel to predict object category scores at each viewpoint.
Our experiments demonstrate the effectiveness of the cylindrical convolutional networks on joint object detection and viewpoint estimation.
arXiv Detail & Related papers (2020-03-25T10:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.