OSP2B: One-Stage Point-to-Box Network for 3D Siamese Tracking
- URL: http://arxiv.org/abs/2304.11584v2
- Date: Tue, 9 May 2023 02:27:49 GMT
- Title: OSP2B: One-Stage Point-to-Box Network for 3D Siamese Tracking
- Authors: Jiahao Nie, Zhiwei He, Yuxiang Yang, Zhengyi Bao, Mingyu Gao, Jing
Zhang
- Abstract summary: Two-stage point-to-box network acts as a critical role in the recent popular 3D Siamese tracking paradigm.
We propose a simple yet effective one-stage point-to-box network for point cloud-based 3D single object tracking.
By integrating the derived classification scores with the center-ness scores, the resulting network can effectively suppress interference proposals.
- Score: 7.868399549570768
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Two-stage point-to-box network acts as a critical role in the recent popular
3D Siamese tracking paradigm, which first generates proposals and then predicts
corresponding proposal-wise scores. However, such a network suffers from
tedious hyper-parameter tuning and task misalignment, limiting the tracking
performance. Towards these concerns, we propose a simple yet effective
one-stage point-to-box network for point cloud-based 3D single object tracking.
It synchronizes 3D proposal generation and center-ness score prediction by a
parallel predictor without tedious hyper-parameters. To guide a task-aligned
score ranking of proposals, a center-aware focal loss is proposed to supervise
the training of the center-ness branch, which enhances the network's
discriminative ability to distinguish proposals of different quality. Besides,
we design a binary target classifier to identify target-relevant points. By
integrating the derived classification scores with the center-ness scores, the
resulting network can effectively suppress interference proposals and further
mitigate task misalignment. Finally, we present a novel one-stage Siamese
tracker OSP2B equipped with the designed network. Extensive experiments on
challenging benchmarks including KITTI and Waymo SOT Dataset show that our
OSP2B achieves leading performance with a considerable real-time speed.Code
will be available at https://github.com/haooozi/OSP2B.
Related papers
- P2Object: Single Point Supervised Object Detection and Instance Segmentation [58.778288785355]
We introduce Point-to-Box Network (P2BNet), which constructs balanced textbftextitinstance-level proposal bags
P2MNet can generate more precise bounding boxes and generalize to segmentation tasks.
Our method largely surpasses the previous methods in terms of the mean average precision on COCO, VOC, and Cityscapes.
arXiv Detail & Related papers (2025-04-10T14:51:08Z) - CS-Net:Contribution-based Sampling Network for Point Cloud Simplification [50.55658910053004]
Point cloud sampling plays a crucial role in reducing computation costs and storage requirements for various vision tasks.
Traditional sampling methods, such as farthest point sampling, lack task-specific information.
We propose a contribution-based sampling network (CS-Net), where the sampling operation is formulated as a Top-k operation.
arXiv Detail & Related papers (2025-01-18T14:56:09Z) - A Point-Based Approach to Efficient LiDAR Multi-Task Perception [49.91741677556553]
PAttFormer is an efficient multi-task architecture for joint semantic segmentation and object detection in point clouds.
Unlike other LiDAR-based multi-task architectures, our proposed PAttFormer does not require separate feature encoders for task-specific point cloud representations.
Our evaluations show substantial gains from multi-task learning, improving LiDAR semantic segmentation by +1.7% in mIou and 3D object detection by +1.7% in mAP.
arXiv Detail & Related papers (2024-04-19T11:24:34Z) - 3D Cascade RCNN: High Quality Object Detection in Point Clouds [122.42455210196262]
We present 3D Cascade RCNN, which allocates multiple detectors based on the voxelized point clouds in a cascade paradigm.
We validate the superiority of our proposed 3D Cascade RCNN, when comparing to state-of-the-art 3D object detection techniques.
arXiv Detail & Related papers (2022-11-15T15:58:36Z) - Point-to-Box Network for Accurate Object Detection via Single Point
Supervision [51.95993495703855]
We introduce a lightweight alternative to the off-the-shelf proposal (OTSP) method.
P2BNet can construct an inter-objects balanced proposal bag by generating proposals in an anchor-like way.
The code will be released at COCO.com/ucas-vg/P2BNet.
arXiv Detail & Related papers (2022-07-14T11:32:00Z) - Not All Points Are Equal: Learning Highly Efficient Point-based
Detectors for 3D LiDAR Point Clouds [29.762645632148097]
We propose a highly-efficient single-stage point-based 3D detector called IA-SSD.
We exploit two learnable, task-oriented, instance-aware downsampling strategies to hierarchically select the foreground points.
Experiments conducted on several large-scale detection benchmarks demonstrate the competitive performance of our IA-SSD.
arXiv Detail & Related papers (2022-03-21T17:14:02Z) - Guided Point Contrastive Learning for Semi-supervised Point Cloud
Semantic Segmentation [90.2445084743881]
We present a method for semi-supervised point cloud semantic segmentation to adopt unlabeled point clouds in training to boost the model performance.
Inspired by the recent contrastive loss in self-supervised tasks, we propose the guided point contrastive loss to enhance the feature representation and model generalization ability.
arXiv Detail & Related papers (2021-10-15T16:38:54Z) - 3D-SiamRPN: An End-to-End Learning Method for Real-Time 3D Single Object
Tracking Using Raw Point Cloud [9.513194898261787]
We propose a 3D tracking method called 3D-SiamRPN Network to track a single target object by using raw 3D point cloud data.
Experimental results on KITTI dataset show that our method has a competitive performance in both Success and Precision.
arXiv Detail & Related papers (2021-08-12T09:52:28Z) - Learning point embedding for 3D data processing [2.12121796606941]
Current point-based methods are essentially spatial relationship processing networks.
Our architecture, PE-Net, learns the representation of point clouds in high-dimensional space.
Experiments show that PE-Net achieves the state-of-the-art performance in multiple challenging datasets.
arXiv Detail & Related papers (2021-07-19T00:25:28Z) - FatNet: A Feature-attentive Network for 3D Point Cloud Processing [1.502579291513768]
We introduce a novel feature-attentive neural network layer, a FAT layer, that combines both global point-based features and local edge-based features in order to generate better embeddings.
Our architecture achieves state-of-the-art results on the task of point cloud classification, as demonstrated on the ModelNet40 dataset.
arXiv Detail & Related papers (2021-04-07T23:13:56Z) - PC-RGNN: Point Cloud Completion and Graph Neural Network for 3D Object
Detection [57.49788100647103]
LiDAR-based 3D object detection is an important task for autonomous driving.
Current approaches suffer from sparse and partial point clouds of distant and occluded objects.
In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions.
arXiv Detail & Related papers (2020-12-18T18:06:43Z) - Corner Proposal Network for Anchor-free, Two-stage Object Detection [174.59360147041673]
The goal of object detection is to determine the class and location of objects in an image.
This paper proposes a novel anchor-free, two-stage framework which first extracts a number of object proposals.
We demonstrate that these two stages are effective solutions for improving recall and precision.
arXiv Detail & Related papers (2020-07-27T19:04:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.