MF-MOS: A Motion-Focused Model for Moving Object Segmentation
- URL: http://arxiv.org/abs/2401.17023v1
- Date: Tue, 30 Jan 2024 13:55:56 GMT
- Title: MF-MOS: A Motion-Focused Model for Moving Object Segmentation
- Authors: Jintao Cheng, Kang Zeng, Zhuoxu Huang, Xiaoyu Tang, Jin Wu, Chengxi
Zhang, Xieyuanli Chen, Rui Fan
- Abstract summary: Moving object segmentation (MOS) provides a reliable solution for detecting traffic participants.
Previous methods capture motion features from the range images directly.
We propose MF-MOS, a novel motion-focused model with a dual-branch structure for LiDAR moving object segmentation.
- Score: 10.533968185642415
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Moving object segmentation (MOS) provides a reliable solution for detecting
traffic participants and thus is of great interest in the autonomous driving
field. Dynamic capture is always critical in the MOS problem. Previous methods
capture motion features from the range images directly. Differently, we argue
that the residual maps provide greater potential for motion information, while
range images contain rich semantic guidance. Based on this intuition, we
propose MF-MOS, a novel motion-focused model with a dual-branch structure for
LiDAR moving object segmentation. Novelly, we decouple the spatial-temporal
information by capturing the motion from residual maps and generating semantic
features from range images, which are used as movable object guidance for the
motion branch. Our straightforward yet distinctive solution can make the most
use of both range images and residual maps, thus greatly improving the
performance of the LiDAR-based MOS task. Remarkably, our MF-MOS achieved a
leading IoU of 76.7% on the MOS leaderboard of the SemanticKITTI dataset upon
submission, demonstrating the current state-of-the-art performance. The
implementation of our MF-MOS has been released at
https://github.com/SCNU-RISLAB/MF-MOS.
Related papers
- CV-MOS: A Cross-View Model for Motion Segmentation [13.378850442525945]
We introduce CV-MOS, a cross-view model for moving object segmentation.
We decouple spatial-temporal information by capturing the motion from BEV and RV residual maps.
Our method achieved leading IoU(%) scores of 77.5% and 79.2% on the validation and test sets of the SemanticKitti dataset.
arXiv Detail & Related papers (2024-08-25T09:39:26Z) - MV-MOS: Multi-View Feature Fusion for 3D Moving Object Segmentation [4.386035726986601]
How to effectively utilize motion and semantic features and avoid information loss during 3D-to-2D projection is still a key challenge.
We propose a novel multi-view MOS model (MV-MOS) by fusing motion-semantic features from different 2D representations of point clouds.
We validated the effectiveness of the proposed multi-branch fusion MOS framework via comprehensive experiments.
arXiv Detail & Related papers (2024-08-20T07:30:00Z) - MambaMOS: LiDAR-based 3D Moving Object Segmentation with Motion-aware State Space Model [15.418115686945056]
LiDAR-based Moving Object (MOS) aims to locate and segment moving objects in point clouds of the current scan using motion information from previous scans.
We propose a novel LiDAR-based 3D Moving Object with Motion-aware State Space Model, termed MambaMOS.
arXiv Detail & Related papers (2024-04-19T11:17:35Z) - Delving into Motion-Aware Matching for Monocular 3D Object Tracking [81.68608983602581]
We find that the motion cue of objects along different time frames is critical in 3D multi-object tracking.
We propose MoMA-M3T, a framework that mainly consists of three motion-aware components.
We conduct extensive experiments on the nuScenes and KITTI datasets to demonstrate our MoMA-M3T achieves competitive performance against state-of-the-art methods.
arXiv Detail & Related papers (2023-08-22T17:53:58Z) - MotionBEV: Attention-Aware Online LiDAR Moving Object Segmentation with
Bird's Eye View based Appearance and Motion Features [5.186531650935954]
We present MotionBEV, a fast and accurate framework for LiDAR moving object segmentation.
Our approach converts 3D LiDAR scans into a 2D polar BEV representation to improve computational efficiency.
We employ a dual-branch network bridged by the Appearance-Motion Co-attention Module (AMCM) to adaptively fuse the LiDAR-temporal information from appearance and motion features.
arXiv Detail & Related papers (2023-05-12T09:28:09Z) - Event-Free Moving Object Segmentation from Moving Ego Vehicle [88.33470650615162]
Moving object segmentation (MOS) in dynamic scenes is an important, challenging, but under-explored research topic for autonomous driving.
Most segmentation methods leverage motion cues obtained from optical flow maps.
We propose to exploit event cameras for better video understanding, which provide rich motion cues without relying on optical flow.
arXiv Detail & Related papers (2023-04-28T23:43:10Z) - MoLo: Motion-augmented Long-short Contrastive Learning for Few-shot
Action Recognition [50.345327516891615]
We develop a Motion-augmented Long-short Contrastive Learning (MoLo) method that contains two crucial components, including a long-short contrastive objective and a motion autodecoder.
MoLo can simultaneously learn long-range temporal context and motion cues for comprehensive few-shot matching.
arXiv Detail & Related papers (2023-04-03T13:09:39Z) - MotionTrack: Learning Robust Short-term and Long-term Motions for
Multi-Object Tracking [56.92165669843006]
We propose MotionTrack, which learns robust short-term and long-term motions in a unified framework to associate trajectories from a short to long range.
For dense crowds, we design a novel Interaction Module to learn interaction-aware motions from short-term trajectories, which can estimate the complex movement of each target.
For extreme occlusions, we build a novel Refind Module to learn reliable long-term motions from the target's history trajectory, which can link the interrupted trajectory with its corresponding detection.
arXiv Detail & Related papers (2023-03-18T12:38:33Z) - Unsupervised 4D LiDAR Moving Object Segmentation in Stationary Settings
with Multivariate Occupancy Time Series [62.997667081978825]
We address the problem of unsupervised moving object segmentation (MOS) in 4D LiDAR data recorded from a stationary sensor, where no ground truth annotations are involved.
We propose a novel 4D LiDAR representation based on a time series that relaxes the problem of unsupervised MOS.
Experiments on stationary scenes from the Raw KITTI dataset show that our fully unsupervised approach achieves performance that is comparable to that of supervised state-of-the-art approaches.
arXiv Detail & Related papers (2022-12-30T14:48:14Z) - Exploring Motion and Appearance Information for Temporal Sentence
Grounding [52.01687915910648]
We propose a Motion-Appearance Reasoning Network (MARN) to solve temporal sentence grounding.
We develop separate motion and appearance branches to learn motion-guided and appearance-guided object relations.
Our proposed MARN significantly outperforms previous state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2022-01-03T02:44:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.