Label Correction for Road Segmentation Using Road-side Cameras
- URL: http://arxiv.org/abs/2502.01281v1
- Date: Mon, 03 Feb 2025 11:52:23 GMT
- Title: Label Correction for Road Segmentation Using Road-side Cameras
- Authors: Henrik Toikka, Eerik Alamikkotervo, Risto Ojala,
- Abstract summary: Existing roadside camera infrastructure is utilized for collecting road data in varying weather conditions automatically.
A novel semi-automatic annotation method for roadside cameras is proposed.
The proposed method is validated with roadside camera data collected from 927 cameras across Finland over 4 month time period during winter.
- Score: 0.44241702149260353
- License:
- Abstract: Reliable road segmentation in all weather conditions is critical for intelligent transportation applications, autonomous vehicles and advanced driver's assistance systems. For robust performance, all weather conditions should be included in the training data of deep learning-based perception models. However, collecting and annotating such a dataset requires extensive resources. In this paper, existing roadside camera infrastructure is utilized for collecting road data in varying weather conditions automatically. Additionally, a novel semi-automatic annotation method for roadside cameras is proposed. For each camera, only one frame is labeled manually and then the label is transferred to other frames of that camera feed. The small camera movements between frames are compensated using frequency domain image registration. The proposed method is validated with roadside camera data collected from 927 cameras across Finland over 4 month time period during winter. Training on the semi-automatically labeled data boosted the segmentation performance of several deep learning segmentation models. Testing was carried out on two different datasets to evaluate the robustness of the resulting models. These datasets were an in-domain roadside camera dataset and out-of-domain dataset captured with a vehicle on-board camera.
Related papers
- Trajectory-based Road Autolabeling with Lidar-Camera Fusion in Winter Conditions [1.3724491757145387]
Trajectory-based self-supervised methods can learn from traversed route without manual labels.
Our method outperforms recent standalone camera- and lidar-based methods when evaluated with a challenging winter driving dataset.
arXiv Detail & Related papers (2024-12-03T10:54:37Z) - Redundancy-Aware Camera Selection for Indoor Scene Neural Rendering [54.468355408388675]
We build a similarity matrix that incorporates both the spatial diversity of the cameras and the semantic variation of the images.
We apply a diversity-based sampling algorithm to optimize the camera selection.
We also develop a new dataset, IndoorTraj, which includes long and complex camera movements captured by humans in virtual indoor environments.
arXiv Detail & Related papers (2024-09-11T08:36:49Z) - XLD: A Cross-Lane Dataset for Benchmarking Novel Driving View Synthesis [84.23233209017192]
This paper presents a novel driving view synthesis dataset and benchmark specifically designed for autonomous driving simulations.
The dataset is unique as it includes testing images captured by deviating from the training trajectory by 1-4 meters.
We establish the first realistic benchmark for evaluating existing NVS approaches under front-only and multi-camera settings.
arXiv Detail & Related papers (2024-06-26T14:00:21Z) - City-Scale Multi-Camera Vehicle Tracking System with Improved Self-Supervised Camera Link Model [0.0]
This article introduces an innovative multi-camera vehicle tracking system that utilizes a self-supervised camera link model.
The proposed method achieves a new state-of-the-art among automatic camera-link based methods in CityFlow V2 benchmarks with 61.07% IDF1 Score.
arXiv Detail & Related papers (2024-05-18T17:28:35Z) - TEDNet: Twin Encoder Decoder Neural Network for 2D Camera and LiDAR Road Detection [2.8038082486377114]
A novel Convolutional Neural Network model is proposed for the accurate estimation of the roadway surface.
Our model is based on the use of a Twin-Decoder Neural Network (TEDNet) for independent camera and LiDAR feature extraction.
Bird's Eye View projections of the camera and LiDAR data are used in this model to perform semantic segmentation on whether each pixel belongs to the road surface.
arXiv Detail & Related papers (2024-05-14T08:45:34Z) - Homography Guided Temporal Fusion for Road Line and Marking Segmentation [73.47092021519245]
Road lines and markings are frequently occluded in the presence of moving vehicles, shadow, and glare.
We propose a Homography Guided Fusion (HomoFusion) module to exploit temporally-adjacent video frames for complementary cues.
We show that exploiting available camera intrinsic data and ground plane assumption for cross-frame correspondence can lead to a light-weight network with significantly improved performances in speed and accuracy.
arXiv Detail & Related papers (2024-04-11T10:26:40Z) - Towards Viewpoint Robustness in Bird's Eye View Segmentation [85.99907496019972]
We study how AV perception models are affected by changes in camera viewpoint.
Small changes to pitch, yaw, depth, or height of the camera at inference time lead to large drops in performance.
We introduce a technique for novel view synthesis and use it to transform collected data to the viewpoint of target rigs.
arXiv Detail & Related papers (2023-09-11T02:10:07Z) - DualCam: A Novel Benchmark Dataset for Fine-grained Real-time Traffic
Light Detection [0.7130302992490973]
We introduce a novel benchmark traffic light dataset captured using a synchronized pair of narrow-angle and wide-angle cameras.
The dataset includes images of resolution 1920$times$1080 covering 10 different classes.
Results show that our technique can strike a balance between speed and accuracy, compared to the conventional approach of using a single camera frame.
arXiv Detail & Related papers (2022-09-03T08:02:55Z) - Extrinsic Camera Calibration with Semantic Segmentation [60.330549990863624]
We present an extrinsic camera calibration approach that automatizes the parameter estimation by utilizing semantic segmentation information.
Our approach relies on a coarse initial measurement of the camera pose and builds on lidar sensors mounted on a vehicle.
We evaluate our method on simulated and real-world data to demonstrate low error measurements in the calibration results.
arXiv Detail & Related papers (2022-08-08T07:25:03Z) - Ithaca365: Dataset and Driving Perception under Repeated and Challenging
Weather Conditions [0.0]
We present a new dataset to enable robust autonomous driving via a novel data collection process.
The dataset includes images and point clouds from cameras and LiDAR sensors, along with high-precision GPS/INS.
We demonstrate the uniqueness of this dataset by analyzing the performance of baselines in amodal segmentation of road and objects.
arXiv Detail & Related papers (2022-08-01T22:55:32Z) - Cross-Camera Trajectories Help Person Retrieval in a Camera Network [124.65912458467643]
Existing methods often rely on purely visual matching or consider temporal constraints but ignore the spatial information of the camera network.
We propose a pedestrian retrieval framework based on cross-camera generation, which integrates both temporal and spatial information.
To verify the effectiveness of our method, we construct the first cross-camera pedestrian trajectory dataset.
arXiv Detail & Related papers (2022-04-27T13:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.