A Two-Stage Adverse Weather Semantic Segmentation Method for WeatherProof Challenge CVPR 2024 Workshop UG2+
- URL: http://arxiv.org/abs/2406.05513v2
- Date: Thu, 11 Jul 2024 02:48:22 GMT
- Title: A Two-Stage Adverse Weather Semantic Segmentation Method for WeatherProof Challenge CVPR 2024 Workshop UG2+
- Authors: Jianzhao Wang, Yanyan Wei, Dehua Hu, Yilin Zhang, Shengeng Tang, Kun Li, Zhao Zhang,
- Abstract summary: We propose a two-stage deep learning framework for the WeatherProof dataset challenge.
In the challenge, our solution achieved a competitive score of 0.43 on the Mean Intersection over Union (mIoU) metric, securing a respectable rank of 4th.
- Score: 10.069192320623031
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This technical report presents our team's solution for the WeatherProof Dataset Challenge: Semantic Segmentation in Adverse Weather at CVPR'24 UG2+. We propose a two-stage deep learning framework for this task. In the first stage, we preprocess the provided dataset by concatenating images into video sequences. Subsequently, we leverage a low-rank video deraining method to generate high-fidelity pseudo ground truths. These pseudo ground truths offer superior alignment compared to the original ground truths, facilitating model convergence during training. In the second stage, we employ the InternImage network to train for the semantic segmentation task using the generated pseudo ground truths. Notably, our meticulously designed framework demonstrates robustness to degraded data captured under adverse weather conditions. In the challenge, our solution achieved a competitive score of 0.43 on the Mean Intersection over Union (mIoU) metric, securing a respectable rank of 4th.
Related papers
- AIM 2024 Sparse Neural Rendering Challenge: Methods and Results [64.19942455360068]
This paper reviews the challenge on Sparse Neural Rendering that was part of the Advances in Image Manipulation (AIM) workshop, held in conjunction with ECCV 2024.
The challenge aims at producing novel camera view synthesis of diverse scenes from sparse image observations.
Participants are asked to optimise objective fidelity to the ground-truth images as measured via the Peak Signal-to-Noise Ratio (PSNR) metric.
arXiv Detail & Related papers (2024-09-23T14:17:40Z) - 1st Place Solution for MeViS Track in CVPR 2024 PVUW Workshop: Motion Expression guided Video Segmentation [81.50620771207329]
We investigate the effectiveness of static-dominant data and frame sampling on referring video object segmentation (RVOS)
Our solution achieves a J&F score of 0.5447 in the competition phase and ranks 1st in the MeViS track of the PVUW Challenge.
arXiv Detail & Related papers (2024-06-11T08:05:26Z) - Technical Report for CVPR 2024 WeatherProof Dataset Challenge: Semantic Segmentation on Paired Real Data [9.128113804878959]
This challenge aims at semantic segmentation of images degraded by various degrees of weather from all around the world.
We introduced a pre-trained large-scale vision foundation model: InternImage, and trained it using images with different levels of noise.
As a result, we achieved 2nd place in the challenge with 45.1 mIOU and fewer submissions than the other winners.
arXiv Detail & Related papers (2024-06-09T17:08:07Z) - Solution for CVPR 2024 UG2+ Challenge Track on All Weather Semantic Segmentation [9.322345758563886]
We present our solution for the semantic segmentation in adverse weather, in UG2+ Challenge at CVPR 2024.
We initialize the InternImage-H backbone with pre-trained weights from the large-scale joint dataset and enhance it with the state-of-the-art Upernet segmentation method.
Our proposed solution demonstrates advanced performance on the test set and achieves 3rd position in this challenge.
arXiv Detail & Related papers (2024-06-09T15:56:35Z) - Enhanced Semantic Segmentation Pipeline for WeatherProof Dataset Challenge [2.722639169181398]
This report describes the winning solution to the WeatherProof dataset challenge.
We propose an enhanced semantic segmentation pipeline for this challenge.
We introduce a new dataset WeatherProofExtra with wider viewing angle and employ data augmentation methods.
Our solution is ranked 1st on the final leaderboard.
arXiv Detail & Related papers (2024-06-06T07:12:50Z) - WeatherProof: Leveraging Language Guidance for Semantic Segmentation in Adverse Weather [8.902960772665482]
We propose a method to infer semantic segmentation maps from images captured under adverse weather conditions.
We begin by examining existing models on images degraded by weather conditions such as rain, fog, or snow.
We propose WeatherProof, the first semantic segmentation dataset with accurate clear and adverse weather image pairs.
arXiv Detail & Related papers (2024-03-21T22:46:27Z) - Genuine Knowledge from Practice: Diffusion Test-Time Adaptation for
Video Adverse Weather Removal [53.15046196592023]
We introduce test-time adaptation into adverse weather removal in videos.
We propose the first framework that integrates test-time adaptation into the iterative diffusion reverse process.
arXiv Detail & Related papers (2024-03-12T14:21:30Z) - WeatherProof: A Paired-Dataset Approach to Semantic Segmentation in
Adverse Weather [9.619700283574533]
We introduce a general paired-training method that leads to improved performance on images in adverse weather conditions.
We create the first semantic segmentation dataset with accurate clear and adverse weather image pairs.
We find that training on these paired clear and adverse weather frames which share an underlying scene results in improved performance on adverse weather data.
arXiv Detail & Related papers (2023-12-15T04:57:54Z) - Robust Monocular Depth Estimation under Challenging Conditions [81.57697198031975]
State-of-the-art monocular depth estimation approaches are highly unreliable under challenging illumination and weather conditions.
We tackle these safety-critical issues with md4all: a simple and effective solution that works reliably under both adverse and ideal conditions.
arXiv Detail & Related papers (2023-08-18T17:59:01Z) - Self-Distilled Masked Auto-Encoders are Efficient Video Anomaly
Detectors [117.61449210940955]
We propose an efficient abnormal event detection model based on a lightweight masked auto-encoder (AE) applied at the video frame level.
We introduce an approach to weight tokens based on motion gradients, thus shifting the focus from the static background scene to the foreground objects.
We generate synthetic abnormal events to augment the training videos, and task the masked AE model to jointly reconstruct the original frames.
arXiv Detail & Related papers (2023-06-21T06:18:05Z) - 1st Place Solution for ECCV 2022 OOD-CV Challenge Image Classification
Track [64.49153847504141]
OOD-CV challenge is an out-of-distribution generalization task.
In this challenge, our core solution can be summarized as that Noisy Label Learning Is A Strong Test-Time Domain Adaptation method.
After integrating Test-Time Augmentation and Model Ensemble strategies, our solution ranks the first place on the Image Classification Leaderboard of the OOD-CV Challenge.
arXiv Detail & Related papers (2023-01-12T03:44:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.