First Place Solution to the ECCV 2024 BRAVO Challenge: Evaluating Robustness of Vision Foundation Models for Semantic Segmentation
- URL: http://arxiv.org/abs/2409.17208v2
- Date: Tue, 8 Oct 2024 10:09:14 GMT
- Title: First Place Solution to the ECCV 2024 BRAVO Challenge: Evaluating Robustness of Vision Foundation Models for Semantic Segmentation
- Authors: Tommie Kerssies, Daan de Geus, Gijs Dubbelman,
- Abstract summary: We present the first place solution to the ECCV 2024 BRAVO Challenge.
A model is trained on Cityscapes and its robustness is evaluated on several out-of-distribution datasets.
This approach outperforms more complex existing approaches, and achieves first place in the challenge.
- Score: 1.8570591025615457
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this report, we present the first place solution to the ECCV 2024 BRAVO Challenge, where a model is trained on Cityscapes and its robustness is evaluated on several out-of-distribution datasets. Our solution leverages the powerful representations learned by vision foundation models, by attaching a simple segmentation decoder to DINOv2 and fine-tuning the entire model. This approach outperforms more complex existing approaches, and achieves first place in the challenge. Our code is publicly available at https://github.com/tue-mps/benchmark-vfm-ss.
Related papers
- Precise Drive with VLM: First Prize Solution for PRCV 2024 Drive LM challenge [8.941623670652389]
This report outlines the methodologies we applied for the PRCV Challenge.
It focuses on cognition and decision-making in driving scenarios.
Our model achieved a score of 0.6064, securing the first prize on the competition's final results.
arXiv Detail & Related papers (2024-11-05T11:00:55Z) - Solution for OOD-CV UNICORN Challenge 2024 Object Detection Assistance LLM Counting Ability Improvement [6.621745547882088]
This report provides a detailed description of the method that we explored and proposed in the ECCV OOD-CV UNICORN Challenge 2024.
The dataset of this competition are OODCA-VQA and SketchyQA.
Our approach ranked second in the final test with a score of 0.86.
arXiv Detail & Related papers (2024-10-05T15:11:47Z) - Zero-Shot Anomaly Detection with Pre-trained Segmentation Models [2.9322869014189985]
This report outlines our submission to the zero-shot track of the Visual Anomaly and Novelty Detection (VAND) 2023 Challenge.
Building on the performance of the WINCLIP framework, we aim to enhance the system's localization capabilities by integrating zero-shot segmentation models.
Our pipeline requires no external data or information, allowing for it to be directly applied to new datasets.
arXiv Detail & Related papers (2023-06-15T16:43:07Z) - 2nd Place Winning Solution for the CVPR2023 Visual Anomaly and Novelty
Detection Challenge: Multimodal Prompting for Data-centric Anomaly Detection [10.682758791557436]
This report introduces the winning solution of the team Segment Any Anomaly for the CVPR2023 Visual Anomaly and Novelty Detection (VAND) challenge.
We present a novel framework, i.e., Segment Any Anomaly + (SAA$+$), for zero-shot anomaly segmentation with multi-modal prompts.
We will release the code of our winning solution for the CVPR2023 VAN.
arXiv Detail & Related papers (2023-06-15T11:49:44Z) - 3rd Place Solution for PVUW2023 VSS Track: A Large Model for Semantic
Segmentation on VSPW [68.56017675820897]
In this paper, we introduce 3rd place solution for PVUW2023 VSS track.
We have explored various image-level visual backbones and segmentation heads to tackle the problem of video semantic segmentation.
arXiv Detail & Related papers (2023-06-04T07:50:38Z) - 1st Place Solution for ECCV 2022 OOD-CV Challenge Object Detection Track [71.12470906323298]
Generalize-then-Adapt (G&A) framework is composed of a two-stage domain generalization part and a one-stage domain adaptation part.
The proposed G&A framework help us achieve the first place on the object detection leaderboard of the OOD-CV challenge.
arXiv Detail & Related papers (2023-01-12T03:50:46Z) - 1st Place Solution of The Robust Vision Challenge (RVC) 2022 Semantic
Segmentation Track [67.56316745239629]
This report describes the winning solution to the semantic segmentation task of the Robust Vision Challenge on ECCV 2022.
Our method adopts the FAN-B-Hybrid model as the encoder and uses Segformer as the segmentation framework.
The proposed method could serve as a strong baseline for the multi-domain segmentation task and benefit future works.
arXiv Detail & Related papers (2022-10-23T20:52:22Z) - Two-Stream Consensus Network: Submission to HACS Challenge 2021
Weakly-Supervised Learning Track [78.64815984927425]
The goal of weakly-supervised temporal action localization is to temporally locate and classify action of interest in untrimmed videos.
We adopt the two-stream consensus network (TSCN) as the main framework in this challenge.
Our solution ranked 2rd in this challenge, and we hope our method can serve as a baseline for future academic research.
arXiv Detail & Related papers (2021-06-21T03:36:36Z) - Large-scale Unsupervised Semantic Segmentation [163.3568726730319]
We propose a new problem of large-scale unsupervised semantic segmentation (LUSS) with a newly created benchmark dataset to track the research progress.
Based on the ImageNet dataset, we propose the ImageNet-S dataset with 1.2 million training images and 40k high-quality semantic segmentation annotations for evaluation.
arXiv Detail & Related papers (2021-06-06T15:02:11Z) - 1st Place Solutions for OpenImage2019 -- Object Detection and Instance
Segmentation [116.25081559037872]
This article introduces the solutions of the two champion teams, MMfruit' for the detection track and MMfruitSeg' for the segmentation track, in OpenImage Challenge 2019.
It is commonly known that for an object detector, the shared feature at the end of the backbone is not appropriate for both classification and regression.
We propose the Decoupling Head (DH) to disentangle the object classification and regression via the self-learned optimal feature extraction.
arXiv Detail & Related papers (2020-03-17T06:45:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.