The Robust Semantic Segmentation UNCV2023 Challenge Results
- URL: http://arxiv.org/abs/2309.15478v1
- Date: Wed, 27 Sep 2023 08:20:03 GMT
- Title: The Robust Semantic Segmentation UNCV2023 Challenge Results
- Authors: Xuanlong Yu, Yi Zuo, Zitao Wang, Xiaowen Zhang, Jiaxuan Zhao, Yuting
Yang, Licheng Jiao, Rui Peng, Xinyi Wang, Junpei Zhang, Kexin Zhang, Fang
Liu, Roberto Alcover-Couso, Juan C. SanMiguel, Marcos Escudero-Vi\~nolo,
Hanlin Tian, Kenta Matsui, Tianhao Wang, Fahmy Adan, Zhitong Gao, Xuming He,
Quentin Bouniot, Hossein Moghaddam, Shyam Nandan Rai, Fabio Cermelli, Carlo
Masone, Andrea Pilzer, Elisa Ricci, Andrei Bursuc, Arno Solin, Martin Trapp,
Rui Li, Angela Yao, Wenlong Chen, Ivor Simpson, Neill D. F. Campbell, Gianni
Franchi
- Abstract summary: This paper outlines the winning solutions employed in addressing the MUAD uncertainty quantification challenge held at ICCV 2023.
The challenge was centered around semantic segmentation in urban environments, with a particular focus on natural adversarial scenarios.
The report presents the results of 19 submitted entries, with numerous techniques drawing inspiration from cutting-edge uncertainty quantification methodologies.
- Score: 99.97867942388486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper outlines the winning solutions employed in addressing the MUAD
uncertainty quantification challenge held at ICCV 2023. The challenge was
centered around semantic segmentation in urban environments, with a particular
focus on natural adversarial scenarios. The report presents the results of 19
submitted entries, with numerous techniques drawing inspiration from
cutting-edge uncertainty quantification methodologies presented at prominent
conferences in the fields of computer vision and machine learning and journals
over the past few years. Within this document, the challenge is introduced,
shedding light on its purpose and objectives, which primarily revolved around
enhancing the robustness of semantic segmentation in urban scenes under varying
natural adversarial conditions. The report then delves into the top-performing
solutions. Moreover, the document aims to provide a comprehensive overview of
the diverse solutions deployed by all participants. By doing so, it seeks to
offer readers a deeper insight into the array of strategies that can be
leveraged to effectively handle the inherent uncertainties associated with
autonomous driving and semantic segmentation, especially within urban
environments.
Related papers
- RescueADI: Adaptive Disaster Interpretation in Remote Sensing Images with Autonomous Agents [11.08910129925713]
This paper introduces Adaptive Disaster Interpretation (ADI), a novel task designed to solve requests by planning and executing multiple correlative interpretation tasks.
We present a new dataset named RescueADI, which contains high-resolution RSIs with annotations for three connected aspects: planning, perception, and recognition.
We propose a new disaster interpretation method employing autonomous agents driven by large language models (LLMs) for task planning and execution.
arXiv Detail & Related papers (2024-10-17T09:36:52Z) - Solving the Challenge Set without Solving the Task: On Winograd Schemas as a Test of Pronominal Coreference Resolution [21.19369044026899]
We show that despite the strong performance of prompted language models (LMs) on the Winograd Challenge set, these same modeling techniques perform relatively poorly at resolving certain pronominal ambiguities attested in OntoNotes.
We propose a method for ensembling a prompted LM with a supervised, task-specific system that is overall more accurate at resolving pronominal coreference across datasets.
arXiv Detail & Related papers (2024-10-12T09:04:53Z) - The BRAVO Semantic Segmentation Challenge Results in UNCV2024 [68.20197719071436]
We define two categories of reliability: (1) semantic reliability, which reflects the model's accuracy and calibration when exposed to various perturbations; and (2) OOD reliability, which measures the model's ability to detect object classes that are unknown during training.
The results reveal interesting insights into the importance of large-scale pre-training and minimal architectural design in developing robust and reliable semantic segmentation models.
arXiv Detail & Related papers (2024-09-23T15:17:30Z) - The Third Monocular Depth Estimation Challenge [134.16634233789776]
This paper discusses the results of the third edition of the Monocular Depth Estimation Challenge (MDEC)
The challenge focuses on zero-shot generalization to the challenging SYNS-Patches dataset, featuring complex scenes in natural and indoor settings.
The challenge winners drastically improved 3D F-Score performance, from 17.51% to 23.72%.
arXiv Detail & Related papers (2024-04-25T17:59:59Z) - Object Detectors in the Open Environment: Challenges, Solutions, and Outlook [95.3317059617271]
The dynamic and intricate nature of the open environment poses novel and formidable challenges to object detectors.
This paper aims to conduct a comprehensive review and analysis of object detectors in open environments.
We propose a framework that includes four quadrants (i.e., out-of-domain, out-of-category, robust learning, and incremental learning) based on the dimensions of the data / target changes.
arXiv Detail & Related papers (2024-03-24T19:32:39Z) - Unmasking the Shadows of AI: Investigating Deceptive Capabilities in Large Language Models [0.0]
This research critically navigates the intricate landscape of AI deception, concentrating on deceptive behaviours of Large Language Models (LLMs)
My objective is to elucidate this issue, examine the discourse surrounding it, and subsequently delve into its categorization and ramifications.
arXiv Detail & Related papers (2024-02-07T00:21:46Z) - Progressive Feature Self-reinforcement for Weakly Supervised Semantic
Segmentation [55.69128107473125]
We propose a single-stage approach for Weakly Supervised Semantic (WSSS) with image-level labels.
We adaptively partition the image content into deterministic regions (e.g., confident foreground and background) and uncertain regions (e.g., object boundaries and misclassified categories) for separate processing.
Building upon this, we introduce a complementary self-enhancement method that constrains the semantic consistency between these confident regions and an augmented image with the same class labels.
arXiv Detail & Related papers (2023-12-14T13:21:52Z) - Semi-supervised multimodal coreference resolution in image narrations [44.66334603518387]
We study multimodal coreference resolution, specifically where a descriptive text is paired with an image.
This poses significant challenges due to fine-grained image-text alignment, inherent ambiguity present in narrative language, and unavailability of large annotated training sets.
We present a data efficient semi-supervised approach that utilizes image-narration pairs to resolve coreferences and narrative grounding in a multimodal context.
arXiv Detail & Related papers (2023-10-20T16:10:14Z) - EFaR 2023: Efficient Face Recognition Competition [51.77649060180531]
The paper presents the summary of the Efficient Face Recognition Competition (EFaR) held at the 2023 International Joint Conference on Biometrics (IJCB 2023)
The competition received 17 submissions from 6 different teams.
The submitted solutions are ranked based on a weighted score of the achieved verification accuracies on a diverse set of benchmarks, as well as the deployability given by the number of floating-point operations and model size.
arXiv Detail & Related papers (2023-08-08T09:58:22Z) - Multimodal Feature Extraction and Fusion for Emotional Reaction
Intensity Estimation and Expression Classification in Videos with
Transformers [47.16005553291036]
We present our solutions to the two sub-challenges of Affective Behavior Analysis in the wild (ABAW) 2023.
For the Expression Classification Challenge, we propose a streamlined approach that handles the challenges of classification effectively.
By studying, analyzing, and combining these features, we significantly enhance the model's accuracy for sentiment prediction in a multimodal context.
arXiv Detail & Related papers (2023-03-16T09:03:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.