Deconstructing Self-Supervised Monocular Reconstruction: The Design
Decisions that Matter
- URL: http://arxiv.org/abs/2208.01489v1
- Date: Tue, 2 Aug 2022 14:38:53 GMT
- Title: Deconstructing Self-Supervised Monocular Reconstruction: The Design
Decisions that Matter
- Authors: Jaime Spencer Martin, Chris Russell, Simon Hadfield, Richard Bowden
- Abstract summary: This paper presents a framework to evaluate state-of-the-art contributions to self-supervised monocular depth estimation.
It includes pretraining, backbone, architectural design choices and loss functions.
We re-implement, validate and re-evaluate 16 state-of-the-art contributions and introduce a new dataset.
- Score: 63.5550818034739
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper presents an open and comprehensive framework to systematically
evaluate state-of-the-art contributions to self-supervised monocular depth
estimation. This includes pretraining, backbone, architectural design choices
and loss functions. Many papers in this field claim novelty in either
architecture design or loss formulation. However, simply updating the backbone
of historical systems results in relative improvements of 25%, allowing them to
outperform the majority of existing systems. A systematic evaluation of papers
in this field was not straightforward. The need to compare like-with-like in
previous papers means that longstanding errors in the evaluation protocol are
ubiquitous in the field. It is likely that many papers were not only optimized
for particular datasets, but also for errors in the data and evaluation
criteria. To aid future research in this area, we release a modular codebase,
allowing for easy evaluation of alternate design decisions against corrected
data and evaluation criteria. We re-implement, validate and re-evaluate 16
state-of-the-art contributions and introduce a new dataset (SYNS-Patches)
containing dense outdoor depth maps in a variety of both natural and urban
scenes. This allows for the computation of informative metrics in complex
regions such as depth boundaries.
Related papers
- On the Evaluation Consistency of Attribution-based Explanations [42.1421504321572]
We introduce Meta-Rank, an open platform for benchmarking attribution methods in the image domain.
Our benchmark reveals three insights in attribution evaluation endeavors: 1) evaluating attribution methods under disparate settings can yield divergent performance rankings; 2) although inconsistent across numerous cases, the performance rankings exhibit remarkable consistency across distinct checkpoints along the same training trajectory; and 3) prior attempts at consistent evaluation fare no better than baselines when extended to more heterogeneous models and datasets.
arXiv Detail & Related papers (2024-07-28T11:49:06Z) - QBI: Quantile-Based Bias Initialization for Efficient Private Data Reconstruction in Federated Learning [0.5497663232622965]
Federated learning enables the training of machine learning models on distributed data without compromising user privacy.
Recent research has shown that the central entity can perfectly reconstruct private data from shared model updates.
arXiv Detail & Related papers (2024-06-26T20:19:32Z) - A Comprehensive Library for Benchmarking Multi-class Visual Anomaly Detection [52.228708947607636]
This paper introduces a comprehensive visual anomaly detection benchmark, ADer, which is a modular framework for new methods.
The benchmark includes multiple datasets from industrial and medical domains, implementing fifteen state-of-the-art methods and nine comprehensive metrics.
We objectively reveal the strengths and weaknesses of different methods and provide insights into the challenges and future directions of multi-class visual anomaly detection.
arXiv Detail & Related papers (2024-06-05T13:40:07Z) - Revisiting Evaluation Metrics for Semantic Segmentation: Optimization
and Evaluation of Fine-grained Intersection over Union [113.20223082664681]
We propose the use of fine-grained mIoUs along with corresponding worst-case metrics.
These fine-grained metrics offer less bias towards large objects, richer statistical information, and valuable insights into model and dataset auditing.
Our benchmark study highlights the necessity of not basing evaluations on a single metric and confirms that fine-grained mIoUs reduce the bias towards large objects.
arXiv Detail & Related papers (2023-10-30T03:45:15Z) - The Second Monocular Depth Estimation Challenge [93.1678025923996]
The second edition of the Monocular Depth Estimation Challenge (MDEC) was open to methods using any form of supervision.
The challenge was based around the SYNS-Patches dataset, which features a wide diversity of environments with high-quality dense ground-truth.
The top supervised submission improved relative F-Score by 27.62%, while the top self-supervised improved it by 16.61%.
arXiv Detail & Related papers (2023-04-14T11:10:07Z) - tieval: An Evaluation Framework for Temporal Information Extraction
Systems [2.3035364984111495]
Temporal information extraction has attracted a great deal of interest over the last two decades.
Having access to a large volume of corpora makes it difficult when it comes to benchmark TIE systems.
tieval is a Python library that provides a concise interface for importing different corpora and facilitates system evaluation.
arXiv Detail & Related papers (2023-01-11T18:55:22Z) - A Closer Look at Debiased Temporal Sentence Grounding in Videos:
Dataset, Metric, and Approach [53.727460222955266]
Temporal Sentence Grounding in Videos (TSGV) aims to ground a natural language sentence in an untrimmed video.
Recent studies have found that current benchmark datasets may have obvious moment annotation biases.
We introduce a new evaluation metric "dR@n,IoU@m" that discounts the basic recall scores to alleviate the inflating evaluation caused by biased datasets.
arXiv Detail & Related papers (2022-03-10T08:58:18Z) - Fairness and underspecification in acoustic scene classification: The
case for disaggregated evaluations [6.186191586944725]
Underspecification and fairness in machine learning (ML) applications have recently become two prominent issues in the ML community.
We argue for the need of a more holistic evaluation process for Acoustic scene classification (ASC) models through disaggregated evaluations.
We demonstrate the effectiveness of the proposed evaluation process in uncovering underspecification and fairness problems when trained on two widely-used ASC datasets.
arXiv Detail & Related papers (2021-10-04T15:23:01Z) - Contextual Fine-to-Coarse Distillation for Coarse-grained Response
Selection in Open-Domain Conversations [48.046725390986595]
We propose a Contextual Fine-to-Coarse (CFC) distilled model for coarse-grained response selection in open-domain conversations.
To evaluate the performance of our proposed model, we construct two new datasets based on the Reddit comments dump and Twitter corpus.
arXiv Detail & Related papers (2021-09-24T08:22:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.