Exploring the Effects of Data Augmentation for Drivable Area
Segmentation
- URL: http://arxiv.org/abs/2208.03437v1
- Date: Sat, 6 Aug 2022 03:39:37 GMT
- Title: Exploring the Effects of Data Augmentation for Drivable Area
Segmentation
- Authors: Srinjoy Bhuiya, Ayushman Kumar, Sankalok Sen
- Abstract summary: We focus on investigating the benefits of data augmentation by analyzing pre-existing image datasets.
Our results show that the performance and robustness of existing state of the art (or SOTA) models can be increased dramatically.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The real-time segmentation of drivable areas plays a vital role in
accomplishing autonomous perception in cars. Recently there have been some
rapid strides in the development of image segmentation models using deep
learning. However, most of the advancements have been made in model
architecture design. In solving any supervised deep learning problem related to
segmentation, the success of the model that one builds depends upon the amount
and quality of input training data we use for that model. This data should
contain well-annotated varied images for better working of the segmentation
model. Issues like this pertaining to annotations in a dataset can lead the
model to conclude with overwhelming Type I and II errors in testing and
validation, causing malicious issues when trying to tackle real world problems.
To address this problem and to make our model more accurate, dynamic, and
robust, data augmentation comes into usage as it helps in expanding our sample
training data and making it better and more diversified overall. Hence, in our
study, we focus on investigating the benefits of data augmentation by analyzing
pre-existing image datasets and performing augmentations accordingly. Our
results show that the performance and robustness of existing state of the art
(or SOTA) models can be increased dramatically without any increase in model
complexity or inference time. The augmentations decided on and used in this
paper were decided only after thorough research of several other augmentation
methodologies and strategies and their corresponding effects that are in
widespread usage today. All our results are being reported on the widely used
Cityscapes Dataset.
Related papers
- A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception [78.26734070960886]
Current perceptive models heavily depend on resource-intensive datasets.
We introduce perception-aware loss (P.A. loss) through segmentation, improving both quality and controllability.
Our method customizes data augmentation by extracting and utilizing perception-aware attribute (P.A. Attr) during generation.
arXiv Detail & Related papers (2024-03-20T04:58:03Z) - Data-efficient Large Vision Models through Sequential Autoregression [58.26179273091461]
We develop an efficient, autoregression-based vision model on a limited dataset.
We demonstrate how this model achieves proficiency in a spectrum of visual tasks spanning both high-level and low-level semantic understanding.
Our empirical evaluations underscore the model's agility in adapting to various tasks, heralding a significant reduction in the parameter footprint.
arXiv Detail & Related papers (2024-02-07T13:41:53Z) - A Survey on Data Augmentation in Large Model Era [16.05117556207015]
Large models, encompassing large language and diffusion models, have shown exceptional promise in approximating human-level intelligence.
With continuous updates to these models, the existing reservoir of high-quality data may soon be depleted.
This paper offers an exhaustive review of large model-driven data augmentation methods.
arXiv Detail & Related papers (2024-01-27T14:19:33Z) - Data-Centric Long-Tailed Image Recognition [49.90107582624604]
Long-tail models exhibit a strong demand for high-quality data.
Data-centric approaches aim to enhance both the quantity and quality of data to improve model performance.
There is currently a lack of research into the underlying mechanisms explaining the effectiveness of information augmentation.
arXiv Detail & Related papers (2023-11-03T06:34:37Z) - Quantifying Overfitting: Introducing the Overfitting Index [0.0]
Overfitting is where a model exhibits superior performance on training data but falters on unseen data.
This paper introduces the Overfitting Index (OI), a novel metric devised to quantitatively assess a model's tendency to overfit.
Our results underscore the variable overfitting behaviors across architectures and highlight the mitigative impact of data augmentation.
arXiv Detail & Related papers (2023-08-16T21:32:57Z) - Scaling Laws Do Not Scale [54.72120385955072]
Recent work has argued that as the size of a dataset increases, the performance of a model trained on that dataset will increase.
We argue that this scaling law relationship depends on metrics used to measure performance that may not correspond with how different groups of people perceive the quality of models' output.
Different communities may also have values in tension with each other, leading to difficult, potentially irreconcilable choices about metrics used for model evaluations.
arXiv Detail & Related papers (2023-07-05T15:32:21Z) - Exploring the Impact of Instruction Data Scaling on Large Language
Models: An Empirical Study on Real-World Use Cases [17.431381376675432]
In this paper we explore the performance of large language models based on instruction tuning across different scales of instruction data.
With Bloomz-7B1-mt as the base model, the results show that merely increasing the amount of instruction data leads to continuous improvement in tasks such as open-ended generation.
We propose potential future research directions such as effectively selecting high-quality training data, scaling base models and training methods specialized for hard tasks.
arXiv Detail & Related papers (2023-03-26T14:49:37Z) - Optimizing the Procedure of CT Segmentation Labeling [1.2891210250935146]
In Computed Tomography, machine learning is often used for automated data processing.
We consider the annotation procedure and its effect on the model performance.
We assume three main virtues of a good dataset collected for a model training to be label quality, diversity, and completeness.
arXiv Detail & Related papers (2023-03-24T15:52:42Z) - Dataset Cartography: Mapping and Diagnosing Datasets with Training
Dynamics [118.75207687144817]
We introduce Data Maps, a model-based tool to characterize and diagnose datasets.
We leverage a largely ignored source of information: the behavior of the model on individual instances during training.
Our results indicate that a shift in focus from quantity to quality of data could lead to robust models and improved out-of-distribution generalization.
arXiv Detail & Related papers (2020-09-22T20:19:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.