A Novel Disaster Image Dataset and Characteristics Analysis using
Attention Model
- URL: http://arxiv.org/abs/2107.01284v1
- Date: Fri, 2 Jul 2021 21:18:20 GMT
- Title: A Novel Disaster Image Dataset and Characteristics Analysis using
Attention Model
- Authors: Fahim Faisal Niloy, Arif, Abu Bakar Siddik Nayem, Anis Sarker, Ovi
Paul, M. Ashraful Amin, Amin Ahsan Ali, Moinul Islam Zaber, AKM Mahbubur
Rahman
- Abstract summary: This dataset contains images collected from various sources for three different disasters: fire, water and land.
There are 13,720 manually annotated images in this dataset, each image is annotated by three individuals.
A three layer attention model (TLAM) is trained and average five fold validation accuracy of 95.88% is achieved.
- Score: 2.1473182295633224
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The advancement of deep learning technology has enabled us to develop systems
that outperform any other classification technique. However, success of any
empirical system depends on the quality and diversity of the data available to
train the proposed system. In this research, we have carefully accumulated a
relatively challenging dataset that contains images collected from various
sources for three different disasters: fire, water and land. Besides this, we
have also collected images for various damaged infrastructure due to natural or
man made calamities and damaged human due to war or accidents. We have also
accumulated image data for a class named non-damage that contains images with
no such disaster or sign of damage in them. There are 13,720 manually annotated
images in this dataset, each image is annotated by three individuals. We are
also providing discriminating image class information annotated manually with
bounding box for a set of 200 test images. Images are collected from different
news portals, social media, and standard datasets made available by other
researchers. A three layer attention model (TLAM) is trained and average five
fold validation accuracy of 95.88% is achieved. Moreover, on the 200 unseen
test images this accuracy is 96.48%. We also generate and compare attention
maps for these test images to determine the characteristics of the trained
attention model. Our dataset is available at
https://niloy193.github.io/Disaster-Dataset
Related papers
- Community Forensics: Using Thousands of Generators to Train Fake Image Detectors [15.166026536032142]
One of the key challenges of detecting AI-generated images is spotting images that have been created by previously unseen generative models.
We propose a new dataset that is significantly larger and more diverse than prior work.
The resulting dataset contains 2.7M images that have been sampled from 4803 different models.
arXiv Detail & Related papers (2024-11-06T18:59:41Z) - Deep Image Composition Meets Image Forgery [0.0]
Image forgery has been studied for many years.
Deep learning models require large amounts of labeled data for training.
We use state of the art image composition deep learning models to generate spliced images close to the quality of real-life manipulations.
arXiv Detail & Related papers (2024-04-03T17:54:37Z) - Invariant Learning via Diffusion Dreamed Distribution Shifts [121.71383835729848]
We propose a dataset called Diffusion Dreamed Distribution Shifts (D3S)
D3S consists of synthetic images generated through StableDiffusion using text prompts and image guides obtained by pasting a sample foreground image onto a background template image.
Due to the incredible photorealism of the diffusion model, our images are much closer to natural images than previous synthetic datasets.
arXiv Detail & Related papers (2022-11-18T17:07:43Z) - From Synthetic to Real: Image Dehazing Collaborating with Unlabeled Real
Data [58.50411487497146]
We propose a novel image dehazing framework collaborating with unlabeled real data.
First, we develop a disentangled image dehazing network (DID-Net), which disentangles the feature representations into three component maps.
Then a disentangled-consistency mean-teacher network (DMT-Net) is employed to collaborate unlabeled real data for boosting single image dehazing.
arXiv Detail & Related papers (2021-08-06T04:00:28Z) - Learning to See by Looking at Noise [87.12788334473295]
We investigate a suite of image generation models that produce images from simple random processes.
These are then used as training data for a visual representation learner with a contrastive loss.
Our findings show that it is important for the noise to capture certain structural properties of real data but that good performance can be achieved even with processes that are far from realistic.
arXiv Detail & Related papers (2021-06-10T17:56:46Z) - Generating Physically-Consistent Satellite Imagery for Climate Visualizations [53.61991820941501]
We train a generative adversarial network to create synthetic satellite imagery of future flooding and reforestation events.
A pure deep learning-based model can generate flood visualizations but hallucinates floods at locations that were not susceptible to flooding.
We publish our code and dataset for segmentation guided image-to-image translation in Earth observation.
arXiv Detail & Related papers (2021-04-10T15:00:15Z) - Unifying Remote Sensing Image Retrieval and Classification with Robust
Fine-tuning [3.6526118822907594]
We aim at unifying remote sensing image retrieval and classification with a new large-scale training and testing dataset, SF300.
We show that our framework systematically achieves a boost of retrieval and classification performance on nine different datasets compared to an ImageNet pretrained baseline.
arXiv Detail & Related papers (2021-02-26T11:01:30Z) - An Empirical Study of the Collapsing Problem in Semi-Supervised 2D Human
Pose Estimation [80.02124918255059]
Semi-supervised learning aims to boost the accuracy of a model by exploring unlabeled images.
We learn two networks to mutually teach each other.
The more reliable predictions on easy images in each network are used to teach the other network to learn about the corresponding hard images.
arXiv Detail & Related papers (2020-11-25T03:29:52Z) - Deep Learning Benchmarks and Datasets for Social Media Image
Classification for Disaster Response [5.610924570214424]
We propose new datasets for disaster type detection, informativeness classification, and damage severity assessment.
We benchmark several state-of-the-art deep learning models and achieve promising results.
We release our datasets and models publicly, aiming to provide proper baselines as well as to spur further research in the crisis informatics community.
arXiv Detail & Related papers (2020-11-17T20:15:49Z) - Background Splitting: Finding Rare Classes in a Sea of Background [55.03789745276442]
We focus on the real-world problem of training accurate deep models for image classification of a small number of rare categories.
In these scenarios, almost all images belong to the background category in the dataset (>95% of the dataset is background)
We demonstrate that both standard fine-tuning approaches and state-of-the-art approaches for training on imbalanced datasets do not produce accurate deep models in the presence of this extreme imbalance.
arXiv Detail & Related papers (2020-08-28T23:05:15Z) - Detecting natural disasters, damage, and incidents in the wild [26.73896031797989]
We present the Incidents dataset, which contains 446,684 images annotated by humans that cover 43 incidents across a variety of scenes.
We employ a baseline classification model that mitigates false-positive errors and we perform image filtering experiments on millions of social media images from Flickr and Twitter.
arXiv Detail & Related papers (2020-08-20T20:09:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.