HOD: A Benchmark Dataset for Harmful Object Detection
- URL: http://arxiv.org/abs/2310.05192v1
- Date: Sun, 8 Oct 2023 15:00:38 GMT
- Title: HOD: A Benchmark Dataset for Harmful Object Detection
- Authors: Eungyeom Ha, Heemook Kim, Sung Chul Hong, Dongbin Na
- Abstract summary: We present a new benchmark dataset for harmful object detection.
Our proposed dataset contains more than 10,000 images across 6 categories that might be harmful.
We have conducted extensive experiments to evaluate the effectiveness of our proposed dataset.
- Score: 3.755082744150185
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent multi-media data such as images and videos have been rapidly spread
out on various online services such as social network services (SNS). With the
explosive growth of online media services, the number of image content that may
harm users is also growing exponentially. Thus, most recent online platforms
such as Facebook and Instagram have adopted content filtering systems to
prevent the prevalence of harmful content and reduce the possible risk of
adverse effects on users. Unfortunately, computer vision research on detecting
harmful content has not yet attracted attention enough. Users of each platform
still manually click the report button to recognize patterns of harmful content
they dislike when exposed to harmful content. However, the problem with manual
reporting is that users are already exposed to harmful content. To address
these issues, our research goal in this work is to develop automatic harmful
object detection systems for online services. We present a new benchmark
dataset for harmful object detection. Unlike most related studies focusing on a
small subset of object categories, our dataset addresses various categories.
Specifically, our proposed dataset contains more than 10,000 images across 6
categories that might be harmful, consisting of not only normal cases but also
hard cases that are difficult to detect. Moreover, we have conducted extensive
experiments to evaluate the effectiveness of our proposed dataset. We have
utilized the recently proposed state-of-the-art (SOTA) object detection
architectures and demonstrated our proposed dataset can be greatly useful for
the real-time harmful object detection task. The whole source codes and
datasets are publicly accessible at
https://github.com/poori-nuna/HOD-Benchmark-Dataset.
Related papers
- T2Vs Meet VLMs: A Scalable Multimodal Dataset for Visual Harmfulness Recognition [24.78672820633581]
Existing harmful datasets are curated by the presence of a narrow range of harmful objects.
This hinders the generalizability of methods based on such datasets, potentially leading to misjudgments.
We propose a comprehensive harmful dataset, consisting of 10,000 images and 1,000 videos, crawled from the Internet and generated by 4 generative models.
arXiv Detail & Related papers (2024-09-29T15:20:00Z) - A Comprehensive Library for Benchmarking Multi-class Visual Anomaly Detection [52.228708947607636]
This paper introduces a comprehensive visual anomaly detection benchmark, ADer, which is a modular framework for new methods.
The benchmark includes multiple datasets from industrial and medical domains, implementing fifteen state-of-the-art methods and nine comprehensive metrics.
We objectively reveal the strengths and weaknesses of different methods and provide insights into the challenges and future directions of multi-class visual anomaly detection.
arXiv Detail & Related papers (2024-06-05T13:40:07Z) - Scrapping The Web For Early Wildfire Detection [0.0]
Pyro is a web-scraping-based dataset composed of videos of wildfires from a network of cameras.
Our dataset was filtered based on a strategy to improve the quality and diversity of the data, reducing the final data to a set of 10,000 images.
arXiv Detail & Related papers (2024-02-08T02:01:36Z) - Into the LAIONs Den: Investigating Hate in Multimodal Datasets [67.21783778038645]
This paper investigates the effect of scaling datasets on hateful content through a comparative audit of two datasets: LAION-400M and LAION-2B.
We found that hate content increased by nearly 12% with dataset scale, measured both qualitatively and quantitatively.
We also found that filtering dataset contents based on Not Safe For Work (NSFW) values calculated based on images alone does not exclude all the harmful content in alt-text.
arXiv Detail & Related papers (2023-11-06T19:00:05Z) - Multimodal datasets: misogyny, pornography, and malignant stereotypes [2.8682942808330703]
We examine the recently released LAION-400M dataset, which is a CLIP-filtered dataset of Image-Alt-text pairs parsed from the Common-Crawl dataset.
We found that the dataset contains, troublesome and explicit images and text pairs of rape, pornography, malign stereotypes, racist and ethnic slurs, and other extremely problematic content.
arXiv Detail & Related papers (2021-10-05T11:47:27Z) - Towards Real-World Prohibited Item Detection: A Large-Scale X-ray
Benchmark [53.9819155669618]
This paper presents a large-scale dataset, named as PIDray, which covers various cases in real-world scenarios for prohibited item detection.
With an intensive amount of effort, our dataset contains $12$ categories of prohibited items in $47,677$ X-ray images with high-quality annotated segmentation masks and bounding boxes.
The proposed method performs favorably against the state-of-the-art methods, especially for detecting the deliberately hidden items.
arXiv Detail & Related papers (2021-08-16T11:14:16Z) - Unsupervised Domain Adaption of Object Detectors: A Survey [87.08473838767235]
Recent advances in deep learning have led to the development of accurate and efficient models for various computer vision applications.
Learning highly accurate models relies on the availability of datasets with a large number of annotated images.
Due to this, model performance drops drastically when evaluated on label-scarce datasets having visually distinct images.
arXiv Detail & Related papers (2021-05-27T23:34:06Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Few-Shot Learning for Video Object Detection in a Transfer-Learning
Scheme [70.45901040613015]
We study the new problem of few-shot learning for video object detection.
We employ a transfer-learning framework to effectively train the video object detector on a large number of base-class objects and a few video clips of novel-class objects.
arXiv Detail & Related papers (2021-03-26T20:37:55Z) - Deep Learning Benchmarks and Datasets for Social Media Image
Classification for Disaster Response [5.610924570214424]
We propose new datasets for disaster type detection, informativeness classification, and damage severity assessment.
We benchmark several state-of-the-art deep learning models and achieve promising results.
We release our datasets and models publicly, aiming to provide proper baselines as well as to spur further research in the crisis informatics community.
arXiv Detail & Related papers (2020-11-17T20:15:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.