Faster, Lighter, More Accurate: A Deep Learning Ensemble for Content
Moderation
- URL: http://arxiv.org/abs/2309.05150v1
- Date: Sun, 10 Sep 2023 21:54:03 GMT
- Title: Faster, Lighter, More Accurate: A Deep Learning Ensemble for Content
Moderation
- Authors: Mohammad Hosseini, Mahmudul Hasan
- Abstract summary: We propose an efficient and lightweight deep classification ensemble structure.
Our approach is based on a combination of simple visual features, designed for high-accuracy classification of violent content.
Our evaluation results demonstrate significant improvements in prediction accuracy, while benefiting from 7.64x faster inference and lower computation cost.
- Score: 2.1756081703276
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To address the increasing need for efficient and accurate content moderation,
we propose an efficient and lightweight deep classification ensemble structure.
Our approach is based on a combination of simple visual features, designed for
high-accuracy classification of violent content with low false positives. Our
ensemble architecture utilizes a set of lightweight models with narrowed-down
color features, and we apply it to both images and videos.
We evaluated our approach using a large dataset of explosion and blast
contents and compared its performance to popular deep learning models such as
ResNet-50. Our evaluation results demonstrate significant improvements in
prediction accuracy, while benefiting from 7.64x faster inference and lower
computation cost.
While our approach is tailored to explosion detection, it can be applied to
other similar content moderation and violence detection use cases as well.
Based on our experiments, we propose a "think small, think many" philosophy in
classification scenarios. We argue that transforming a single, large,
monolithic deep model into a verification-based step model ensemble of multiple
small, simple, and lightweight models with narrowed-down visual features can
possibly lead to predictions with higher accuracy.
Related papers
- Few-shot Image Classification based on Gradual Machine Learning [6.935034849731568]
Few-shot image classification aims to accurately classify unlabeled images using only a few labeled samples.
We propose a novel approach based on the non-i.i.d paradigm of gradual machine learning (GML)
We show that the proposed approach can improve the SOTA performance by 1-5% in terms of accuracy.
arXiv Detail & Related papers (2023-07-28T12:30:41Z) - Evaluation of Confidence-based Ensembling in Deep Learning Image
Classification [0.6445605125467573]
Conf-Ensemble is an adaptation to Boosting to create ensembles based on model confidence instead of model errors.
We evaluate the Conf-Ensemble approach in the much more complex task of image classification with the ImageNet dataset.
arXiv Detail & Related papers (2023-03-03T16:29:22Z) - Part-Based Models Improve Adversarial Robustness [57.699029966800644]
We show that combining human prior knowledge with end-to-end learning can improve the robustness of deep neural networks.
Our model combines a part segmentation model with a tiny classifier and is trained end-to-end to simultaneously segment objects into parts.
Our experiments indicate that these models also reduce texture bias and yield better robustness against common corruptions and spurious correlations.
arXiv Detail & Related papers (2022-09-15T15:41:47Z) - Simple Open-Vocabulary Object Detection with Vision Transformers [51.57562920090721]
We propose a strong recipe for transferring image-text models to open-vocabulary object detection.
We use a standard Vision Transformer architecture with minimal modifications, contrastive image-text pre-training, and end-to-end detection fine-tuning.
We provide the adaptation strategies and regularizations needed to attain very strong performance on zero-shot text-conditioned and one-shot image-conditioned object detection.
arXiv Detail & Related papers (2022-05-12T17:20:36Z) - Sparse MoEs meet Efficient Ensembles [49.313497379189315]
We study the interplay of two popular classes of such models: ensembles of neural networks and sparse mixture of experts (sparse MoEs)
We present Efficient Ensemble of Experts (E$3$), a scalable and simple ensemble of sparse MoEs that takes the best of both classes of models, while using up to 45% fewer FLOPs than a deep ensemble.
arXiv Detail & Related papers (2021-10-07T11:58:35Z) - The Case for High-Accuracy Classification: Think Small, Think Many! [4.817521691828748]
We propose an efficient and lightweight deep classification ensemble structure based on a combination of simple color features.
Our evaluation results show considerable improvements on the prediction accuracy compared to the popular ResNet-50 model.
arXiv Detail & Related papers (2021-03-18T16:15:31Z) - Firearm Detection via Convolutional Neural Networks: Comparing a
Semantic Segmentation Model Against End-to-End Solutions [68.8204255655161]
Threat detection of weapons and aggressive behavior from live video can be used for rapid detection and prevention of potentially deadly incidents.
One way for achieving this is through the use of artificial intelligence and, in particular, machine learning for image analysis.
We compare a traditional monolithic end-to-end deep learning model and a previously proposed model based on an ensemble of simpler neural networks detecting fire-weapons via semantic segmentation.
arXiv Detail & Related papers (2020-12-17T15:19:29Z) - A Compact Deep Architecture for Real-time Saliency Prediction [42.58396452892243]
Saliency models aim to imitate the attention mechanism in the human visual system.
Deep models have a high number of parameters which makes them less suitable for real-time applications.
Here we propose a compact yet fast model for real-time saliency prediction.
arXiv Detail & Related papers (2020-08-30T17:47:16Z) - Few-shot Classification via Adaptive Attention [93.06105498633492]
We propose a novel few-shot learning method via optimizing and fast adapting the query sample representation based on very few reference samples.
As demonstrated experimentally, the proposed model achieves state-of-the-art classification results on various benchmark few-shot classification and fine-grained recognition datasets.
arXiv Detail & Related papers (2020-08-06T05:52:59Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.