Rethinking the Image Feature Biases Exhibited by Deep CNN Models
- URL: http://arxiv.org/abs/2111.02058v1
- Date: Wed, 3 Nov 2021 08:04:06 GMT
- Title: Rethinking the Image Feature Biases Exhibited by Deep CNN Models
- Authors: Dawei Dai and Yutang Li and Huanan Bao and Sy Xia and Guoyin Wang and
Xiaoli Ma
- Abstract summary: We train two classification tasks based on human intuition to identify anticipated biases.
We conclude that the combined effect of certain features is typically far more influential than any single feature.
- Score: 14.690952990358095
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, convolutional neural networks (CNNs) have been applied
successfully in many fields. However, such deep neural models are still
regarded as black box in most tasks. One of the fundamental issues underlying
this problem is understanding which features are most influential in image
recognition tasks and how they are processed by CNNs. It is widely accepted
that CNN models combine low-level features to form complex shapes until the
object can be readily classified, however, several recent studies have argued
that texture features are more important than other features. In this paper, we
assume that the importance of certain features varies depending on specific
tasks, i.e., specific tasks exhibit a feature bias. We designed two
classification tasks based on human intuition to train deep neural models to
identify anticipated biases. We devised experiments comprising many tasks to
test these biases for the ResNet and DenseNet models. From the results, we
conclude that (1) the combined effect of certain features is typically far more
influential than any single feature; (2) in different tasks, neural models can
perform different biases, that is, we can design a specific task to make a
neural model biased toward a specific anticipated feature.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Biased Attention: Do Vision Transformers Amplify Gender Bias More than
Convolutional Neural Networks? [2.8391805742728553]
Deep neural networks used in computer vision have been shown to exhibit many social biases such as gender bias.
Vision Transformers (ViTs) have become increasingly popular in computer vision applications, outperforming Convolutional Neural Networks (CNNs) in many tasks such as image classification.
This research found that ViTs amplified gender bias to a greater extent than CNNs.
arXiv Detail & Related papers (2023-09-15T20:59:12Z) - Provable Multi-Task Representation Learning by Two-Layer ReLU Neural Networks [69.38572074372392]
We present the first results proving that feature learning occurs during training with a nonlinear model on multiple tasks.
Our key insight is that multi-task pretraining induces a pseudo-contrastive loss that favors representations that align points that typically have the same label across tasks.
arXiv Detail & Related papers (2023-07-13T16:39:08Z) - Bias Loss for Mobile Neural Networks [14.08237958134373]
Compact convolutional neural networks (CNNs) have witnessed exceptional improvements in performance in recent years.
The diverse and even abundant features captured by the layers is an important characteristic of these successful CNNs.
In compact CNNs, due to the limited number of parameters, abundant features are unlikely to be obtained, and feature diversity becomes an essential characteristic.
arXiv Detail & Related papers (2021-07-23T12:37:56Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Teaching CNNs to mimic Human Visual Cognitive Process & regularise
Texture-Shape bias [18.003188982585737]
Recent experiments in computer vision demonstrate texture bias as the primary reason for supreme results in models employing Convolutional Neural Networks (CNNs)
It is believed that the cost function forces the CNN to take a greedy approach and develop a proclivity for local information like texture to increase accuracy, thus failing to explore any global statistics.
We propose CognitiveCNN, a new intuitive architecture, inspired from feature integration theory in psychology to utilise human interpretable feature like shape, texture, edges etc. to reconstruct, and classify the image.
arXiv Detail & Related papers (2020-06-25T22:32:54Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z) - When CNNs Meet Random RNNs: Towards Multi-Level Analysis for RGB-D
Object and Scene Recognition [10.796613905980609]
We propose a novel framework that extracts discriminative feature representations from multi-modal RGB-D images for object and scene recognition tasks.
To cope with the high dimensionality of CNN activations, a random weighted pooling scheme has been proposed.
Experiments verify that fully randomized structure in RNN stage encodes CNN activations to discriminative solid features successfully.
arXiv Detail & Related papers (2020-04-26T10:58:27Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z) - NeurAll: Towards a Unified Visual Perception Model for Automated Driving [8.49826472556323]
We propose a joint multi-task network design for learning several tasks simultaneously.
The main bottleneck in automated driving systems is the limited processing power available on deployment hardware.
arXiv Detail & Related papers (2019-02-10T12:45:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.