Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans
by measuring error consistency
- URL: http://arxiv.org/abs/2006.16736v3
- Date: Fri, 18 Dec 2020 15:39:48 GMT
- Title: Beyond accuracy: quantifying trial-by-trial behaviour of CNNs and humans
by measuring error consistency
- Authors: Robert Geirhos, Kristof Meding, Felix A. Wichmann
- Abstract summary: A central problem in cognitive science and behavioural neuroscience is to ascertain whether two or more decision makers (be they brains or algorithms) use the same strategy.
We introduce trial-by-trial error consistency, a quantitative analysis for measuring whether two decision making systems systematically make errors on the same inputs.
- Score: 10.028543085687803
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A central problem in cognitive science and behavioural neuroscience as well
as in machine learning and artificial intelligence research is to ascertain
whether two or more decision makers (be they brains or algorithms) use the same
strategy. Accuracy alone cannot distinguish between strategies: two systems may
achieve similar accuracy with very different strategies. The need to
differentiate beyond accuracy is particularly pressing if two systems are near
ceiling performance, like Convolutional Neural Networks (CNNs) and humans on
visual object recognition. Here we introduce trial-by-trial error consistency,
a quantitative analysis for measuring whether two decision making systems
systematically make errors on the same inputs. Making consistent errors on a
trial-by-trial basis is a necessary condition for similar processing strategies
between decision makers. Our analysis is applicable to compare algorithms with
algorithms, humans with humans, and algorithms with humans. When applying error
consistency to object recognition we obtain three main findings: (1.)
Irrespective of architecture, CNNs are remarkably consistent with one another.
(2.) The consistency between CNNs and human observers, however, is little above
what can be expected by chance alone -- indicating that humans and CNNs are
likely implementing very different strategies. (3.) CORnet-S, a recurrent model
termed the "current best model of the primate ventral visual stream", fails to
capture essential characteristics of human behavioural data and behaves
essentially like a standard purely feedforward ResNet-50 in our analysis. Taken
together, error consistency analysis suggests that the strategies used by human
and machine vision are still very different -- but we envision our
general-purpose error consistency analysis to serve as a fruitful tool for
quantifying future progress.
Related papers
- Cognitive Networks and Performance Drive fMRI-Based State Classification Using DNN Models [0.0]
We employ two structurally different and complementary DNN-based models to classify individual cognitive states.
We show that despite the architectural differences, both models consistently produce a robust relationship between prediction accuracy and individual cognitive performance.
arXiv Detail & Related papers (2024-08-14T15:25:51Z) - Towards Neural Network based Cognitive Models of Dynamic Decision-Making by Humans [16.72938921687168]
We build on Instance Based Learning (IBL) that posits that human decisions are based on similar situations encountered in the past.
We propose two new attention-based neural network models to have open form non-linear functions to model distinct and heterogeneous human decision-making.
arXiv Detail & Related papers (2024-07-24T20:28:03Z) - Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes [72.13373216644021]
We study the societal impact of machine learning by considering the collection of models that are deployed in a given context.
We find deployed machine learning is prone to systemic failure, meaning some users are exclusively misclassified by all models available.
These examples demonstrate ecosystem-level analysis has unique strengths for characterizing the societal impact of machine learning.
arXiv Detail & Related papers (2023-07-12T01:11:52Z) - Neural Causal Models for Counterfactual Identification and Estimation [62.30444687707919]
We study the evaluation of counterfactual statements through neural models.
First, we show that neural causal models (NCMs) are expressive enough.
Second, we develop an algorithm for simultaneously identifying and estimating counterfactual distributions.
arXiv Detail & Related papers (2022-09-30T18:29:09Z) - Neural Architecture Dilation for Adversarial Robustness [56.18555072877193]
A shortcoming of convolutional neural networks is that they are vulnerable to adversarial attacks.
This paper aims to improve the adversarial robustness of the backbone CNNs that have a satisfactory accuracy.
Under a minimal computational overhead, a dilation architecture is expected to be friendly with the standard performance of the backbone CNN.
arXiv Detail & Related papers (2021-08-16T03:58:00Z) - The Causal Neural Connection: Expressiveness, Learnability, and
Inference [125.57815987218756]
An object called structural causal model (SCM) represents a collection of mechanisms and sources of random variation of the system under investigation.
In this paper, we show that the causal hierarchy theorem (Thm. 1, Bareinboim et al., 2020) still holds for neural models.
We introduce a special type of SCM called a neural causal model (NCM), and formalize a new type of inductive bias to encode structural constraints necessary for performing causal inferences.
arXiv Detail & Related papers (2021-07-02T01:55:18Z) - Are Convolutional Neural Networks or Transformers more like human
vision? [9.83454308668432]
We show that attention-based networks can achieve higher accuracy than CNNs on vision tasks.
These results have implications both for building more human-like vision models, as well as for understanding visual object recognition in humans.
arXiv Detail & Related papers (2021-05-15T10:33:35Z) - Vulnerability Under Adversarial Machine Learning: Bias or Variance? [77.30759061082085]
We investigate the effect of adversarial machine learning on the bias and variance of a trained deep neural network.
Our analysis sheds light on why the deep neural networks have poor performance under adversarial perturbation.
We introduce a new adversarial machine learning algorithm with lower computational complexity than well-known adversarial machine learning strategies.
arXiv Detail & Related papers (2020-08-01T00:58:54Z) - An Uncertainty-based Human-in-the-loop System for Industrial Tool Wear
Analysis [68.8204255655161]
We show that uncertainty measures based on Monte-Carlo dropout in the context of a human-in-the-loop system increase the system's transparency and performance.
A simulation study demonstrates that the uncertainty-based human-in-the-loop system increases performance for different levels of human involvement.
arXiv Detail & Related papers (2020-07-14T15:47:37Z) - VINNAS: Variational Inference-based Neural Network Architecture Search [2.685668802278155]
We present a differentiable variational inference-based NAS method for searching sparse convolutional neural networks.
Our method finds diverse network cells, while showing state-of-the-art accuracy with up to almost 2 times fewer non-zero parameters.
arXiv Detail & Related papers (2020-07-12T21:47:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.