Introspective Learning : A Two-Stage Approach for Inference in Neural
Networks
- URL: http://arxiv.org/abs/2209.08425v1
- Date: Sat, 17 Sep 2022 23:31:03 GMT
- Title: Introspective Learning : A Two-Stage Approach for Inference in Neural
Networks
- Authors: Mohit Prabhushankar, Ghassan AlRegib
- Abstract summary: We advocate for two stages in a neural network's decision making process.
The first is the existing feed-forward inference framework where patterns in given data are sensed and associated with previously learned patterns.
The second is a slower reflection stage where we ask the network to reflect on its feed-forward decision by considering and evaluating all available choices.
- Score: 18.32369721322249
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we advocate for two stages in a neural network's decision
making process. The first is the existing feed-forward inference framework
where patterns in given data are sensed and associated with previously learned
patterns. The second stage is a slower reflection stage where we ask the
network to reflect on its feed-forward decision by considering and evaluating
all available choices. Together, we term the two stages as introspective
learning. We use gradients of trained neural networks as a measurement of this
reflection. A simple three-layered Multi Layer Perceptron is used as the second
stage that predicts based on all extracted gradient features. We perceptually
visualize the post-hoc explanations from both stages to provide a visual
grounding to introspection. For the application of recognition, we show that an
introspective network is 4% more robust and 42% less prone to calibration
errors when generalizing to noisy data. We also illustrate the value of
introspective networks in downstream tasks that require generalizability and
calibration including active learning, out-of-distribution detection, and
uncertainty estimation. Finally, we ground the proposed machine introspection
to human introspection for the application of image quality assessment.
Related papers
- A Novel Information-Theoretic Objective to Disentangle Representations
for Fair Classification [46.884905701771004]
One of the main application for such disentangled representations is fair classification.
We adopt an information-theoretic view of this problem which motivates a novel family of regularizers.
The resulting set of losses, called CLINIC, is parameter free and thus, it is easier and faster to train.
arXiv Detail & Related papers (2023-10-21T12:35:48Z) - A Study of Forward-Forward Algorithm for Self-Supervised Learning [65.268245109828]
We study the performance of forward-forward vs. backpropagation for self-supervised representation learning.
Our main finding is that while the forward-forward algorithm performs comparably to backpropagation during (self-supervised) training, the transfer performance is significantly lagging behind in all the studied settings.
arXiv Detail & Related papers (2023-09-21T10:14:53Z) - Gaussian Switch Sampling: A Second Order Approach to Active Learning [11.775252660867285]
In active learning, acquisition functions define informativeness directly on the representation position within the model manifold.
We propose a grounded second-order definition of information content and sample importance within the context of active learning.
We show that our definition produces highly accurate importance scores even when the model representations are constrained by the lack of training data.
arXiv Detail & Related papers (2023-02-16T15:24:56Z) - Neural Maximum A Posteriori Estimation on Unpaired Data for Motion
Deblurring [87.97330195531029]
We propose a Neural Maximum A Posteriori (NeurMAP) estimation framework for training neural networks to recover blind motion information and sharp content from unpaired data.
The proposed NeurMAP is an approach to existing deblurring neural networks, and is the first framework that enables training image deblurring networks on unpaired datasets.
arXiv Detail & Related papers (2022-04-26T08:09:47Z) - DIAL: Deep Interactive and Active Learning for Semantic Segmentation in
Remote Sensing [34.209686918341475]
We propose to build up a collaboration between a deep neural network and a human in the loop.
In a nutshell, the agent iteratively interacts with the network to correct its initially flawed predictions.
We show that active learning based on uncertainty estimation enables to quickly lead the user towards mistakes.
arXiv Detail & Related papers (2022-01-04T09:11:58Z) - Learning Neural Network Subspaces [74.44457651546728]
Recent observations have advanced our understanding of the neural network optimization landscape.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
With a similar computational cost as training one model, we learn lines, curves, and simplexes of high-accuracy neural networks.
arXiv Detail & Related papers (2021-02-20T23:26:58Z) - Training Binary Neural Networks through Learning with Noisy Supervision [76.26677550127656]
This paper formalizes the binarization operations over neural networks from a learning perspective.
Experimental results on benchmark datasets indicate that the proposed binarization technique attains consistent improvements over baselines.
arXiv Detail & Related papers (2020-10-10T01:59:39Z) - Scalable Polyhedral Verification of Recurrent Neural Networks [9.781772283276734]
We present a scalable and precise verifier for recurrent neural networks, called Prover.
Our evaluation shows that Prover successfully verifies several challenging recurrent models in computer vision, speech, and motion sensor classification.
arXiv Detail & Related papers (2020-05-27T11:57:01Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z) - Binary Neural Networks: A Survey [126.67799882857656]
The binary neural network serves as a promising technique for deploying deep models on resource-limited devices.
The binarization inevitably causes severe information loss, and even worse, its discontinuity brings difficulty to the optimization of the deep network.
We present a survey of these algorithms, mainly categorized into the native solutions directly conducting binarization, and the optimized ones using techniques like minimizing the quantization error, improving the network loss function, and reducing the gradient error.
arXiv Detail & Related papers (2020-03-31T16:47:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.