Cost-effective Interactive Attention Learning with Neural Attention
Processes
- URL: http://arxiv.org/abs/2006.05419v1
- Date: Tue, 9 Jun 2020 17:36:41 GMT
- Title: Cost-effective Interactive Attention Learning with Neural Attention
Processes
- Authors: Jay Heo, Junhyeon Park, Hyewon Jeong, Kwang Joon Kim, Juho Lee, Eunho
Yang, Sung Ju Hwang
- Abstract summary: We propose a novel interactive learning framework which we refer to as Interactive Attention Learning (IAL)
IAL is prone to overfitting due to scarcity of human annotations, and requires costly retraining.
We tackle these challenges by proposing a sample-efficient attention mechanism and a cost-effective reranking algorithm for instances and features.
- Score: 79.8115563067513
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel interactive learning framework which we refer to as
Interactive Attention Learning (IAL), in which the human supervisors
interactively manipulate the allocated attentions, to correct the model's
behavior by updating the attention-generating network. However, such a model is
prone to overfitting due to scarcity of human annotations, and requires costly
retraining. Moreover, it is almost infeasible for the human annotators to
examine attentions on tons of instances and features. We tackle these
challenges by proposing a sample-efficient attention mechanism and a
cost-effective reranking algorithm for instances and features. First, we
propose Neural Attention Process (NAP), which is an attention generator that
can update its behavior by incorporating new attention-level supervisions
without any retraining. Secondly, we propose an algorithm which prioritizes the
instances and the features by their negative impacts, such that the model can
yield large improvements with minimal human feedback. We validate IAL on
various time-series datasets from multiple domains (healthcare, real-estate,
and computer vision) on which it significantly outperforms baselines with
conventional attention mechanisms, or without cost-effective reranking, with
substantially less retraining and human-model interaction cost.
Related papers
- Understanding Human Activity with Uncertainty Measure for Novelty in Graph Convolutional Networks [2.223052975765005]
We introduce the Temporal Fusion Graph Convolutional Network.
It aims to rectify the inadequate boundary estimation of individual actions within an activity stream.
It also mitigates the issue of over-segmentation in the temporal dimension.
arXiv Detail & Related papers (2024-10-10T13:44:18Z) - A Primal-Dual Framework for Transformers and Neural Networks [52.814467832108875]
Self-attention is key to the remarkable success of transformers in sequence modeling tasks.
We show that the self-attention corresponds to the support vector expansion derived from a support vector regression problem.
We propose two new attentions: Batch Normalized Attention (Attention-BN) and Attention with Scaled Head (Attention-SH)
arXiv Detail & Related papers (2024-06-19T19:11:22Z) - Learning from Observer Gaze:Zero-Shot Attention Prediction Oriented by Human-Object Interaction Recognition [13.956664101032006]
We first collect a novel gaze fixation dataset named IG, comprising 530,000 fixation points across 740 diverse interaction categories.
We then introduce the zero-shot interaction-oriented attention prediction task ZeroIA, which challenges models to predict visual cues for interactions not encountered during training.
Thirdly, we present the Interactive Attention model IA, designed to emulate human observers cognitive processes to tackle the ZeroIA problem.
arXiv Detail & Related papers (2024-05-16T09:34:57Z) - Efficient Human-in-the-loop System for Guiding DNNs Attention [25.501443892795614]
We propose an efficient human-in-the-loop system to interactively direct the attention of classifiers to the regions specified by users.
Previous approaches for attention guidance require the preparation of pixel-level annotations and are not designed as interactive systems.
arXiv Detail & Related papers (2022-06-13T09:04:32Z) - Learning Self-Modulating Attention in Continuous Time Space with
Applications to Sequential Recommendation [102.24108167002252]
We propose a novel attention network, named self-modulating attention, that models the complex and non-linearly evolving dynamic user preferences.
We empirically demonstrate the effectiveness of our method on top-N sequential recommendation tasks, and the results on three large-scale real-world datasets show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-03-30T03:54:11Z) - Aligning Eyes between Humans and Deep Neural Network through Interactive
Attention Alignment [17.653966477405024]
We propose a novel framework of Interactive Attention Alignment (IAA) that aims at realizing human-steerable Deep Neural Networks (DNNs)
IAA leverages DNN model explanation method as an interactive medium that humans can use to unveil the cases of biased model attention and directly adjust the attention.
In improving the DNN using human-generated adjusted attention, we introduce GRADIA, a novel computational pipeline that jointly maximizes attention quality and prediction accuracy.
arXiv Detail & Related papers (2022-02-06T19:22:06Z) - Unlocking Pixels for Reinforcement Learning via Implicit Attention [61.666538764049854]
We make use of new efficient attention algorithms, recently shown to be highly effective for Transformers.
This allows our attention-based controllers to scale to larger visual inputs, and facilitate the use of smaller patches.
In addition, we propose a new efficient algorithm approximating softmax attention with what we call hybrid random features.
arXiv Detail & Related papers (2021-02-08T17:00:26Z) - Parrot: Data-Driven Behavioral Priors for Reinforcement Learning [79.32403825036792]
We propose a method for pre-training behavioral priors that can capture complex input-output relationships observed in successful trials.
We show how this learned prior can be used for rapidly learning new tasks without impeding the RL agent's ability to try out novel behaviors.
arXiv Detail & Related papers (2020-11-19T18:47:40Z) - Untangling tradeoffs between recurrence and self-attention in neural
networks [81.30894993852813]
We present a formal analysis of how self-attention affects gradient propagation in recurrent networks.
We prove that it mitigates the problem of vanishing gradients when trying to capture long-term dependencies.
We propose a relevancy screening mechanism that allows for a scalable use of sparse self-attention with recurrence.
arXiv Detail & Related papers (2020-06-16T19:24:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.