PeTra: A Sparsely Supervised Memory Model for People Tracking
- URL: http://arxiv.org/abs/2005.02990v1
- Date: Wed, 6 May 2020 17:45:35 GMT
- Title: PeTra: A Sparsely Supervised Memory Model for People Tracking
- Authors: Shubham Toshniwal, Allyson Ettinger, Kevin Gimpel, and Karen Livescu
- Abstract summary: We propose PeTra, a memory-augmented neural network designed to track entities in its memory slots.
We empirically compare key modeling choices, finding that we can simplify several aspects of the design of the memory module while retaining strong performance.
PeTra is highly effective in both evaluations, demonstrating its ability to track people in its memory despite being trained with limited annotation.
- Score: 50.98911178059019
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose PeTra, a memory-augmented neural network designed to track
entities in its memory slots. PeTra is trained using sparse annotation from the
GAP pronoun resolution dataset and outperforms a prior memory model on the task
while using a simpler architecture. We empirically compare key modeling
choices, finding that we can simplify several aspects of the design of the
memory module while retaining strong performance. To measure the people
tracking capability of memory models, we (a) propose a new diagnostic
evaluation based on counting the number of unique entities in text, and (b)
conduct a small scale human evaluation to compare evidence of people tracking
in the memory logs of PeTra relative to a previous approach. PeTra is highly
effective in both evaluations, demonstrating its ability to track people in its
memory despite being trained with limited annotation.
Related papers
- Parameter-Efficient and Memory-Efficient Tuning for Vision Transformer: A Disentangled Approach [87.8330887605381]
We show how to adapt a pre-trained Vision Transformer to downstream recognition tasks with only a few learnable parameters.
We synthesize a task-specific query with a learnable and lightweight module, which is independent of the pre-trained model.
Our method achieves state-of-the-art performance under memory constraints, showcasing its applicability in real-world situations.
arXiv Detail & Related papers (2024-07-09T15:45:04Z) - Causal Estimation of Memorisation Profiles [58.20086589761273]
Understanding memorisation in language models has practical and societal implications.
Memorisation is the causal effect of training with an instance on the model's ability to predict that instance.
This paper proposes a new, principled, and efficient method to estimate memorisation based on the difference-in-differences design from econometrics.
arXiv Detail & Related papers (2024-06-06T17:59:09Z) - Pluggable Neural Machine Translation Models via Memory-augmented Adapters [25.26982333390014]
We propose a memory-augmented adapter to steer pretrained NMT models in a pluggable manner.
Specifically, we construct a multi-granular memory based on the user-provided text samples.
We also propose a training strategy using memory dropout to reduce spurious dependencies between the NMT model and the memory.
arXiv Detail & Related papers (2023-07-12T09:23:41Z) - Eigen Memory Tree [27.33148786536804]
This work introduces the Eigen Memory Tree (EMT), a novel online memory model for sequential learning scenarios.
We demonstrate that EMT outperforms existing online memory approaches, and provide a hybridized EMT-parametric algorithm that enjoys drastically improved performance.
Our findings are validated using 206 datasets from the OpenML repository in both bounded and infinite memory budget situations.
arXiv Detail & Related papers (2022-10-25T14:57:41Z) - A Memory Transformer Network for Incremental Learning [64.0410375349852]
We study class-incremental learning, a training setup in which new classes of data are observed over time for the model to learn from.
Despite the straightforward problem formulation, the naive application of classification models to class-incremental learning results in the "catastrophic forgetting" of previously seen classes.
One of the most successful existing methods has been the use of a memory of exemplars, which overcomes the issue of catastrophic forgetting by saving a subset of past data into a memory bank and utilizing it to prevent forgetting when training future tasks.
arXiv Detail & Related papers (2022-10-10T08:27:28Z) - A Model or 603 Exemplars: Towards Memory-Efficient Class-Incremental
Learning [56.450090618578]
Class-Incremental Learning (CIL) aims to train a model with limited memory size to meet this requirement.
We show that when counting the model size into the total budget and comparing methods with aligned memory size, saving models do not consistently work.
We propose a simple yet effective baseline, denoted as MEMO for Memory-efficient Expandable MOdel.
arXiv Detail & Related papers (2022-05-26T08:24:01Z) - Learning Dynamic Compact Memory Embedding for Deformable Visual Object
Tracking [82.34356879078955]
We propose a compact memory embedding to enhance the discrimination of the segmentation-based deformable visual tracking method.
Our method outperforms the excellent segmentation-based trackers, i.e., D3S and SiamMask on DAVIS 2017 benchmark.
arXiv Detail & Related papers (2021-11-23T03:07:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.