Leveraging recent advances in Pre-Trained Language Models
forEye-Tracking Prediction
- URL: http://arxiv.org/abs/2110.04475v1
- Date: Sat, 9 Oct 2021 06:46:48 GMT
- Title: Leveraging recent advances in Pre-Trained Language Models
forEye-Tracking Prediction
- Authors: Varun Madhavan, Aditya Girish Pawate, Shraman Pal, Abhranil Chandra
- Abstract summary: Natural Language Pro-cessing uses human-derived behavioral data like eye-tracking data to augment the neural nets to solve arange of tasks spanning syntax and semantics.
In this paper,we use the ZuCo 1.0 and ZuCo 2.0 dataset to explore differ-ent linguistic models to directly predict thesegaze features for each word with respect to itssentence.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cognitively inspired Natural Language Pro-cessing uses human-derived
behavioral datalike eye-tracking data, which reflect the seman-tic
representations of language in the humanbrain to augment the neural nets to
solve arange of tasks spanning syntax and semanticswith the aim of teaching
machines about lan-guage processing mechanisms. In this paper,we use the ZuCo
1.0 and ZuCo 2.0 dataset con-taining the eye-gaze features to explore
differ-ent linguistic models to directly predict thesegaze features for each
word with respect to itssentence. We tried different neural networkmodels with
the words as inputs to predict thetargets. And after lots of experimentation
andfeature engineering finally devised a novel ar-chitecture consisting of
RoBERTa Token Clas-sifier with a dense layer on top for languagemodeling and a
stand-alone model consistingof dense layers followed by a transformer layerfor
the extra features we engineered. Finally,we took the mean of the outputs of
both thesemodels to make the final predictions. We eval-uated the models using
mean absolute error(MAE) and the R2 score for each target.
Related papers
- Unveiling Multilinguality in Transformer Models: Exploring Language
Specificity in Feed-Forward Networks [12.7259425362286]
We investigate how multilingual models might leverage key-value memories.
For autoregressive models trained on two or more languages, do all neurons (across layers) respond equally to all languages?
Our findings reveal that the layers closest to the network's input or output tend to exhibit more language-specific behaviour compared to the layers in the middle.
arXiv Detail & Related papers (2023-10-24T06:45:00Z) - Humans and language models diverge when predicting repeating text [52.03471802608112]
We present a scenario in which the performance of humans and LMs diverges.
Human and GPT-2 LM predictions are strongly aligned in the first presentation of a text span, but their performance quickly diverges when memory begins to play a role.
We hope that this scenario will spur future work in bringing LMs closer to human behavior.
arXiv Detail & Related papers (2023-10-10T08:24:28Z) - Few-shot Subgoal Planning with Language Models [58.11102061150875]
We show that language priors encoded in pre-trained language models allow us to infer fine-grained subgoal sequences.
In contrast to recent methods which make strong assumptions about subgoal supervision, our experiments show that language models can infer detailed subgoal sequences without any fine-tuning.
arXiv Detail & Related papers (2022-05-28T01:03:30Z) - A Generative Language Model for Few-shot Aspect-Based Sentiment Analysis [90.24921443175514]
We focus on aspect-based sentiment analysis, which involves extracting aspect term, category, and predicting their corresponding polarities.
We propose to reformulate the extraction and prediction tasks into the sequence generation task, using a generative language model with unidirectional attention.
Our approach outperforms the previous state-of-the-art (based on BERT) on average performance by a large margins in few-shot and full-shot settings.
arXiv Detail & Related papers (2022-04-11T18:31:53Z) - Zero Shot Crosslingual Eye-Tracking Data Prediction using Multilingual
Transformer Models [0.0]
We describe our submission to the CMCL 2022 shared task on predicting human reading patterns for multi-lingual dataset.
Our model uses text representations from transformers and some hand engineered features with a regression layer on top to predict statistical measures of mean and standard deviation.
We train an end to end model to extract meaningful information from different languages and test our model on two seperate datasets.
arXiv Detail & Related papers (2022-03-30T17:11:48Z) - Dependency-based Mixture Language Models [53.152011258252315]
We introduce the Dependency-based Mixture Language Models.
In detail, we first train neural language models with a novel dependency modeling objective.
We then formulate the next-token probability by mixing the previous dependency modeling probability distributions with self-attention.
arXiv Detail & Related papers (2022-03-19T06:28:30Z) - Learning Contextual Representations for Semantic Parsing with
Generation-Augmented Pre-Training [86.91380874390778]
We present Generation-Augmented Pre-training (GAP), that jointly learns representations of natural language utterances and table schemas by leveraging generation models to generate pre-train data.
Based on experimental results, neural semantics that leverage GAP MODEL obtain new state-of-the-art results on both SPIDER and CRITERIA-TO-generative benchmarks.
arXiv Detail & Related papers (2020-12-18T15:53:50Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.