Synthesizing Human Gaze Feedback for Improved NLP Performance
- URL: http://arxiv.org/abs/2302.05721v1
- Date: Sat, 11 Feb 2023 15:34:23 GMT
- Title: Synthesizing Human Gaze Feedback for Improved NLP Performance
- Authors: Varun Khurana, Yaman Kumar Singla, Nora Hollenstein, Rajesh Kumar,
Balaji Krishnamurthy
- Abstract summary: ScanTextGAN is a novel model for generating human scanpaths over text.
We show that ScanTextGAN-generated scanpaths can approximate meaningful cognitive signals in human gaze patterns.
- Score: 20.837790838762036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Integrating human feedback in models can improve the performance of natural
language processing (NLP) models. Feedback can be either explicit (e.g. ranking
used in training language models) or implicit (e.g. using human cognitive
signals in the form of eyetracking). Prior eye tracking and NLP research reveal
that cognitive processes, such as human scanpaths, gleaned from human gaze
patterns aid in the understanding and performance of NLP models. However, the
collection of real eyetracking data for NLP tasks is challenging due to the
requirement of expensive and precise equipment coupled with privacy invasion
issues. To address this challenge, we propose ScanTextGAN, a novel model for
generating human scanpaths over text. We show that ScanTextGAN-generated
scanpaths can approximate meaningful cognitive signals in human gaze patterns.
We include synthetically generated scanpaths in four popular NLP tasks spanning
six different datasets as proof of concept and show that the models augmented
with generated scanpaths improve the performance of all downstream NLP tasks.
Related papers
- Caption-Driven Explorations: Aligning Image and Text Embeddings through Human-Inspired Foveated Vision [3.3295510777293837]
We introduce CapMIT1003, a dataset with captions and click-contingent image explorations, to study human attention during the captioning task.
We also present NevaClip, a zero-shot method for predicting visual scanpaths by combining CLIP models with NeVA algorithms.
arXiv Detail & Related papers (2024-08-19T12:41:46Z) - Neural Sign Actors: A diffusion model for 3D sign language production from text [51.81647203840081]
Sign Languages (SL) serve as the primary mode of communication for the Deaf and Hard of Hearing communities.
This work makes an important step towards realistic neural sign avatars, bridging the communication gap between Deaf and hearing communities.
arXiv Detail & Related papers (2023-12-05T12:04:34Z) - Generative Input: Towards Next-Generation Input Methods Paradigm [49.98958865125018]
We propose a novel Generative Input paradigm named GeneInput.
It uses prompts to handle all input scenarios and other intelligent auxiliary input functions, optimizing the model with user feedback to deliver personalized results.
The results demonstrate that we have achieved state-of-the-art performance for the first time in the Full-mode Key-sequence to Characters(FK2C) task.
arXiv Detail & Related papers (2023-11-02T12:01:29Z) - ScanDL: A Diffusion Model for Generating Synthetic Scanpaths on Texts [0.5520145204626482]
Eye movements in reading play a crucial role in psycholinguistic research.
The scarcity of eye movement data and its unavailability at application time poses a major challenge for this line of research.
We propose ScanDL, a novel discrete sequence-to-sequence diffusion model that generates synthetic scanpaths on texts.
arXiv Detail & Related papers (2023-10-24T07:52:19Z) - Pre-Trained Language Models Augmented with Synthetic Scanpaths for
Natural Language Understanding [3.6498648388765513]
We develop a model that integrates synthetic scanpath generation with a scanpath-augmented language model.
We find that the proposed model not only outperforms the underlying language model, but achieves a performance that is comparable to a language model augmented with real human gaze data.
arXiv Detail & Related papers (2023-10-23T08:15:38Z) - Contrastive Language-Image Pretrained Models are Zero-Shot Human
Scanpath Predictors [2.524526956420465]
CapMIT1003 is a database of captions and click-contingent image explorations collected during captioning tasks.
NevaClip is a novel zero-shot method for predicting visual scanpaths.
arXiv Detail & Related papers (2023-05-21T07:24:50Z) - Artificial Text Detection via Examining the Topology of Attention Maps [58.46367297712477]
We propose three novel types of interpretable topological features for this task based on Topological Data Analysis (TDA)
We empirically show that the features derived from the BERT model outperform count- and neural-based baselines up to 10% on three common datasets.
The probing analysis of the features reveals their sensitivity to the surface and syntactic properties.
arXiv Detail & Related papers (2021-09-10T12:13:45Z) - Graph Neural Networks for Natural Language Processing: A Survey [64.36633422999905]
We present a comprehensive overview onGraph Neural Networks (GNNs) for Natural Language Processing.
We propose a new taxonomy of GNNs for NLP, which organizes existing research of GNNs for NLP along three axes: graph construction,graph representation learning, and graph based encoder-decoder models.
arXiv Detail & Related papers (2021-06-10T23:59:26Z) - CogAlign: Learning to Align Textual Neural Representations to Cognitive
Language Processing Signals [60.921888445317705]
We propose a CogAlign approach to integrate cognitive language processing signals into natural language processing models.
We show that CogAlign achieves significant improvements with multiple cognitive features over state-of-the-art models on public datasets.
arXiv Detail & Related papers (2021-06-10T07:10:25Z) - Word Shape Matters: Robust Machine Translation with Visual Embedding [78.96234298075389]
We introduce a new encoding of the input symbols for character-level NLP models.
It encodes the shape of each character through the images depicting the letters when printed.
We name this new strategy visual embedding and it is expected to improve the robustness of NLP models.
arXiv Detail & Related papers (2020-10-20T04:08:03Z) - Improving Natural Language Processing Tasks with Human Gaze-Guided
Neural Attention [14.940723222424749]
A lack of corpora has so far limited advances in integrating human gaze data as a supervisory signal in neural attention mechanisms.
We propose a novel hybrid text saliency model that combines a cognitive model of reading with explicit human gaze supervision.
Our work introduces a practical approach for bridging between data-driven and cognitive models and demonstrates a new way to integrate human gaze-guided neural attention into NLP tasks.
arXiv Detail & Related papers (2020-10-15T17:14:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.