Do I Have Your Attention: A Large Scale Engagement Prediction Dataset
and Baselines
- URL: http://arxiv.org/abs/2302.00431v2
- Date: Thu, 17 Aug 2023 09:50:29 GMT
- Title: Do I Have Your Attention: A Large Scale Engagement Prediction Dataset
and Baselines
- Authors: Monisha Singh, Ximi Hoque, Donghuo Zeng, Yanan Wang, Kazushi Ikeda,
Abhinav Dhall
- Abstract summary: The degree of concentration, enthusiasm, optimism, and passion displayed by individual(s) while interacting with a machine is referred to as user engagement'
To create engagement prediction systems that can work in real-world conditions, it is quintessential to learn from rich, diverse datasets.
Large scale multi-faceted engagement in the wild dataset EngageNet is proposed.
- Score: 9.896915478880635
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The degree of concentration, enthusiasm, optimism, and passion displayed by
individual(s) while interacting with a machine is referred to as `user
engagement'. Engagement comprises of behavioral, cognitive, and affect related
cues. To create engagement prediction systems that can work in real-world
conditions, it is quintessential to learn from rich, diverse datasets. To this
end, a large scale multi-faceted engagement in the wild dataset EngageNet is
proposed. 31 hours duration data of 127 participants representing different
illumination conditions are recorded. Thorough experiments are performed
exploring the applicability of different features, action units, eye gaze, head
pose, and MARLIN. Data from user interactions (question-answer) are analyzed to
understand the relationship between effective learning and user engagement. To
further validate the rich nature of the dataset, evaluation is also performed
on the EngageWild dataset. The experiments show the usefulness of the proposed
dataset. The code, models, and dataset link are publicly available at
https://github.com/engagenet/engagenet_baselines.
Related papers
- DIPSER: A Dataset for In-Person Student Engagement Recognition in the Wild [1.766742562532995]
In this paper, a novel dataset is introduced, designed to assess student attention within in-person classroom settings.
This dataset encompasses RGB camera data, featuring multiple cameras per student to capture both posture and facial expressions.
A comprehensive suite of attention and emotion labels for each student is provided, generated through self-reporting and evaluations by four different experts.
Our dataset uniquely combines facial and environmental camera data, smartwatch metrics, and includes underrepresented ethnicities in similar datasets, all within in-the-wild, in-person settings.
arXiv Detail & Related papers (2025-02-27T15:50:21Z) - Scaling Laws for the Value of Individual Data Points in Machine Learning [55.596413470429475]
We introduce a new perspective by investigating scaling behavior for the value of individual data points.
We provide learning theory to support our scaling law, and we observe empirically that it holds across diverse model classes.
Our work represents a first step towards understanding and utilizing scaling properties for the value of individual data points.
arXiv Detail & Related papers (2024-05-30T20:10:24Z) - Weakly Supervised Multi-Task Representation Learning for Human Activity
Analysis Using Wearables [2.398608007786179]
We propose a weakly supervised multi-output siamese network that learns to map the data into multiple representation spaces.
The representation of the data samples are positioned in the space such that the data with the same semantic meaning in that aspect are closely located to each other.
arXiv Detail & Related papers (2023-08-06T08:20:07Z) - FaceFusion: Exploiting Full Spectrum of Multiple Datasets [4.438240667468304]
We present a novel training method, named FaceFusion.
It creates a fused view of different datasets that is untainted by identity conflicts, while concurrently training an embedding network using the view.
Using the unified view of combined datasets enables the embedding network to be trained against the entire spectrum of the datasets, leading to a noticeable performance boost.
arXiv Detail & Related papers (2023-05-24T00:51:04Z) - Can Population-based Engagement Improve Personalisation? A Novel Dataset
and Experiments [21.12546768556595]
VLE is a novel dataset that consists of content and video based features extracted from publicly available scientific video lectures.
Our experimental results indicate that the newly proposed VLE dataset leads to building context-agnostic engagement prediction models.
Experiments in combining the built model with a personalising algorithm show promising improvements in addressing the cold-start problem encountered in educational recommenders.
arXiv Detail & Related papers (2022-06-22T15:53:24Z) - Representation Matters: Assessing the Importance of Subgroup Allocations
in Training Data [85.43008636875345]
We show that diverse representation in training data is key to increasing subgroup performances and achieving population level objectives.
Our analysis and experiments describe how dataset compositions influence performance and provide constructive results for using trends in existing data, alongside domain knowledge, to help guide intentional, objective-aware dataset design.
arXiv Detail & Related papers (2021-03-05T00:27:08Z) - Diverse Complexity Measures for Dataset Curation in Self-driving [80.55417232642124]
We propose a new data selection method that exploits a diverse set of criteria that quantize interestingness of traffic scenes.
Our experiments show that the proposed curation pipeline is able to select datasets that lead to better generalization and higher performance.
arXiv Detail & Related papers (2021-01-16T23:45:02Z) - Interactive Weak Supervision: Learning Useful Heuristics for Data
Labeling [19.24454872492008]
Weak supervision offers a promising alternative for producing labeled datasets without ground truth labels.
We develop the first framework for interactive weak supervision in which a method proposes iterations and learns from user feedback.
Our experiments demonstrate that only a small number of feedback are needed to train models that achieve highly competitive test set performance.
arXiv Detail & Related papers (2020-12-11T00:10:38Z) - Towards Understanding Sample Variance in Visually Grounded Language
Generation: Evaluations and Observations [67.4375210552593]
We design experiments to understand an important but often ignored problem in visually grounded language generation.
Given that humans have different utilities and visual attention, how will the sample variance in multi-reference datasets affect the models' performance?
We show that it is of paramount importance to report variance in experiments; that human-generated references could vary drastically in different datasets/tasks, revealing the nature of each task.
arXiv Detail & Related papers (2020-10-07T20:45:14Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - Mining Implicit Entity Preference from User-Item Interaction Data for
Knowledge Graph Completion via Adversarial Learning [82.46332224556257]
We propose a novel adversarial learning approach by leveraging user interaction data for the Knowledge Graph Completion task.
Our generator is isolated from user interaction data, and serves to improve the performance of the discriminator.
To discover implicit entity preference of users, we design an elaborate collaborative learning algorithms based on graph neural networks.
arXiv Detail & Related papers (2020-03-28T05:47:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.