X2T: Training an X-to-Text Typing Interface with Online Learning from
User Feedback
- URL: http://arxiv.org/abs/2203.02072v2
- Date: Mon, 7 Mar 2022 01:39:28 GMT
- Title: X2T: Training an X-to-Text Typing Interface with Online Learning from
User Feedback
- Authors: Jensen Gao, Siddharth Reddy, Glen Berseth, Nicholas Hardy, Nikhilesh
Natraj, Karunesh Ganguly, Anca D. Dragan, Sergey Levine
- Abstract summary: We focus on assistive typing applications in which a user cannot operate a keyboard, but can supply other inputs.
Standard methods train a model on a fixed dataset of user inputs, then deploy a static interface that does not learn from its mistakes.
We investigate a simple idea that would enable such interfaces to improve over time, with minimal additional effort from the user.
- Score: 83.95599156217945
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We aim to help users communicate their intent to machines using flexible,
adaptive interfaces that translate arbitrary user input into desired actions.
In this work, we focus on assistive typing applications in which a user cannot
operate a keyboard, but can instead supply other inputs, such as webcam images
that capture eye gaze or neural activity measured by a brain implant. Standard
methods train a model on a fixed dataset of user inputs, then deploy a static
interface that does not learn from its mistakes; in part, because extracting an
error signal from user behavior can be challenging. We investigate a simple
idea that would enable such interfaces to improve over time, with minimal
additional effort from the user: online learning from user feedback on the
accuracy of the interface's actions. In the typing domain, we leverage
backspaces as feedback that the interface did not perform the desired action.
We propose an algorithm called x-to-text (X2T) that trains a predictive model
of this feedback signal, and uses this model to fine-tune any existing, default
interface for translating user input into actions that select words or
characters. We evaluate X2T through a small-scale online user study with 12
participants who type sentences by gazing at their desired words, a large-scale
observational study on handwriting samples from 60 users, and a pilot study
with one participant using an electrocorticography-based brain-computer
interface. The results show that X2T learns to outperform a non-adaptive
default interface, stimulates user co-adaptation to the interface, personalizes
the interface to individual users, and can leverage offline data collected from
the default interface to improve its initial performance and accelerate online
learning.
Related papers
- Bootstrapping Adaptive Human-Machine Interfaces with Offline
Reinforcement Learning [82.91837418721182]
Adaptive interfaces can help users perform sequential decision-making tasks.
Recent advances in human-in-the-loop machine learning enable such systems to improve by interacting with users.
We propose a reinforcement learning algorithm to train an interface to map raw command signals to actions.
arXiv Detail & Related papers (2023-09-07T16:52:27Z) - Interactive Text Generation [75.23894005664533]
We introduce a new Interactive Text Generation task that allows training generation models interactively without the costs of involving real users.
We train our interactive models using Imitation Learning, and our experiments against competitive non-interactive generation models show that models trained interactively are superior to their non-interactive counterparts.
arXiv Detail & Related papers (2023-03-02T01:57:17Z) - First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual
Information Maximization [112.40598205054994]
We formalize this idea as a completely unsupervised objective for optimizing interfaces.
We conduct an observational study on 540K examples of users operating various keyboard and eye gaze interfaces for typing, controlling simulated robots, and playing video games.
The results show that our mutual information scores are predictive of the ground-truth task completion metrics in a variety of domains.
arXiv Detail & Related papers (2022-05-24T21:57:18Z) - ASHA: Assistive Teleoperation via Human-in-the-Loop Reinforcement
Learning [91.58711082348293]
Reinforcement learning from online user feedback on the system's performance presents a natural solution to this problem.
This approach tends to require a large amount of human-in-the-loop training data, especially when feedback is sparse.
We propose a hierarchical solution that learns efficiently from sparse user feedback.
arXiv Detail & Related papers (2022-02-05T02:01:19Z) - One Chatbot Per Person: Creating Personalized Chatbots based on Implicit
User Profiles [31.432585994256375]
Existing personalized approaches tried to incorporate several text descriptions as explicit user profiles.
We train a personalized language model to construct a general user profile from the user's historical responses.
We design a personalized decoder to fuse two decoding strategies, including generating a word from the generic vocabulary and copying one word from the user's personalized vocabulary.
arXiv Detail & Related papers (2021-08-20T20:33:12Z) - Federated Learning of User Authentication Models [69.93965074814292]
We propose Federated User Authentication (FedUA), a framework for privacy-preserving training of machine learning models.
FedUA adopts federated learning framework to enable a group of users to jointly train a model without sharing the raw inputs.
We show our method is privacy-preserving, scalable with number of users, and allows new users to be added to training without changing the output layer.
arXiv Detail & Related papers (2020-07-09T08:04:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.