Analysis of hidden feedback loops in continuous machine learning systems
- URL: http://arxiv.org/abs/2101.05673v2
- Date: Sun, 17 Jan 2021 18:38:51 GMT
- Title: Analysis of hidden feedback loops in continuous machine learning systems
- Authors: Anton Khritankov
- Abstract summary: We signify a problem of implicit feedback loops, demonstrate how they intervene with user behavior on an exemplary housing prices prediction system.
Based on a preliminary model, we highlight conditions when such feedback loops arise and discuss possible solution approaches.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this concept paper, we discuss intricacies of specifying and verifying the
quality of continuous and lifelong learning artificial intelligence systems as
they interact with and influence their environment causing a so-called concept
drift. We signify a problem of implicit feedback loops, demonstrate how they
intervene with user behavior on an exemplary housing prices prediction system.
Based on a preliminary model, we highlight conditions when such feedback loops
arise and discuss possible solution approaches.
Related papers
- Joint trajectory and network inference via reference fitting [0.0]
We propose an approach for leveraging both dynamical and perturbational single cell data to jointly learn cellular trajectories and power network inference.
Our approach is motivated by min-entropy estimation for dynamics and can infer directed and signed networks from time-stamped single cell snapshots.
arXiv Detail & Related papers (2024-09-10T21:49:57Z) - Beyond Thumbs Up/Down: Untangling Challenges of Fine-Grained Feedback for Text-to-Image Generation [67.88747330066049]
Fine-grained feedback captures nuanced distinctions in image quality and prompt-alignment.
We show that demonstrating its superiority to coarse-grained feedback is not automatic.
We identify key challenges in eliciting and utilizing fine-grained feedback.
arXiv Detail & Related papers (2024-06-24T17:19:34Z) - A Mathematical Model of the Hidden Feedback Loop Effect in Machine Learning Systems [44.99833362998488]
We introduce a repeated learning process to jointly describe several phenomena attributed to unintended hidden feedback loops.
A distinctive feature of such repeated learning setting is that the state of the environment becomes causally dependent on the learner itself over time.
We present a novel dynamical systems model of the repeated learning process and prove the limiting set of probability distributions for positive and negative feedback loop modes.
arXiv Detail & Related papers (2024-05-04T17:57:24Z) - Continually Improving Extractive QA via Human Feedback [59.49549491725224]
We study continually improving an extractive question answering (QA) system via human user feedback.
We conduct experiments involving thousands of user interactions under diverse setups to broaden the understanding of learning from feedback over time.
arXiv Detail & Related papers (2023-05-21T14:35:32Z) - A Classification of Feedback Loops and Their Relation to Biases in
Automated Decision-Making Systems [0.0]
We study the different types of feedback loops in the ML-based decision-making pipeline.
We find that the existence of feedback loops can perpetuate, reinforce, or even reduce ML biases.
arXiv Detail & Related papers (2023-05-10T11:15:22Z) - From Concept Drift to Model Degradation: An Overview on
Performance-Aware Drift Detectors [1.757501664210825]
Changes in the system on which a predictive machine learning model has been trained may lead to performance degradation during the system's life cycle.
Different terms have been used in the literature to refer to the same type of concept drift and the same term for various types.
This lack of unified terminology is set out to create confusion on distinguishing between different concept drift variants.
arXiv Detail & Related papers (2022-03-21T15:48:13Z) - Simulating Bandit Learning from User Feedback for Extractive Question
Answering [51.97943858898579]
We study learning from user feedback for extractive question answering by simulating feedback using supervised data.
We show that systems initially trained on a small number of examples can dramatically improve given feedback from users on model-predicted answers.
arXiv Detail & Related papers (2022-03-18T17:47:58Z) - Stateful Offline Contextual Policy Evaluation and Learning [88.9134799076718]
We study off-policy evaluation and learning from sequential data.
We formalize the relevant causal structure of problems such as dynamic personalized pricing.
We show improved out-of-sample policy performance in this class of relevant problems.
arXiv Detail & Related papers (2021-10-19T16:15:56Z) - Improving Conversational Question Answering Systems after Deployment
using Feedback-Weighted Learning [69.42679922160684]
We propose feedback-weighted learning based on importance sampling to improve upon an initial supervised system using binary user feedback.
Our work opens the prospect to exploit interactions with real users and improve conversational systems after deployment.
arXiv Detail & Related papers (2020-11-01T19:50:34Z) - Counter-example Guided Learning of Bounds on Environment Behavior [11.357397596759172]
We present a data-driven solution that allows for a system to be evaluated for specification conformance without an accurate model of the environment.
Our approach involves learning a conservative reactive bound of the environment's behavior using data and specification of the system's desired behavior.
arXiv Detail & Related papers (2020-01-20T19:58:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.