Designing Optimal Behavioral Experiments Using Machine Learning
- URL: http://arxiv.org/abs/2305.07721v2
- Date: Sun, 26 Nov 2023 19:58:15 GMT
- Title: Designing Optimal Behavioral Experiments Using Machine Learning
- Authors: Simon Valentin, Steven Kleinegesse, Neil R. Bramley, Peggy Seri\`es,
Michael U. Gutmann, Christopher G. Lucas
- Abstract summary: We provide a tutorial on leveraging recent advances in BOED and machine learning to find optimal experiments for any kind of model.
We consider theories of how people balance exploration and exploitation in multi-armed bandit decision-making tasks.
As compared to experimental designs commonly used in the literature, we show that our optimal designs more efficiently determine which of a set of models best account for individual human behavior.
- Score: 8.759299724881219
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Computational models are powerful tools for understanding human cognition and
behavior. They let us express our theories clearly and precisely, and offer
predictions that can be subtle and often counter-intuitive. However, this same
richness and ability to surprise means our scientific intuitions and
traditional tools are ill-suited to designing experiments to test and compare
these models. To avoid these pitfalls and realize the full potential of
computational modeling, we require tools to design experiments that provide
clear answers about what models explain human behavior and the auxiliary
assumptions those models must make. Bayesian optimal experimental design (BOED)
formalizes the search for optimal experimental designs by identifying
experiments that are expected to yield informative data. In this work, we
provide a tutorial on leveraging recent advances in BOED and machine learning
to find optimal experiments for any kind of model that we can simulate data
from, and show how by-products of this procedure allow for quick and
straightforward evaluation of models and their parameters against real
experimental data. As a case study, we consider theories of how people balance
exploration and exploitation in multi-armed bandit decision-making tasks. We
validate the presented approach using simulations and a real-world experiment.
As compared to experimental designs commonly used in the literature, we show
that our optimal designs more efficiently determine which of a set of models
best account for individual human behavior, and more efficiently characterize
behavior given a preferred model. At the same time, formalizing a scientific
question such that it can be adequately addressed with BOED can be challenging
and we discuss several potential caveats and pitfalls that practitioners should
be aware of. We provide code and tutorial notebooks to replicate all analyses.
Related papers
- Evaluating Alternative Training Interventions Using Personalized Computational Models of Learning [0.0]
evaluating different training interventions to determine which produce the best learning outcomes is one of the main challenges faced by instructional designers.
We present an approach for automatically tuning models to specific individuals and show that personalized models make better predictions of students' behavior than generic ones.
Our approach makes predictions that align with previous human findings, as well as testable predictions that might be evaluated with future human experiments.
arXiv Detail & Related papers (2024-08-24T22:51:57Z) - Online simulator-based experimental design for cognitive model selection [74.76661199843284]
We propose BOSMOS: an approach to experimental design that can select between computational models without tractable likelihoods.
In simulated experiments, we demonstrate that the proposed BOSMOS technique can accurately select models in up to 2 orders of magnitude less time than existing LFI alternatives.
arXiv Detail & Related papers (2023-03-03T21:41:01Z) - GFlowNets for AI-Driven Scientific Discovery [74.27219800878304]
We present a new probabilistic machine learning framework called GFlowNets.
GFlowNets can be applied in the modeling, hypotheses generation and experimental design stages of the experimental science loop.
We argue that GFlowNets can become a valuable tool for AI-driven scientific discovery.
arXiv Detail & Related papers (2023-02-01T17:29:43Z) - Design Amortization for Bayesian Optimal Experimental Design [70.13948372218849]
We build off of successful variational approaches, which optimize a parameterized variational model with respect to bounds on the expected information gain (EIG)
We present a novel neural architecture that allows experimenters to optimize a single variational model that can estimate the EIG for potentially infinitely many designs.
arXiv Detail & Related papers (2022-10-07T02:12:34Z) - Bayesian Optimal Experimental Design for Simulator Models of Cognition [14.059933880568908]
We combine recent advances in BOED and approximate inference for intractable models to find optimal experimental designs.
Our simulation experiments on multi-armed bandit tasks show that our method results in improved model discrimination and parameter estimation.
arXiv Detail & Related papers (2021-10-29T09:04:01Z) - Cognitive simulation models for inertial confinement fusion: Combining
simulation and experimental data [0.0]
Researchers rely heavily on computer simulations to explore the design space in search of high-performing implosions.
For more effective design and investigation, simulations require input from past experimental data to better predict future performance.
We describe a cognitive simulation method for combining simulation and experimental data into a common, predictive model.
arXiv Detail & Related papers (2021-03-19T02:00:14Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z) - Predicting Performance for Natural Language Processing Tasks [128.34208911925424]
We build regression models to predict the evaluation score of an NLP experiment given the experimental settings as input.
Experimenting on 9 different NLP tasks, we find that our predictors can produce meaningful predictions over unseen languages and different modeling architectures.
arXiv Detail & Related papers (2020-05-02T16:02:18Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.