Information-Seeking Decision Strategies Mitigate Risk in Dynamic, Uncertain Environments
- URL: http://arxiv.org/abs/2503.19107v1
- Date: Mon, 24 Mar 2025 19:55:41 GMT
- Title: Information-Seeking Decision Strategies Mitigate Risk in Dynamic, Uncertain Environments
- Authors: Nicholas W. Barendregt, Joshua I. Gold, Krešimir Josić, Zachary P. Kilpatrick,
- Abstract summary: We compare the performance of normative reward- and information-seeking strategies in a foraging task.<n>We find subtle disparities in the actions they take, resulting in meaningful performance differences.<n>Our findings support the adaptive value of information-seeking behaviors that can mitigate risk with minimal reward loss.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To survive in dynamic and uncertain environments, individuals must develop effective decision strategies that balance information gathering and decision commitment. Models of such strategies often prioritize either optimizing tangible payoffs, like reward rate, or gathering information to support a diversity of (possibly unknown) objectives. However, our understanding of the relative merits of these two approaches remains incomplete, in part because direct comparisons have been limited to idealized, static environments that lack the dynamic complexity of the real world. Here we compared the performance of normative reward- and information-seeking strategies in a dynamic foraging task. Both strategies show similar transitions between exploratory and exploitative behaviors as environmental uncertainty changes. However, we find subtle disparities in the actions they take, resulting in meaningful performance differences: whereas reward-seeking strategies generate slightly more reward on average, information-seeking strategies provide more consistent and predictable outcomes. Our findings support the adaptive value of information-seeking behaviors that can mitigate risk with minimal reward loss.
Related papers
- Grounded Test-Time Adaptation for LLM Agents [75.62784644919803]
Large language model (LLM)-based agents struggle to generalize to novel and complex environments.<n>We propose two strategies for adapting LLM agents by leveraging environment-specific information available during deployment.
arXiv Detail & Related papers (2025-11-06T22:24:35Z) - MORSE: Multi-Objective Reinforcement Learning via Strategy Evolution for Supply Chain Optimization [0.0]
In supply chain management, decision-making often involves balancing multiple objectives, such as cost reduction, service level improvement, and environmental sustainability.<n>Traditional multi-objective optimization methods, such as linear programming and evolutionary algorithms, struggle to adapt in real-time to the dynamic nature of supply chains.<n>We propose an approach that combines Reinforcement Learning (RL) and Multi-Objective Evolutionary Algorithms (MOEAs) to address these challenges for dynamic multi-objective optimization under uncertainty.
arXiv Detail & Related papers (2025-09-08T09:51:24Z) - Towards Robust Model Evolution with Algorithmic Recourse [9.608265445696125]
Algorithmic Recourse is a way for users to modify their attributes to align with a model's expectations.<n>In real-world scenarios, users often need to strategically adjust their attributes to compete for limited resources.<n>These shifts arise from user competition, resource constraints, and adaptive user responses.
arXiv Detail & Related papers (2025-03-12T12:17:34Z) - Steganography and Probabilistic Risk Analysis: A Game Theoretical Framework for Quantifying Adversary Advantage and Impact [1.2937020918620652]
This study introduces a novel steganographic game-theoretic model to analyze the strategic interactions between a defending company and an adversary.<n>By framing the scenario as a non-cooperative game, there is systematic evaluation of optimal strategies for both parties.
arXiv Detail & Related papers (2024-12-23T20:02:52Z) - Improving Active Learning with a Bayesian Representation of Epistemic Uncertainty [0.0]
A popular strategy for active learning is to specifically target a reduction in epistemic uncertainty.<n>We show how this combination leads to new active learning strategies that have desirable properties.<n>In order to demonstrate the efficiency of these strategies in non-trivial settings, we introduce the notion of a possibilistic Gaussian process (GP)
arXiv Detail & Related papers (2024-12-11T09:19:20Z) - Non-linear Welfare-Aware Strategic Learning [10.448052192725168]
This paper studies algorithmic decision-making in the presence of strategic individual behaviors.
We first generalize the agent best response model in previous works to the non-linear setting.
We show the three welfare can attain the optimum simultaneously only under restrictive conditions.
arXiv Detail & Related papers (2024-05-03T01:50:03Z) - Reduced-Rank Multi-objective Policy Learning and Optimization [57.978477569678844]
In practice, causal researchers do not have a single outcome in mind a priori.
In government-assisted social benefit programs, policymakers collect many outcomes to understand the multidimensional nature of poverty.
We present a data-driven dimensionality-reduction methodology for multiple outcomes in the context of optimal policy learning.
arXiv Detail & Related papers (2024-04-29T08:16:30Z) - On strategies for risk management and decision making under uncertainty shared across multiple fields [55.2480439325792]
The paper finds more than 110 examples of such strategies and this approach to risk is termed RDOT: Risk-reducing Design and Operations Toolkit.<n>RDOT strategies fall into six broad categories: structural, reactive, formal, adversarial, multi-stage and positive.<n>Overall, RDOT represents an overlooked class of versatile responses to uncertainty.
arXiv Detail & Related papers (2023-09-06T16:14:32Z) - Structured Dynamic Pricing: Optimal Regret in a Global Shrinkage Model [50.06663781566795]
We consider a dynamic model with the consumers' preferences as well as price sensitivity varying over time.
We measure the performance of a dynamic pricing policy via regret, which is the expected revenue loss compared to a clairvoyant that knows the sequence of model parameters in advance.
Our regret analysis results not only demonstrate optimality of the proposed policy but also show that for policy planning it is essential to incorporate available structural information.
arXiv Detail & Related papers (2023-03-28T00:23:23Z) - Meta-control of social learning strategies [9.419484512715242]
Social learning, copying other's behavior without actual experience, offers a cost-effective means of knowledge acquisition.
We show here that while the success-based strategy fully exploits the benign environment of low uncertainly, it fails in uncertain environments.
On the other hand, the conformist strategy can effectively mitigate this adverse effect.
arXiv Detail & Related papers (2021-06-18T09:17:21Z) - A Deep Reinforcement Learning Approach to Marginalized Importance
Sampling with the Successor Representation [61.740187363451746]
Marginalized importance sampling (MIS) measures the density ratio between the state-action occupancy of a target policy and that of a sampling distribution.
We bridge the gap between MIS and deep reinforcement learning by observing that the density ratio can be computed from the successor representation of the target policy.
We evaluate the empirical performance of our approach on a variety of challenging Atari and MuJoCo environments.
arXiv Detail & Related papers (2021-06-12T20:21:38Z) - Reliable Off-policy Evaluation for Reinforcement Learning [53.486680020852724]
In a sequential decision-making problem, off-policy evaluation estimates the expected cumulative reward of a target policy.
We propose a novel framework that provides robust and optimistic cumulative reward estimates using one or multiple logged data.
arXiv Detail & Related papers (2020-11-08T23:16:19Z) - Learning Adaptive Exploration Strategies in Dynamic Environments Through
Informed Policy Regularization [100.72335252255989]
We study the problem of learning exploration-exploitation strategies that effectively adapt to dynamic environments.
We propose a novel algorithm that regularizes the training of an RNN-based policy using informed policies trained to maximize the reward in each task.
arXiv Detail & Related papers (2020-05-06T16:14:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.