Strategic Hypothesis Testing
- URL: http://arxiv.org/abs/2508.03289v1
- Date: Tue, 05 Aug 2025 10:08:17 GMT
- Title: Strategic Hypothesis Testing
- Authors: Safwan Hossain, Yatong Chen, Yiling Chen,
- Abstract summary: We develop a game-theoretic model that captures how the agent's participation and reporting behavior respond to the principal's statistical decision rule.<n>Despite the complexity of the interaction, we show that the principal's errors exhibit clear monotonic behavior when segmented by an efficiently computable critical p-value threshold.
- Score: 7.960225901913547
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We examine hypothesis testing within a principal-agent framework, where a strategic agent, holding private beliefs about the effectiveness of a product, submits data to a principal who decides on approval. The principal employs a hypothesis testing rule, aiming to pick a p-value threshold that balances false positives and false negatives while anticipating the agent's incentive to maximize expected profitability. Building on prior work, we develop a game-theoretic model that captures how the agent's participation and reporting behavior respond to the principal's statistical decision rule. Despite the complexity of the interaction, we show that the principal's errors exhibit clear monotonic behavior when segmented by an efficiently computable critical p-value threshold, leading to an interpretable characterization of their optimal p-value threshold. We empirically validate our model and these insights using publicly available data on drug approvals. Overall, our work offers a comprehensive perspective on strategic interactions within the hypothesis testing framework, providing technical and regulatory insights.
Related papers
- Generating Causal Explanations of Vehicular Agent Behavioural Interactions with Learnt Reward Profiles [13.450023647228843]
We learn a weighting of reward metrics for agents such that explanations for agent interactions can be causally inferred.<n>We validate our approach quantitatively and qualitatively across three real-world driving datasets.
arXiv Detail & Related papers (2025-03-18T01:53:59Z) - PredictaBoard: Benchmarking LLM Score Predictability [50.47497036981544]
Large Language Models (LLMs) often fail unpredictably.<n>This poses a significant challenge to ensuring their safe deployment.<n>We present PredictaBoard, a novel collaborative benchmarking framework.
arXiv Detail & Related papers (2025-02-20T10:52:38Z) - Sharp Results for Hypothesis Testing with Risk-Sensitive Agents [32.38246810091696]
We study a game-theoretic version of hypothesis testing in which a statistician, also known as a principal, interacts with strategic agents that can generate data.<n>The statistician seeks to design a testing protocol with controlled error, while the data-generating agents, guided by their utility and prior information, choose whether or not to opt in.
arXiv Detail & Related papers (2024-12-21T02:51:56Z) - Incentive-Theoretic Bayesian Inference for Collaborative Science [59.15962177829337]
We study hypothesis testing when there is an agent with a private prior about an unknown parameter.
We show how the principal can conduct statistical inference that leverages the information that is revealed by an agent's strategic behavior.
arXiv Detail & Related papers (2023-07-07T17:59:01Z) - An active learning method for solving competitive multi-agent decision-making and control problems [1.2430809884830318]
We introduce a novel active-learning scheme to identify a stationary action profile for a population of competitive agents.
We show that the proposed learning-based approach can be applied to typical multi-agent control and decision-making problems.
arXiv Detail & Related papers (2022-12-23T19:37:39Z) - Domain Adaptation with Adversarial Training on Penultimate Activations [82.9977759320565]
Enhancing model prediction confidence on unlabeled target data is an important objective in Unsupervised Domain Adaptation (UDA)
We show that this strategy is more efficient and better correlated with the objective of boosting prediction confidence than adversarial training on input images or intermediate features.
arXiv Detail & Related papers (2022-08-26T19:50:46Z) - Alternative Microfoundations for Strategic Classification [33.67797984699066]
We show that rational agents with perfect information produce discontinuities in the aggregate response to a decision rule.
optimal decision rules under standard microfoundations maximize a measure of negative externality known as social burden.
Our model retains analytical tractability, leads to more robust insights about stable points, and imposes a lower social burden at optimality.
arXiv Detail & Related papers (2021-06-24T00:30:58Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - A General Framework for Distributed Inference with Uncertain Models [14.8884251609335]
We study the problem of distributed classification with a network of heterogeneous agents.
We build upon the concept of uncertain models to incorporate the agents' uncertainty in the likelihoods.
arXiv Detail & Related papers (2020-11-20T22:17:12Z) - Catch Me if I Can: Detecting Strategic Behaviour in Peer Assessment [61.24399136715106]
We consider the issue of strategic behaviour in various peer-assessment tasks, including peer grading of exams or homeworks and peer review in hiring or promotions.
Our focus is on designing methods for detection of such manipulations.
Specifically, we consider a setting in which agents evaluate a subset of their peers and output rankings that are later aggregated to form a final ordering.
arXiv Detail & Related papers (2020-10-08T15:08:40Z) - On the model-based stochastic value gradient for continuous
reinforcement learning [50.085645237597056]
We show that simple model-based agents can outperform state-of-the-art model-free agents in terms of both sample-efficiency and final reward.
Our findings suggest that model-based policy evaluation deserves closer attention.
arXiv Detail & Related papers (2020-08-28T17:58:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.