Consensus in Motion: A Case of Dynamic Rationality of Sequential Learning in Probability Aggregation
- URL: http://arxiv.org/abs/2504.14624v1
- Date: Sun, 20 Apr 2025 14:04:39 GMT
- Title: Consensus in Motion: A Case of Dynamic Rationality of Sequential Learning in Probability Aggregation
- Authors: Polina Gordienko, Christoph Jansen, Thomas Augustin, Martin Rechenauer,
- Abstract summary: We propose a framework for probability aggregation based on propositional probability logic.<n>We show that any consensus-compatible and independent aggregation rule on a non-nested agenda is necessarily linear.
- Score: 0.562479170374811
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a framework for probability aggregation based on propositional probability logic. Unlike conventional judgment aggregation, which focuses on static rationality, our model addresses dynamic rationality by ensuring that collective beliefs update consistently with new information. We show that any consensus-compatible and independent aggregation rule on a non-nested agenda is necessarily linear. Furthermore, we provide sufficient conditions for a fair learning process, where individuals initially agree on a specified subset of propositions known as the common ground, and new information is restricted to this shared foundation. This guarantees that updating individual judgments via Bayesian conditioning-whether performed before or after aggregation-yields the same collective belief. A distinctive feature of our framework is its treatment of sequential decision-making, which allows new information to be incorporated progressively through multiple stages while maintaining the established common ground. We illustrate our findings with a running example in a political scenario concerning healthcare and immigration policies.
Related papers
- Confidence Estimation via Sequential Likelihood Mixing [46.69347918899963]
We present a universal framework for constructing confidence sets based on sequential likelihood mixing.<n>We establish fundamental connections between sequential mixing, Bayesian inference and regret inequalities from online estimation.<n>We illustrate the power of the framework by deriving tighter confidence sequences for classical settings.
arXiv Detail & Related papers (2025-02-20T16:16:34Z) - Policy Aggregation [21.21314301021803]
We consider the challenge of AI value alignment with multiple individuals with different reward functions and optimal policies in an underlying Markov decision process.
We formalize this problem as one of policy aggregation, where the goal is to identify a desirable collective policy.
Key insight is that social choice methods can be reinterpreted by identifying ordinal preferences with volumes of subsets of the state-action occupancy polytope.
arXiv Detail & Related papers (2024-11-06T04:19:50Z) - Sequential Representation Learning via Static-Dynamic Conditional Disentanglement [58.19137637859017]
This paper explores self-supervised disentangled representation learning within sequential data, focusing on separating time-independent and time-varying factors in videos.
We propose a new model that breaks the usual independence assumption between those factors by explicitly accounting for the causal relationship between the static/dynamic variables.
Experiments show that the proposed approach outperforms previous complex state-of-the-art techniques in scenarios where the dynamics of a scene are influenced by its content.
arXiv Detail & Related papers (2024-08-10T17:04:39Z) - Sequential Manipulation Against Rank Aggregation: Theory and Algorithm [119.57122943187086]
We leverage an online attack on the vulnerable data collection process.
From the game-theoretic perspective, the confrontation scenario is formulated as a distributionally robust game.
The proposed method manipulates the results of rank aggregation methods in a sequential manner.
arXiv Detail & Related papers (2024-07-02T03:31:21Z) - Designing Long-term Group Fair Policies in Dynamical Systems [12.115106776644156]
We propose a novel framework for achieving long-term group fairness in dynamical systems.
Our framework allows us to identify a time-independent policy that converges, if deployed, to the targeted fair stationary state of the system in the long term.
arXiv Detail & Related papers (2023-11-21T08:58:50Z) - Generative Intrinsic Optimization: Intrinsic Control with Model Learning [5.439020425819001]
Future sequence represents the outcome after executing the action into the environment.
Explicit outcomes may vary across state, return, or trajectory serving different purposes such as credit assignment or imitation learning.
We propose a policy scheme that seamlessly incorporates the mutual information, ensuring convergence to the optimal policy.
arXiv Detail & Related papers (2023-10-12T07:50:37Z) - Auditing Fairness by Betting [47.53732591434]
We provide practical, efficient, and nonparametric methods for auditing the fairness of deployed classification and regression models.
Our methods are sequential and allow for the continuous monitoring of incoming data.
We demonstrate the efficacy of our approach on three benchmark fairness datasets.
arXiv Detail & Related papers (2023-05-27T20:14:11Z) - Is Vertical Logistic Regression Privacy-Preserving? A Comprehensive
Privacy Analysis and Beyond [57.10914865054868]
We consider vertical logistic regression (VLR) trained with mini-batch descent gradient.
We provide a comprehensive and rigorous privacy analysis of VLR in a class of open-source Federated Learning frameworks.
arXiv Detail & Related papers (2022-07-19T05:47:30Z) - Constructing a Good Behavior Basis for Transfer using Generalized Policy
Updates [63.58053355357644]
We study the problem of learning a good set of policies, so that when combined together, they can solve a wide variety of unseen reinforcement learning tasks.
We show theoretically that having access to a specific set of diverse policies, which we call a set of independent policies, can allow for instantaneously achieving high-level performance.
arXiv Detail & Related papers (2021-12-30T12:20:46Z) - Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement
Learning [70.01650994156797]
Off- evaluation of sequential decision policies from observational data is necessary in batch reinforcement learning such as education healthcare.
We develop an approach that estimates the bounds of a given policy.
We prove convergence to the sharp bounds as we collect more confounded data.
arXiv Detail & Related papers (2020-02-11T16:18:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.