Causal Bandits without Graph Learning
- URL: http://arxiv.org/abs/2301.11401v2
- Date: Thu, 8 Jun 2023 17:11:38 GMT
- Title: Causal Bandits without Graph Learning
- Authors: Mikhail Konobeev, Jalal Etesami, Negar Kiyavash
- Abstract summary: We develop an efficient algorithm for finding the parent node of the reward node using atomic interventions.
We extend our algorithm to the case when the reward node has multiple parents.
- Score: 28.021500949026766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the causal bandit problem when the causal graph is unknown and
develop an efficient algorithm for finding the parent node of the reward node
using atomic interventions. We derive the exact equation for the expected
number of interventions performed by the algorithm and show that under certain
graphical conditions it could perform either logarithmically fast or, under
more general assumptions, slower but still sublinearly in the number of
variables. We formally show that our algorithm is optimal as it meets the
universal lower bound we establish for any algorithm that performs atomic
interventions. Finally, we extend our algorithm to the case when the reward
node has multiple parents. Using this algorithm together with a standard
algorithm from bandit literature leads to improved regret bounds.
Related papers
- Nearest Neighbour with Bandit Feedback [4.9094025705644695]
Our algorithm handles the fully adversarial setting in which no assumptions at all are made about the data-generation process.
We give generic regret for our algorithm and further analyse them when applied to the bandit problem in euclidean space.
arXiv Detail & Related papers (2023-06-23T20:09:01Z) - Optimal Gradient-based Algorithms for Non-concave Bandit Optimization [76.57464214864756]
This work considers a large family of bandit problems where the unknown underlying reward function is non-concave.
Our algorithms are based on a unified zeroth-order optimization paradigm that applies in great generality.
We show that the standard optimistic algorithms are sub-optimal by dimension factors.
arXiv Detail & Related papers (2021-07-09T16:04:24Z) - Causal Bandits with Unknown Graph Structure [24.58691421788476]
We develop novel causal bandit algorithms without knowing the causal graph.
Our algorithms work well for causal trees, causal forests and a general class of causal graphs.
arXiv Detail & Related papers (2021-06-05T23:41:38Z) - Efficient Permutation Discovery in Causal DAGs [9.22466799504763]
We introduce an efficient algorithm for finding sparse permutations of a directed acyclic graph.
We show that our method with depth $w$ runs in $O(pw+3)$ time.
We also compare our algorithm to provably consistent causal structure learning algorithms, such as the PC algorithm, GES, and GSP, and show that our method achieves comparable performance with a shorter runtime.
arXiv Detail & Related papers (2020-11-06T21:56:41Z) - An Asymptotically Optimal Primal-Dual Incremental Algorithm for
Contextual Linear Bandits [129.1029690825929]
We introduce a novel algorithm improving over the state-of-the-art along multiple dimensions.
We establish minimax optimality for any learning horizon in the special case of non-contextual linear bandits.
arXiv Detail & Related papers (2020-10-23T09:12:47Z) - Efficient Computation of Expectations under Spanning Tree Distributions [67.71280539312536]
We propose unified algorithms for the important cases of first-order expectations and second-order expectations in edge-factored, non-projective spanning-tree models.
Our algorithms exploit a fundamental connection between gradients and expectations, which allows us to derive efficient algorithms.
arXiv Detail & Related papers (2020-08-29T14:58:26Z) - Online Dense Subgraph Discovery via Blurred-Graph Feedback [87.9850024070244]
We introduce a novel learning problem for dense subgraph discovery.
We first propose a edge-time algorithm that obtains a nearly-optimal solution with high probability.
We then design a more scalable algorithm with a theoretical guarantee.
arXiv Detail & Related papers (2020-06-24T11:37:33Z) - Corralling Stochastic Bandit Algorithms [54.10645564702416]
We show that the regret of the corralling algorithms is no worse than that of the best algorithm containing the arm with the highest reward.
We show that the gap between the highest reward and other rewards depends on the gap between the highest reward and other rewards.
arXiv Detail & Related papers (2020-06-16T15:33:12Z) - Lagrangian Decomposition for Neural Network Verification [148.0448557991349]
A fundamental component of neural network verification is the computation of bounds on the values their outputs can take.
We propose a novel approach based on Lagrangian Decomposition.
We show that we obtain bounds comparable with off-the-shelf solvers in a fraction of their running time.
arXiv Detail & Related papers (2020-02-24T17:55:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.