Counterfactual Inference of Second Opinions
- URL: http://arxiv.org/abs/2203.08653v2
- Date: Thu, 30 Jun 2022 16:46:33 GMT
- Title: Counterfactual Inference of Second Opinions
- Authors: Nina L. Corvelo Benz and Manuel Gomez Rodriguez
- Abstract summary: Automated decision support systems that are able to infer second opinions from experts can potentially facilitate a more efficient allocation of resources.
This paper looks at the design of this type of support systems from the perspective of counterfactual inference.
Experiments on both synthetic and real data show that our model can be used to infer second opinions more accurately than its non-causal counterpart.
- Score: 13.93477033094828
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated decision support systems that are able to infer second opinions
from experts can potentially facilitate a more efficient allocation of
resources; they can help decide when and from whom to seek a second opinion. In
this paper, we look at the design of this type of support systems from the
perspective of counterfactual inference. We focus on a multiclass
classification setting and first show that, if experts make predictions on
their own, the underlying causal mechanism generating their predictions needs
to satisfy a desirable set invariant property. Further, we show that, for any
causal mechanism satisfying this property, there exists an equivalent mechanism
where the predictions by each expert are generated by independent
sub-mechanisms governed by a common noise. This motivates the design of a set
invariant Gumbel-Max structural causal model where the structure of the noise
governing the sub-mechanisms underpinning the model depends on an intuitive
notion of similarity between experts which can be estimated from data.
Experiments on both synthetic and real data show that our model can be used to
infer second opinions more accurately than its non-causal counterpart.
Related papers
- Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Invariant Causal Set Covering Machines [64.86459157191346]
Rule-based models, such as decision trees, appeal to practitioners due to their interpretable nature.
However, the learning algorithms that produce such models are often vulnerable to spurious associations and thus, they are not guaranteed to extract causally-relevant insights.
We propose Invariant Causal Set Covering Machines, an extension of the classical Set Covering Machine algorithm for conjunctions/disjunctions of binary-valued rules that provably avoids spurious associations.
arXiv Detail & Related papers (2023-06-07T20:52:01Z) - Designing Decision Support Systems Using Counterfactual Prediction Sets [15.121082690769525]
Decision support systems for classification tasks are predominantly designed to predict the value of the ground truth labels.
This paper revisits the design of this type of systems from the perspective of online learning.
We develop a methodology that does not require, nor assumes, an expert model.
arXiv Detail & Related papers (2023-06-06T18:00:09Z) - Differentially Private Counterfactuals via Functional Mechanism [47.606474009932825]
We propose a novel framework to generate differentially private counterfactual (DPC) without touching the deployed model or explanation set.
In particular, we train an autoencoder with the functional mechanism to construct noisy class prototypes, and then derive the DPC from the latent prototypes.
arXiv Detail & Related papers (2022-08-04T20:31:22Z) - Properties from Mechanisms: An Equivariance Perspective on Identifiable
Representation Learning [79.4957965474334]
Key goal of unsupervised representation learning is "inverting" a data generating process to recover its latent properties.
This paper asks, "Can we instead identify latent properties by leveraging knowledge of the mechanisms that govern their evolution?"
We provide a complete characterization of the sources of non-identifiability as we vary knowledge about a set of possible mechanisms.
arXiv Detail & Related papers (2021-10-29T14:04:08Z) - Understanding Interlocking Dynamics of Cooperative Rationalization [90.6863969334526]
Selective rationalization explains the prediction of complex neural networks by finding a small subset of the input that is sufficient to predict the neural model output.
We reveal a major problem with such cooperative rationalization paradigm -- model interlocking.
We propose a new rationalization framework, called A2R, which introduces a third component into the architecture, a predictor driven by soft attention as opposed to selection.
arXiv Detail & Related papers (2021-10-26T17:39:18Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - Model-Based Counterfactual Synthesizer for Interpretation [40.01787107375103]
We propose a Model-based Counterfactual Synthesizer (MCS) framework for interpreting machine learning models.
We first analyze the model-based counterfactual process and construct a base synthesizer using a conditional generative adversarial net (CGAN)
To better approximate the counterfactual universe for those rare queries, we novelly employ the umbrella sampling technique to conduct the MCS framework training.
arXiv Detail & Related papers (2021-06-16T17:09:57Z) - A theory of independent mechanisms for extrapolation in generative
models [20.794692397859755]
Generative models can be trained to emulate complex empirical data, but are they useful to make predictions in the context of previously unobserved environments?
We develop a theoretical framework to address this challenging situation by defining a weaker form of identifiability, based on the principle of independence of mechanisms.
We demonstrate on toy examples that classical gradient descent can hinder the model's extrapolation capabilities, suggesting independence of mechanisms should be enforced explicitly during training.
arXiv Detail & Related papers (2020-04-01T01:01:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.