A New Score for Adaptive Tests in Bayesian and Credal Networks
- URL: http://arxiv.org/abs/2105.12205v1
- Date: Tue, 25 May 2021 20:35:42 GMT
- Title: A New Score for Adaptive Tests in Bayesian and Credal Networks
- Authors: Alessandro Antonucci and Francesca Mangili and Claudio Bonesana and
Giorgia Adorni
- Abstract summary: A test is adaptive when its sequence and number of questions is dynamically tuned on the basis of the estimated skills of the taker.
We present an alternative family of scores, based on the mode of the posterior probabilities, and hence easier to explain.
- Score: 64.80185026979883
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A test is adaptive when its sequence and number of questions is dynamically
tuned on the basis of the estimated skills of the taker. Graphical models, such
as Bayesian networks, are used for adaptive tests as they allow to model the
uncertainty about the questions and the skills in an explainable fashion,
especially when coping with multiple skills. A better elicitation of the
uncertainty in the question/skills relations can be achieved by interval
probabilities. This turns the model into a credal network, thus making more
challenging the inferential complexity of the queries required to select
questions. This is especially the case for the information theoretic quantities
used as scores to drive the adaptive mechanism. We present an alternative
family of scores, based on the mode of the posterior probabilities, and hence
easier to explain. This makes considerably simpler the evaluation in the credal
case, without significantly affecting the quality of the adaptive process.
Numerical tests on synthetic and real-world data are used to support this
claim.
Related papers
- Adaptive Data-Free Quantization [44.00726062583708]
Data-free quantization (DFQ) recovers the performance of quantized network (Q) without the original data.
DFQ generates the fake sample via a generator (G) by learning from full-precision network (P)
We propose an Adaptive Data-Free Quantization (AdaDFQ) method, which revisits DFQ from a zero-sum game perspective.
arXiv Detail & Related papers (2023-03-13T05:37:40Z) - Efficient Test-Time Model Adaptation without Forgetting [60.36499845014649]
Test-time adaptation seeks to tackle potential distribution shifts between training and testing data.
We propose an active sample selection criterion to identify reliable and non-redundant samples.
We also introduce a Fisher regularizer to constrain important model parameters from drastic changes.
arXiv Detail & Related papers (2022-04-06T06:39:40Z) - Logically Consistent Loss for Visual Question Answering [66.83963844316561]
The current advancement in neural-network based Visual Question Answering (VQA) cannot ensure such consistency due to identically distribution (i.i.d.) assumption.
We propose a new model-agnostic logic constraint to tackle this issue by formulating a logically consistent loss in the multi-task learning framework.
Experiments confirm that the proposed loss formulae and introduction of hybrid-batch leads to more consistency as well as better performance.
arXiv Detail & Related papers (2020-11-19T20:31:05Z) - On the Robustness of Active Learning [0.7340017786387767]
Active Learning is concerned with how to identify the most useful samples for a Machine Learning algorithm to be trained with.
We find that it is often applied with not enough care and domain knowledge.
We propose the new "Sum of Squared Logits" method based on the Simpson diversity index and investigate the effect of using the confusion matrix for balancing in sample selection.
arXiv Detail & Related papers (2020-06-18T09:07:23Z) - Selective Question Answering under Domain Shift [90.021577320085]
Abstention policies based solely on the model's softmax probabilities fare poorly, since models are overconfident on out-of-domain inputs.
We train a calibrator to identify inputs on which the QA model errs, and abstain when it predicts an error is likely.
Our method answers 56% of questions while maintaining 80% accuracy; in contrast, directly using the model's probabilities only answers 48% at 80% accuracy.
arXiv Detail & Related papers (2020-06-16T19:13:21Z) - Robust Question Answering Through Sub-part Alignment [53.94003466761305]
We model question answering as an alignment problem.
We train our model on SQuAD v1.1 and test it on several adversarial and out-of-domain datasets.
arXiv Detail & Related papers (2020-04-30T09:10:57Z) - Manifold-based Test Generation for Image Classifiers [7.226144684379191]
To test an image classification neural network, one must obtain realistic test data adequate enough to inspire confidence.
This paper proposes a novel framework to address these challenges.
Experiments show that this approach enables generation of thousands of realistic yet fault-revealing test cases efficiently even for well-trained models.
arXiv Detail & Related papers (2020-02-15T07:53:34Z) - ManyModalQA: Modality Disambiguation and QA over Diverse Inputs [73.93607719921945]
We present a new multimodal question answering challenge, ManyModalQA, in which an agent must answer a question by considering three distinct modalities.
We collect our data by scraping Wikipedia and then utilize crowdsourcing to collect question-answer pairs.
arXiv Detail & Related papers (2020-01-22T14:39:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.