Parametric Constraints for Bayesian Knowledge Tracing from First
Principles
- URL: http://arxiv.org/abs/2401.09456v1
- Date: Sat, 23 Dec 2023 03:58:41 GMT
- Title: Parametric Constraints for Bayesian Knowledge Tracing from First
Principles
- Authors: Denis Shchepakin, Sreecharan Sankaranarayanan, Dawn Zimmaro
- Abstract summary: This paper takes a "from first principles" approach to deriving constraints that can be imposed on the BKT parameter space.
The paper further introduces a novel algorithm for estimating BKT parameters subject to the newly defined constraints.
- Score: 0.276240219662896
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Bayesian Knowledge Tracing (BKT) is a probabilistic model of a learner's
state of mastery corresponding to a knowledge component. It considers the
learner's state of mastery as a "hidden" or latent binary variable and updates
this state based on the observed correctness of the learner's response using
parameters that represent transition probabilities between states. BKT is often
represented as a Hidden Markov Model and the Expectation-Maximization (EM)
algorithm is used to infer these parameters. However, this algorithm can suffer
from several issues including producing multiple viable sets of parameters,
settling into a local minima, producing degenerate parameter values, and a high
computational cost during fitting. This paper takes a "from first principles"
approach to deriving constraints that can be imposed on the BKT parameter
space. Starting from the basic mathematical truths of probability and building
up to the behaviors expected of the BKT parameters in real systems, this paper
presents a mathematical derivation that results in succinct constraints that
can be imposed on the BKT parameter space. Since these constraints are
necessary conditions, they can be applied prior to fitting in order to reduce
computational cost and the likelihood of issues that can emerge from the EM
procedure. In order to see that promise through, the paper further introduces a
novel algorithm for estimating BKT parameters subject to the newly defined
constraints. While the issue of degenerate parameter values has been reported
previously, this paper is the first, to our best knowledge, to derive the
constrains from first principles while also presenting an algorithm that
respects those constraints.
Related papers
- A Comparative Study of MAP and LMMSE Estimators for Blind Inverse Problems [0.17188280334580194]
We show how two synthetic MAP approaches can be used to reduce the inherent non-dimensionality problem.<n>We also show that the LMMSE estimator can serve as an alternative that can circumvent the limitations.
arXiv Detail & Related papers (2026-02-12T10:49:45Z) - Learning Algorithms for Verification of Markov Decision Processes [20.5951492453299]
We present a general framework for applying learning algorithms to the verification of Markov decision processes (MDPs)
The presented framework focuses on probabilistic reachability, which is a core problem in verification.
arXiv Detail & Related papers (2024-03-14T08:54:19Z) - Online non-parametric likelihood-ratio estimation by Pearson-divergence
functional minimization [55.98760097296213]
We introduce a new framework for online non-parametric LRE (OLRE) for the setting where pairs of iid observations $(x_t sim p, x'_t sim q)$ are observed over time.
We provide theoretical guarantees for the performance of the OLRE method along with empirical validation in synthetic experiments.
arXiv Detail & Related papers (2023-11-03T13:20:11Z) - Sub-linear Regret in Adaptive Model Predictive Control [56.705978425244496]
We present STT-MPC (Self-Tuning Tube-based Model Predictive Control), an online oracle that combines the certainty-equivalence principle and polytopic tubes.
We analyze the regret of the algorithm, when compared to an algorithm initially aware of the system dynamics.
arXiv Detail & Related papers (2023-10-07T15:07:10Z) - Geometry-Aware Approaches for Balancing Performance and Theoretical
Guarantees in Linear Bandits [6.907555940790131]
Thompson sampling and Greedy demonstrate promising empirical performance, yet this contrasts with their pessimistic theoretical regret bounds.
We propose a new data-driven technique that tracks the geometric properties of the uncertainty ellipsoid.
We identify and course-correct" problem instances in which the base algorithms perform poorly.
arXiv Detail & Related papers (2023-06-26T17:38:45Z) - Stochastic Marginal Likelihood Gradients using Neural Tangent Kernels [78.6096486885658]
We introduce lower bounds to the linearized Laplace approximation of the marginal likelihood.
These bounds are amenable togradient-based optimization and allow to trade off estimation accuracy against computational complexity.
arXiv Detail & Related papers (2023-06-06T19:02:57Z) - PriorCVAE: scalable MCMC parameter inference with Bayesian deep
generative modelling [12.820453440015553]
Recent have shown that GP priors can be encoded using deep generative models such as variational autoencoders (VAEs)
We show how VAEs can serve as drop-in replacements for the original priors during MCMC inference.
We propose PriorCVAE to encode solutions of ODEs.
arXiv Detail & Related papers (2023-04-09T20:23:26Z) - History-Based, Bayesian, Closure for Stochastic Parameterization:
Application to Lorenz '96 [0.09137554315375918]
We develop a new type of parameterization based on a Bayesian formalism for neural networks, to account for uncertainty quantification.
We apply the proposed Bayesian history-based parameterization to the Lorenz '96 model in the presence of noisy and sparse data.
This approach paves the way for the use of Bayesian approaches for closure problems.
arXiv Detail & Related papers (2022-10-26T05:22:50Z) - Sample-Then-Optimize Batch Neural Thompson Sampling [50.800944138278474]
We introduce two algorithms for black-box optimization based on the Thompson sampling (TS) policy.
To choose an input query, we only need to train an NN and then choose the query by maximizing the trained NN.
Our algorithms sidestep the need to invert the large parameter matrix yet still preserve the validity of the TS policy.
arXiv Detail & Related papers (2022-10-13T09:01:58Z) - A Statistical Decision-Theoretical Perspective on the Two-Stage Approach
to Parameter Estimation [7.599399338954307]
Two-Stage (TS) Approach can be applied to obtain reliable parametric estimates.
We show how to apply the TS approach on models for independent and identically distributed samples.
arXiv Detail & Related papers (2022-03-31T18:19:47Z) - Bayesian Error-in-Variables Models for the Identification of Power
Networks [0.0]
A reliable estimate of the admittance matrix may either be missing or quicklybecome obsolete for temporally varying grids.
We propose a data-driven identificationmethod utilising voltage and current measurements collected from micro-PMUs.
arXiv Detail & Related papers (2021-07-09T15:10:47Z) - FLIP: A flexible initializer for arbitrarily-sized parametrized quantum
circuits [105.54048699217668]
We propose a FLexible Initializer for arbitrarily-sized Parametrized quantum circuits.
FLIP can be applied to any family of PQCs, and instead of relying on a generic set of initial parameters, it is tailored to learn the structure of successful parameters.
We illustrate the advantage of using FLIP in three scenarios: a family of problems with proven barren plateaus, PQC training to solve max-cut problem instances, and PQC training for finding the ground state energies of 1D Fermi-Hubbard models.
arXiv Detail & Related papers (2021-03-15T17:38:33Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.