Prob2Vec: Mathematical Semantic Embedding for Problem Retrieval in
Adaptive Tutoring
- URL: http://arxiv.org/abs/2003.10838v1
- Date: Sat, 21 Mar 2020 00:16:14 GMT
- Title: Prob2Vec: Mathematical Semantic Embedding for Problem Retrieval in
Adaptive Tutoring
- Authors: Du Su, Ali Yekkehkhany, Yi Lu, Wenmiao Lu
- Abstract summary: It is difficult for humans to determine a similarity score that is consistent across a large enough training set.
We propose a hierarchical problem embedding algorithm, called Prob2Vec, that consists of abstraction and embedding steps.
Prob2Vec 96.88% accuracy on a problem similarity test, in contrast to 75% from directly applying state-of-the-art sentence embedding methods.
- Score: 4.230510356675453
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new application of embedding techniques for problem retrieval in
adaptive tutoring. The objective is to retrieve problems whose mathematical
concepts are similar. There are two challenges: First, like sentences, problems
helpful to tutoring are never exactly the same in terms of the underlying
concepts. Instead, good problems mix concepts in innovative ways, while still
displaying continuity in their relationships. Second, it is difficult for
humans to determine a similarity score that is consistent across a large enough
training set. We propose a hierarchical problem embedding algorithm, called
Prob2Vec, that consists of abstraction and embedding steps. Prob2Vec achieves
96.88\% accuracy on a problem similarity test, in contrast to 75\% from
directly applying state-of-the-art sentence embedding methods. It is
interesting that Prob2Vec is able to distinguish very fine-grained differences
among problems, an ability humans need time and effort to acquire. In addition,
the sub-problem of concept labeling with imbalanced training data set is
interesting in its own right. It is a multi-label problem suffering from
dimensionality explosion, which we propose ways to ameliorate. We propose the
novel negative pre-training algorithm that dramatically reduces false negative
and positive ratios for classification, using an imbalanced training data set.
Related papers
- From Variance to Veracity: Unbundling and Mitigating Gradient Variance in Differentiable Bundle Adjustment Layers [10.784222655465264]
Various pose estimation and tracking problems in robotics can be decomposed into a correspondence estimation problem and a weighted least squares optimization problem.
Recent work has shown that coupling the two problems by iteratively refining one conditioned on the other's output yields SOTA results across domains.
We show that training these models has proved challenging, requiring a litany of tricks to stabilize and speed up training.
arXiv Detail & Related papers (2024-06-12T00:41:25Z) - Semisupervised score based matching algorithm to evaluate the effect of public health interventions [3.221788913179251]
In one-to-one matching algorithms, a large number of "pairs" to be matched could mean both the information from a large sample and a large number of tasks.
We propose a novel one-to-one matching algorithm based on a quadratic score function $S_beta(x_i,x_j)= betaT (x_i-x_j)(x_i-x_j)T beta$.
arXiv Detail & Related papers (2024-03-19T02:24:16Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Recovering Top-Two Answers and Confusion Probability in Multi-Choice
Crowdsourcing [10.508187462682308]
We consider crowdsourcing tasks with the goal of recovering not only the ground truth, but also the most confusing answer and the confusion probability.
We propose a model in which there are the top two plausible answers for each task, distinguished from the rest of the choices.
Under this model, we propose a two-stage inference algorithm to infer both the top two answers and the confusion probability.
arXiv Detail & Related papers (2022-12-29T09:46:39Z) - Rebalanced Siamese Contrastive Mining for Long-Tailed Recognition [120.80038161330623]
We show that supervised contrastive learning suffers a dual class-imbalance problem at both the original batch and Siamese batch levels.
We propose supervised hard positive and negative pairs mining to pick up informative pairs for contrastive computation and improve representation learning.
arXiv Detail & Related papers (2022-03-22T07:30:38Z) - Deep Probabilistic Graph Matching [72.6690550634166]
We propose a deep learning-based graph matching framework that works for the original QAP without compromising on the matching constraints.
The proposed method is evaluated on three popularly tested benchmarks (Pascal VOC, Willow Object and SPair-71k) and it outperforms all previous state-of-the-arts on all benchmarks.
arXiv Detail & Related papers (2022-01-05T13:37:27Z) - Simple Stochastic and Online Gradient DescentAlgorithms for Pairwise
Learning [65.54757265434465]
Pairwise learning refers to learning tasks where the loss function depends on a pair instances.
Online descent (OGD) is a popular approach to handle streaming data in pairwise learning.
In this paper, we propose simple and online descent to methods for pairwise learning.
arXiv Detail & Related papers (2021-11-23T18:10:48Z) - Scalable Personalised Item Ranking through Parametric Density Estimation [53.44830012414444]
Learning from implicit feedback is challenging because of the difficult nature of the one-class problem.
Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem.
We propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart.
arXiv Detail & Related papers (2021-05-11T03:38:16Z) - Iterative Weak Learnability and Multi-Class AdaBoost [0.0]
We construct an efficient ensemble algorithm for the multi-class classification problem inspired by SAMME.
In contrast to SAMME, our algorithm's final hypothesis converges to the correct label with probability 1.
The sum of the training error and an additional term, that depends only on the sample size, bounds the generalization error of our algorithm as the Adaptive Boosting algorithm.
arXiv Detail & Related papers (2021-01-26T03:30:30Z) - Learning to Actively Learn: A Robust Approach [22.75298609290053]
This work proposes a procedure for designing algorithms for adaptive data collection tasks like active learning and pure-exploration multi-armed bandits.
Our adaptive algorithm is learned via adversarial training over equivalence classes of problems derived from information theoretic lower bounds.
We perform synthetic experiments to justify the stability and effectiveness of the training procedure, and then evaluate the method on tasks derived from real data.
arXiv Detail & Related papers (2020-10-29T06:48:22Z) - Multi-task Supervised Learning via Cross-learning [102.64082402388192]
We consider a problem known as multi-task learning, consisting of fitting a set of regression functions intended for solving different tasks.
In our novel formulation, we couple the parameters of these functions, so that they learn in their task specific domains while staying close to each other.
This facilitates cross-fertilization in which data collected across different domains help improving the learning performance at each other task.
arXiv Detail & Related papers (2020-10-24T21:35:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.