Imputing Knowledge Tracing Data with Subject-Based Training via LSTM
Variational Autoencoders Frameworks
- URL: http://arxiv.org/abs/2302.12910v1
- Date: Fri, 24 Feb 2023 21:56:03 GMT
- Title: Imputing Knowledge Tracing Data with Subject-Based Training via LSTM
Variational Autoencoders Frameworks
- Authors: Jia Tracy Shen, Dongwon Lee
- Abstract summary: We adopt a subject-based training method to split and impute data by student IDs instead of row number splitting.
We leverage two existing deep generative frameworks, namely variational Autoencoders (VAE) and Longitudinal Variational Autoencoders (LVAE)
We demonstrate that the generated data from LSTM-VAE and LSTM-LVAE can boost the original model performance by about 50%.
- Score: 6.24828623162058
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The issue of missing data poses a great challenge on boosting performance and
application of deep learning models in the {\em Knowledge Tracing} (KT)
problem. However, there has been the lack of understanding on the issue in the
literature. %are not sufficient studies tackling this problem. In this work, to
address this challenge, we adopt a subject-based training method to split and
impute data by student IDs instead of row number splitting which we call
non-subject based training. The benefit of subject-based training can retain
the complete sequence for each student and hence achieve efficient training.
Further, we leverage two existing deep generative frameworks, namely
variational Autoencoders (VAE) and Longitudinal Variational Autoencoders (LVAE)
frameworks and build LSTM kernels into them to form LSTM-VAE and LSTM LVAE
(noted as VAE and LVAE for simplicity) models to generate quality data. In
LVAE, a Gaussian Process (GP) model is trained to disentangle the correlation
between the subject (i.e., student) descriptor information (e.g., age, gender)
and the latent space. The paper finally compare the model performance between
training the original data and training the data imputed with generated data
from non-subject based model VAE-NS and subject-based training models (i.e.,
VAE and LVAE). We demonstrate that the generated data from LSTM-VAE and
LSTM-LVAE can boost the original model performance by about 50%. Moreover, the
original model just needs 10% more student data to surpass the original
performance if the prediction model is small and 50\% more data if the
prediction model is large with our proposed frameworks.
Related papers
- Rethinking Data Synthesis: A Teacher Model Training Recipe with Interpretation [8.013158752919722]
Recent advances in large language model (LLM) training have highlighted the need for diverse, high-quality instruction data.
We propose a paradigm shift named textbfNOMAD by investigating how to specifically train models for data generation.
arXiv Detail & Related papers (2024-10-27T07:38:39Z) - Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration [90.41908331897639]
Large language models (LLMs) have significantly benefited from training on diverse, high-quality task-specific data.
We present a novel approach, ReverseGen, designed to automatically generate effective training samples.
arXiv Detail & Related papers (2024-10-22T06:43:28Z) - Data Shapley in One Training Run [88.59484417202454]
Data Shapley provides a principled framework for attributing data's contribution within machine learning contexts.
Existing approaches require re-training models on different data subsets, which is computationally intensive.
This paper introduces In-Run Data Shapley, which addresses these limitations by offering scalable data attribution for a target model of interest.
arXiv Detail & Related papers (2024-06-16T17:09:24Z) - How to Train Data-Efficient LLMs [56.41105687693619]
We study data-efficient approaches for pre-training language models (LLMs)
We find that Ask-LLM and Density sampling are the best methods in their respective categories.
In our comparison of 19 samplers, involving hundreds of evaluation tasks and pre-training runs, we find that Ask-LLM and Density are the best methods in their respective categories.
arXiv Detail & Related papers (2024-02-15T02:27:57Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - Comparison of Transfer Learning based Additive Manufacturing Models via
A Case Study [3.759936323189418]
This paper defines a case study based on an open-source dataset about metal AM products.
Five TL methods are integrated with decision tree regression (DTR) and artificial neural network (ANN) to construct six TL-based models.
The comparisons are used to quantify the performance of applied TL methods and are discussed from the perspective of similarity, training data size, and data preprocessing.
arXiv Detail & Related papers (2023-05-17T00:29:25Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Self-Supervised Pre-Training for Transformer-Based Person
Re-Identification [54.55281692768765]
Transformer-based supervised pre-training achieves great performance in person re-identification (ReID)
Due to the domain gap between ImageNet and ReID datasets, it usually needs a larger pre-training dataset to boost the performance.
This work aims to mitigate the gap between the pre-training and ReID datasets from the perspective of data and model structure.
arXiv Detail & Related papers (2021-11-23T18:59:08Z) - Reconstructing Training Data from Diverse ML Models by Ensemble
Inversion [8.414622657659168]
Model Inversion (MI), in which an adversary abuses access to a trained Machine Learning (ML) model, has attracted increasing research attention.
We propose an ensemble inversion technique that estimates the distribution of original training data by training a generator constrained by an ensemble of trained models.
We achieve high quality results without any dataset and show how utilizing an auxiliary dataset that's similar to the presumed training data improves the results.
arXiv Detail & Related papers (2021-11-05T18:59:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.