Solving ESL Sentence Completion Questions via Pre-trained Neural
Language Models
- URL: http://arxiv.org/abs/2107.07122v1
- Date: Thu, 15 Jul 2021 05:01:39 GMT
- Title: Solving ESL Sentence Completion Questions via Pre-trained Neural
Language Models
- Authors: Qiongqiong Liu, Tianqiao Liu, Jiafu Zhao, Qiang Fang, Wenbiao Ding,
Zhongqin Wu, Feng Xia, Jiliang Tang, Zitao Liu
- Abstract summary: Sentence completion (SC) questions present a sentence with one or more blanks that need to be filled in.
We propose a neural framework to solve SC questions in English examinations by utilizing pre-trained language models.
- Score: 33.41201869566935
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sentence completion (SC) questions present a sentence with one or more blanks
that need to be filled in, three to five possible words or phrases as options.
SC questions are widely used for students learning English as a Second Language
(ESL) and building computational approaches to automatically solve such
questions is beneficial to language learners. In this work, we propose a neural
framework to solve SC questions in English examinations by utilizing
pre-trained language models. We conduct extensive experiments on a real-world
K-12 ESL SC question dataset and the results demonstrate the superiority of our
model in terms of prediction accuracy. Furthermore, we run precision-recall
trade-off analysis to discuss the practical issues when deploying it in
real-life scenarios. To encourage reproducible results, we make our code
publicly available at \url{https://github.com/AIED2021/ESL-SentenceCompletion}.
Related papers
- A Novel Cartography-Based Curriculum Learning Method Applied on RoNLI: The First Romanian Natural Language Inference Corpus [71.77214818319054]
Natural language inference is a proxy for natural language understanding.
There is no publicly available NLI corpus for the Romanian language.
We introduce the first Romanian NLI corpus (RoNLI) comprising 58K training sentence pairs.
arXiv Detail & Related papers (2024-05-20T08:41:15Z) - Question Translation Training for Better Multilingual Reasoning [108.10066378240879]
Large language models show compelling performance on reasoning tasks but they tend to perform much worse in languages other than English.
A typical solution is to translate instruction data into all languages of interest, and then train on the resulting multilingual data, which is called translate-training.
In this paper we explore the benefits of question alignment, where we train the model to translate reasoning questions into English by finetuning on X-English parallel question data.
arXiv Detail & Related papers (2024-01-15T16:39:10Z) - SC-Ques: A Sentence Completion Question Dataset for English as a Second
Language Learners [22.566710467490182]
Sentence completion (SC) questions present a sentence with one or more blanks that need to be filled in, three to five possible words or phrases as options.
We present a large-scale SC dataset, textscSC-Ques, which is made up of 289,148 ESL SC questions from real-world standardized English examinations.
arXiv Detail & Related papers (2022-06-24T02:17:13Z) - Automatic Short Math Answer Grading via In-context Meta-learning [2.0263791972068628]
We study the problem of automatic short answer grading for students' responses to math questions.
We use MathBERT, a variant of the popular language model BERT adapted to mathematical content, as our base model.
Second, we use an in-context learning approach that provides scoring examples as input to the language model.
arXiv Detail & Related papers (2022-05-30T16:26:02Z) - From Good to Best: Two-Stage Training for Cross-lingual Machine Reading
Comprehension [51.953428342923885]
We develop a two-stage approach to enhance the model performance.
The first stage targets at recall: we design a hard-learning (HL) algorithm to maximize the likelihood that the top-k predictions contain the accurate answer.
The second stage focuses on precision: an answer-aware contrastive learning mechanism is developed to learn the fine difference between the accurate answer and other candidates.
arXiv Detail & Related papers (2021-12-09T07:31:15Z) - Intent Classification Using Pre-Trained Embeddings For Low Resource
Languages [67.40810139354028]
Building Spoken Language Understanding systems that do not rely on language specific Automatic Speech Recognition is an important yet less explored problem in language processing.
We present a comparative study aimed at employing a pre-trained acoustic model to perform Spoken Language Understanding in low resource scenarios.
We perform experiments across three different languages: English, Sinhala, and Tamil each with different data sizes to simulate high, medium, and low resource scenarios.
arXiv Detail & Related papers (2021-10-18T13:06:59Z) - Question Generation for Adaptive Education [7.23389716633927]
We show how to fine-tune pre-trained language models for deep knowledge tracing (LM-KT)
This model accurately predicts the probability of a student answering a question correctly, and generalizes to questions not seen in training.
We then use LM-KT to specify the objective and data for training a model to generate questions conditioned on the student and target difficulty.
arXiv Detail & Related papers (2021-06-08T11:46:59Z) - Learning Spoken Language Representations with Neural Lattice Language
Modeling [39.50831917042577]
We propose a framework that trains neural lattice language models to provide contextualized representations for spoken language understanding tasks.
The proposed two-stage pre-training approach reduces the demands of speech data and has better efficiency.
arXiv Detail & Related papers (2020-07-06T10:38:03Z) - How to Probe Sentence Embeddings in Low-Resource Languages: On
Structural Design Choices for Probing Task Evaluation [82.96358326053115]
We investigate sensitivity of probing task results to structural design choices.
We probe embeddings in a multilingual setup with design choices that lie in a'stable region', as we identify for English.
We find that results on English do not transfer to other languages.
arXiv Detail & Related papers (2020-06-16T12:37:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.