Italian Crossword Generator: Enhancing Education through Interactive
Word Puzzles
- URL: http://arxiv.org/abs/2311.15723v1
- Date: Mon, 27 Nov 2023 11:17:29 GMT
- Title: Italian Crossword Generator: Enhancing Education through Interactive
Word Puzzles
- Authors: Kamyar Zeinalipour, Tommaso laquinta, Asya Zanollo, Giovanni Angelini,
Leonardo Rigutini, Marco Maggini, Marco Gori
- Abstract summary: We develop a comprehensive system for generating and verifying crossword clues.
A dataset of clue-answer pairs was compiled to fine-tune the models.
For generating crossword clues from a given text, Zero/Few-shot learning techniques were used.
- Score: 9.84767617576152
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Educational crosswords offer numerous benefits for students, including
increased engagement, improved understanding, critical thinking, and memory
retention. Creating high-quality educational crosswords can be challenging, but
recent advances in natural language processing and machine learning have made
it possible to use language models to generate nice wordplays. The exploitation
of cutting-edge language models like GPT3-DaVinci, GPT3-Curie, GPT3-Babbage,
GPT3-Ada, and BERT-uncased has led to the development of a comprehensive system
for generating and verifying crossword clues. A large dataset of clue-answer
pairs was compiled to fine-tune the models in a supervised manner to generate
original and challenging clues from a given keyword. On the other hand, for
generating crossword clues from a given text, Zero/Few-shot learning techniques
were used to extract clues from the input text, adding variety and creativity
to the puzzles. We employed the fine-tuned model to generate data and labeled
the acceptability of clue-answer parts with human supervision. To ensure
quality, we developed a classifier by fine-tuning existing language models on
the labeled dataset. Conversely, to assess the quality of clues generated from
the given text using zero/few-shot learning, we employed a zero-shot learning
approach to check the quality of generated clues. The results of the evaluation
have been very promising, demonstrating the effectiveness of the approach in
creating high-standard educational crosswords that offer students engaging and
rewarding learning experiences.
Related papers
- Clue-Instruct: Text-Based Clue Generation for Educational Crossword Puzzles [10.375451846093327]
We propose a methodology to build educational clue generation datasets that can be used to instruct Large Language Models.
By gathering from Wikipedia pages informative content associated with relevant keywords, we use Large Language Models to automatically generate pedagogical clues.
We used clue-instruct to instruct different LLMs to generate educational clues from a given input content and keyword.
arXiv Detail & Related papers (2024-04-09T10:12:34Z) - Retrieval is Accurate Generation [99.24267226311157]
We introduce a novel method that selects context-aware phrases from a collection of supporting documents.
Our model achieves the best performance and the lowest latency among several retrieval-augmented baselines.
arXiv Detail & Related papers (2024-02-27T14:16:19Z) - ArabIcros: AI-Powered Arabic Crossword Puzzle Generation for Educational
Applications [11.881406917880287]
This paper presents the first Arabic crossword puzzle generator driven by advanced AI technology.
Leveraging cutting-edge large language models including GPT4, GPT3-Davinci, GPT3-Curie, GPT3-Babbage, GPT3-Ada, and BERT, the system generates distinctive and challenging clues.
arXiv Detail & Related papers (2023-12-03T10:03:50Z) - Investigating the Efficacy of Large Language Models in Reflective
Assessment Methods through Chain of Thoughts Prompting [0.2552922646705803]
Chain of Thought(CoT) prompting method has been proposed as a means to enhance LLMs' proficiency in complex reasoning tasks.
The primary aim of this research is to assess how well four language models can grade reflective essays of third-year medical students.
arXiv Detail & Related papers (2023-09-30T06:25:27Z) - MOCHA: A Multi-Task Training Approach for Coherent Text Generation from
Cognitive Perspective [22.69509556890676]
We propose a novel multi-task training strategy for coherent text generation grounded on the cognitive theory of writing.
We extensively evaluate our model on three open-ended generation tasks including story generation, news article writing and argument generation.
arXiv Detail & Related papers (2022-10-26T11:55:41Z) - Informative Text Generation from Knowledge Triples [56.939571343797304]
We propose a novel memory augmented generator that employs a memory network to memorize the useful knowledge learned during the training.
We derive a dataset from WebNLG for our new setting and conduct extensive experiments to investigate the effectiveness of our model.
arXiv Detail & Related papers (2022-09-26T14:35:57Z) - Evidentiality-guided Generation for Knowledge-Intensive NLP Tasks [59.761411682238645]
Retrieval-augmented generation models have shown state-of-the-art performance across many knowledge-intensive NLP tasks.
We introduce a method to incorporate evidentiality of passages -- whether a passage contains correct evidence to support the output -- into training the generator.
arXiv Detail & Related papers (2021-12-16T08:18:47Z) - VidLanKD: Improving Language Understanding via Video-Distilled Knowledge
Transfer [76.3906723777229]
We present VidLanKD, a video-language knowledge distillation method for improving language understanding.
We train a multi-modal teacher model on a video-text dataset, and then transfer its knowledge to a student language model with a text dataset.
In our experiments, VidLanKD achieves consistent improvements over text-only language models and vokenization models.
arXiv Detail & Related papers (2021-07-06T15:41:32Z) - Decrypting Cryptic Crosswords: Semantically Complex Wordplay Puzzles as
a Target for NLP [5.447716844779342]
Cryptic crosswords are the dominant English-language crossword variety in the United Kingdom.
We present a dataset of cryptic crossword clues that can be used as a benchmark and train a sequence-to-sequence model to solve them.
We show that performance can be substantially improved using a novel curriculum learning approach.
arXiv Detail & Related papers (2021-04-17T18:54:00Z) - Language Generation with Multi-Hop Reasoning on Commonsense Knowledge
Graph [124.45799297285083]
We argue that exploiting both the structural and semantic information of the knowledge graph facilitates commonsense-aware text generation.
We propose Generation with Multi-Hop Reasoning Flow (GRF) that enables pre-trained models with dynamic multi-hop reasoning on multi-relational paths extracted from the external commonsense knowledge graph.
arXiv Detail & Related papers (2020-09-24T13:55:32Z) - Exploiting Structured Knowledge in Text via Graph-Guided Representation
Learning [73.0598186896953]
We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs.
Building upon entity-level masked language models, our first contribution is an entity masking scheme.
In contrast to existing paradigms, our approach uses knowledge graphs implicitly, only during pre-training.
arXiv Detail & Related papers (2020-04-29T14:22:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.