Assisting Language Learners: Automated Trans-Lingual Definition
Generation via Contrastive Prompt Learning
- URL: http://arxiv.org/abs/2306.06058v1
- Date: Fri, 9 Jun 2023 17:32:45 GMT
- Title: Assisting Language Learners: Automated Trans-Lingual Definition
Generation via Contrastive Prompt Learning
- Authors: Hengyuan Zhang, Dawei Li, Yanran Li, Chenming Shang, Chufan Shi, Yong
Jiang
- Abstract summary: The standard definition generation task requires to automatically produce mono-lingual definitions.
We propose a novel task of Trans-Lingual Definition Generation (TLDG), which aims to generate definitions in another language.
- Score: 25.851611353632926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The standard definition generation task requires to automatically produce
mono-lingual definitions (e.g., English definitions for English words), but
ignores that the generated definitions may also consist of unfamiliar words for
language learners. In this work, we propose a novel task of Trans-Lingual
Definition Generation (TLDG), which aims to generate definitions in another
language, i.e., the native speaker's language. Initially, we explore the
unsupervised manner of this task and build up a simple implementation of
fine-tuning the multi-lingual machine translation model. Then, we develop two
novel methods, Prompt Combination and Contrastive Prompt Learning, for further
enhancing the quality of the generation. Our methods are evaluated against the
baseline Pipeline method in both rich- and low-resource settings, and we
empirically establish its superiority in generating higher-quality
trans-lingual definitions.
Related papers
- DECIDER: A Dual-System Rule-Controllable Decoding Framework for Language Generation [57.07295906718989]
Constrained decoding approaches aim to control the meaning or style of text generated by a Pre-trained Language Model (PLM) using specific target words during inference.
We propose a novel decoding framework, DECIDER, which enables us to program rules on how we complete tasks to control a PLM.
arXiv Detail & Related papers (2024-03-04T11:49:08Z) - Zur Darstellung eines mehrstufigen Prototypbegriffs in der
multilingualen automatischen Sprachgenerierung: vom Korpus \"uber word
embeddings bis hin zum automatischen W\"orterbuch [0.0]
The multilingual dictionary of noun valency Portlex is considered to be the trigger for the creation of the automatic language generators Xera and Combinatoria.
Both prototypes are used for the automatic generation of nominal phrases.
arXiv Detail & Related papers (2023-12-26T19:39:25Z) - Why Does Zero-Shot Cross-Lingual Generation Fail? An Explanation and a
Solution [0.9085116579988537]
We show that the fine-tuning process learns language invariant representations, which is beneficial for classification tasks but harmful for generation tasks.
Experiments on three semantically diverse generation tasks show that our method reduces the accidental translation problem by 68% and improves the ROUGE-L score by 1.5 on average.
arXiv Detail & Related papers (2023-05-27T02:04:19Z) - Learning to Solve Voxel Building Embodied Tasks from Pixels and Natural
Language Instructions [53.21504989297547]
We propose a new method that combines a language model and reinforcement learning for the task of building objects in a Minecraft-like environment.
Our method first generates a set of consistently achievable sub-goals from the instructions and then completes associated sub-tasks with a pre-trained RL policy.
arXiv Detail & Related papers (2022-11-01T18:30:42Z) - Multitasking Framework for Unsupervised Simple Definition Generation [5.2221935174520056]
We propose a novel task of Simple Definition Generation to help language learners and low literacy readers.
A significant challenge of this task is the lack of learner's dictionaries in many languages.
We propose a multitasking framework SimpDefiner that only requires a standard dictionary with complex definitions and a corpus containing arbitrary simple texts.
arXiv Detail & Related papers (2022-03-24T08:16:04Z) - Controllable Natural Language Generation with Contrastive Prefixes [120.12778570283956]
GPT2 generation utilizes a set of small attribute-specific vectors, called prefixes, to steer natural language generation.
We propose a novel supervised method and also an unsupervised method to train the prefixes for single-aspect control.
Experimental results on both single-aspect and multi-aspect control show that our methods can guide generation towards the desired attributes while keeping high linguistic quality.
arXiv Detail & Related papers (2022-02-27T00:31:03Z) - Analyzing the Limits of Self-Supervision in Handling Bias in Language [52.26068057260399]
We evaluate how well language models capture the semantics of four tasks for bias: diagnosis, identification, extraction and rephrasing.
Our analyses indicate that language models are capable of performing these tasks to widely varying degrees across different bias dimensions, such as gender and political affiliation.
arXiv Detail & Related papers (2021-12-16T05:36:08Z) - Neural Abstructions: Abstractions that Support Construction for Grounded
Language Learning [69.1137074774244]
Leveraging language interactions effectively requires addressing limitations in the two most common approaches to language grounding.
We introduce the idea of neural abstructions: a set of constraints on the inference procedure of a label-conditioned generative model.
We show that with this method a user population is able to build a semantic modification for an open-ended house task in Minecraft.
arXiv Detail & Related papers (2021-07-20T07:01:15Z) - Toward Cross-Lingual Definition Generation for Language Learners [10.45755551957024]
We propose to generate definitions in English for words in various languages.
Models can be directly applied to other languages after trained on the English dataset.
Experiments and manual analyses show that our models have a strong cross-lingual transfer ability.
arXiv Detail & Related papers (2020-10-12T08:45:28Z) - On the Importance of Word Order Information in Cross-lingual Sequence
Labeling [80.65425412067464]
Cross-lingual models that fit into the word order of the source language might fail to handle target languages.
We investigate whether making models insensitive to the word order of the source language can improve the adaptation performance in target languages.
arXiv Detail & Related papers (2020-01-30T03:35:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.