Towards Red Teaming in Multimodal and Multilingual Translation
- URL: http://arxiv.org/abs/2401.16247v1
- Date: Mon, 29 Jan 2024 15:49:40 GMT
- Title: Towards Red Teaming in Multimodal and Multilingual Translation
- Authors: Christophe Ropers, David Dale, Prangthip Hansanti, Gabriel Mejia
Gonzalez, Ivan Evtimov, Corinne Wong, Christophe Touret, Kristina Pereyra,
Seohyun Sonia Kim, Cristian Canton Ferrer, Pierre Andrews and Marta R.
Costa-juss\`a
- Abstract summary: This paper presents the first study on human-based red teaming for Machine Translation.
It marks a significant step towards understanding and improving the performance of translation models.
We report lessons learned and provide recommendations for both translation models and red teaming drills.
- Score: 7.440772334845366
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Assessing performance in Natural Language Processing is becoming increasingly
complex. One particular challenge is the potential for evaluation datasets to
overlap with training data, either directly or indirectly, which can lead to
skewed results and overestimation of model performance. As a consequence, human
evaluation is gaining increasing interest as a means to assess the performance
and reliability of models. One such method is the red teaming approach, which
aims to generate edge cases where a model will produce critical errors. While
this methodology is becoming standard practice for generative AI, its
application to the realm of conditional AI remains largely unexplored. This
paper presents the first study on human-based red teaming for Machine
Translation (MT), marking a significant step towards understanding and
improving the performance of translation models. We delve into both human-based
red teaming and a study on automation, reporting lessons learned and providing
recommendations for both translation models and red teaming drills. This
pioneering work opens up new avenues for research and development in the field
of MT.
Related papers
- Self-Evolved Reward Learning for LLMs [45.6910747154447]
Reinforcement Learning from Human Feedback (RLHF) is a crucial technique for aligning language models with human preferences.
We propose Self-Evolved Reward Learning (SER), a novel approach where the RM generates additional training data to iteratively improve itself.
Our results demonstrate that even with limited human-annotated data, learning from self-feedback can robustly enhance RM performance.
arXiv Detail & Related papers (2024-11-01T07:29:03Z) - Rethinking Human-like Translation Strategy: Integrating Drift-Diffusion
Model with Large Language Models for Machine Translation [15.333148705267012]
We propose Thinker with the Drift-Diffusion Model to emulate human translators' dynamic decision-making under constrained resources.
We conduct experiments under the high-resource, low-resource, and commonsense translation settings using the WMT22 and CommonMT datasets.
We also perform additional analysis and evaluation on commonsense translation to illustrate the high effectiveness and efficacy of the proposed method.
arXiv Detail & Related papers (2024-02-16T14:00:56Z) - Generative Judge for Evaluating Alignment [84.09815387884753]
We propose a generative judge with 13B parameters, Auto-J, designed to address these challenges.
Our model is trained on user queries and LLM-generated responses under massive real-world scenarios.
Experimentally, Auto-J outperforms a series of strong competitors, including both open-source and closed-source models.
arXiv Detail & Related papers (2023-10-09T07:27:15Z) - Red Teaming Language Models to Reduce Harms: Methods, Scaling Behaviors,
and Lessons Learned [10.836210010868932]
We investigate scaling behaviors for red teaming across 3 model sizes (2.7B, 13B, and 52B parameters) and 4 model types.
We release our dataset of 38,961 red team attacks for others to analyze and learn from.
arXiv Detail & Related papers (2022-08-23T23:37:14Z) - Efficient Entity Candidate Generation for Low-Resource Languages [13.789451365205665]
Candidate generation is a crucial module in entity linking.
It plays a key role in multiple NLP tasks that have been proven to beneficially leverage knowledge bases.
This paper constitutes an in-depth analysis of the candidate generation problem in the context of cross-lingual entity linking.
arXiv Detail & Related papers (2022-06-30T09:49:53Z) - From Good to Best: Two-Stage Training for Cross-lingual Machine Reading
Comprehension [51.953428342923885]
We develop a two-stage approach to enhance the model performance.
The first stage targets at recall: we design a hard-learning (HL) algorithm to maximize the likelihood that the top-k predictions contain the accurate answer.
The second stage focuses on precision: an answer-aware contrastive learning mechanism is developed to learn the fine difference between the accurate answer and other candidates.
arXiv Detail & Related papers (2021-12-09T07:31:15Z) - NoiER: An Approach for Training more Reliable Fine-TunedDownstream Task
Models [54.184609286094044]
We propose noise entropy regularisation (NoiER) as an efficient learning paradigm that solves the problem without auxiliary models and additional data.
The proposed approach improved traditional OOD detection evaluation metrics by 55% on average compared to the original fine-tuned models.
arXiv Detail & Related papers (2021-08-29T06:58:28Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Unsupervised Neural Machine Translation for Low-Resource Domains via
Meta-Learning [27.86606560170401]
We present a novel meta-learning algorithm for unsupervised neural machine translation (UNMT)
We train the model to adapt to another domain by utilizing only a small amount of training data.
Our model surpasses a transfer learning-based approach by up to 2-4 BLEU scores.
arXiv Detail & Related papers (2020-10-18T17:54:13Z) - Pre-training Multilingual Neural Machine Translation by Leveraging
Alignment Information [72.2412707779571]
mRASP is an approach to pre-train a universal multilingual neural machine translation model.
We carry out experiments on 42 translation directions across a diverse setting, including low, medium, rich resource, and as well as transferring to exotic language pairs.
arXiv Detail & Related papers (2020-10-07T03:57:54Z) - Joint Contextual Modeling for ASR Correction and Language Understanding [60.230013453699975]
We propose multi-task neural approaches to perform contextual language correction on ASR outputs jointly with language understanding (LU)
We show that the error rates of off the shelf ASR and following LU systems can be reduced significantly by 14% relative with joint models trained using small amounts of in-domain data.
arXiv Detail & Related papers (2020-01-28T22:09:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.