Towards Accurate Translation via Semantically Appropriate Application of
  Lexical Constraints
        - URL: http://arxiv.org/abs/2306.12089v1
 - Date: Wed, 21 Jun 2023 08:08:15 GMT
 - Title: Towards Accurate Translation via Semantically Appropriate Application of
  Lexical Constraints
 - Authors: Yujin Baek (1), Koanho Lee (1), Dayeon Ki (2), Hyoung-Gyu Lee (3),
  Cheonbok Park (3) and Jaegul Choo (1) ((1) KAIST, (2) Korea University, (3)
  Papago, Naver Corp.)
 - Abstract summary: We focus on two important but under-studied issues that lie in the current evaluation process of LNMT studies.
The model needs to cope with challenging lexical constraints that are "homographs" or "unseen" during training.
We propose PLUMCOT, which integrates contextually rich information about unseen lexical constraints from pre-trained language models.
We also release HOLLY, an evaluation benchmark for assessing the ability of a model to cope with "homographic" and "unseen" lexical constraints.
 - Score: 0.0
 - License: http://creativecommons.org/licenses/by/4.0/
 - Abstract:   Lexically-constrained NMT (LNMT) aims to incorporate user-provided
terminology into translations. Despite its practical advantages, existing work
has not evaluated LNMT models under challenging real-world conditions. In this
paper, we focus on two important but under-studied issues that lie in the
current evaluation process of LNMT studies. The model needs to cope with
challenging lexical constraints that are "homographs" or "unseen" during
training. To this end, we first design a homograph disambiguation module to
differentiate the meanings of homographs. Moreover, we propose PLUMCOT, which
integrates contextually rich information about unseen lexical constraints from
pre-trained language models and strengthens a copy mechanism of the pointer
network via direct supervision of a copying score. We also release HOLLY, an
evaluation benchmark for assessing the ability of a model to cope with
"homographic" and "unseen" lexical constraints. Experiments on HOLLY and the
previous test setup show the effectiveness of our method. The effects of
PLUMCOT are shown to be remarkable in "unseen" constraints. Our dataset is
available at https://github.com/papago-lab/HOLLY-benchmark
 
       
      
        Related papers
        - Model Utility Law: Evaluating LLMs beyond Performance through Mechanism   Interpretable Metric [99.56567010306807]
Large Language Models (LLMs) have become indispensable across academia, industry, and daily applications.<n>One core challenge of evaluation in the large language model (LLM) era is the generalization issue.<n>We propose Model Utilization Index (MUI), a mechanism interpretability enhanced metric that complements traditional performance scores.
arXiv  Detail & Related papers  (2025-04-10T04:09:47Z) - Refining Sentence Embedding Model through Ranking Sentences Generation   with Large Language Models [60.00178316095646]
Sentence embedding is essential for many NLP tasks, with contrastive learning methods achieving strong performance using datasets like NLI.
Recent studies leverage large language models (LLMs) to generate sentence pairs, reducing annotation dependency.
We propose a method for controlling the generation direction of LLMs in the latent space. Unlike unconstrained generation, the controlled approach ensures meaningful semantic divergence.
 Experiments on multiple benchmarks demonstrate that our method achieves new SOTA performance with a modest cost in ranking sentence synthesis.
arXiv  Detail & Related papers  (2025-02-19T12:07:53Z) - How Hard is this Test Set? NLI Characterization by Exploiting Training   Dynamics [49.9329723199239]
We propose a method for the automated creation of a challenging test set without relying on the manual construction of artificial and unrealistic examples.
We categorize the test set of popular NLI datasets into three difficulty levels by leveraging methods that exploit training dynamics.
When our characterization method is applied to the training set, models trained with only a fraction of the data achieve comparable performance to those trained on the full dataset.
arXiv  Detail & Related papers  (2024-10-04T13:39:21Z) - Pretraining Data Detection for Large Language Models: A Divergence-based   Calibration Method [108.56493934296687]
We introduce a divergence-based calibration method, inspired by the divergence-from-randomness concept, to calibrate token probabilities for pretraining data detection.
We have developed a Chinese-language benchmark, PatentMIA, to assess the performance of detection approaches for LLMs on Chinese text.
arXiv  Detail & Related papers  (2024-09-23T07:55:35Z) - ALMol: Aligned Language-Molecule Translation LLMs through Offline   Preference Contrastive Optimisation [2.296475290901356]
We focus on machine language-molecule translation and deploy a novel training approach called contrastive preference optimisation.
Our results demonstrate that our models achieve up to a 32% improvement compared to counterpart models.
arXiv  Detail & Related papers  (2024-05-14T13:59:24Z) - Contextual Spelling Correction with Language Model for Low-resource   Setting [0.0]
A small-scale word-based transformer LM is trained to provide the SC model with contextual understanding.
Probability of error happening(error model) is extracted from the corpus.
Combination of LM and error model is used to develop the SC model through the well-known noisy channel framework.
arXiv  Detail & Related papers  (2024-04-28T05:29:35Z) - Evaluating Generative Language Models in Information Extraction as   Subjective Question Correction [49.729908337372436]
We propose a new evaluation method, SQC-Score.
Inspired by the principles in subjective question correction, we propose a new evaluation method, SQC-Score.
Results on three information extraction tasks show that SQC-Score is more preferred by human annotators than the baseline metrics.
arXiv  Detail & Related papers  (2024-04-04T15:36:53Z) - Unsupervised Approach to Evaluate Sentence-Level Fluency: Do We Really
  Need Reference? [3.2528685897001455]
This paper adapts an existing unsupervised technique for measuring text fluency without the need for any reference.
Our approach leverages various word embeddings and trains language models using Recurrent Neural Network (RNN) architectures.
To assess the performance of the models, we conduct a comparative analysis across 10 Indic languages.
arXiv  Detail & Related papers  (2023-12-03T20:09:23Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
  Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv  Detail & Related papers  (2023-09-20T10:31:17Z) - Towards Fine-Grained Information: Identifying the Type and Location of
  Translation Errors [80.22825549235556]
Existing approaches can not synchronously consider error position and type.
We build an FG-TED model to predict the textbf addition and textbfomission errors.
 Experiments show that our model can identify both error type and position concurrently, and gives state-of-the-art results.
arXiv  Detail & Related papers  (2023-02-17T16:20:33Z) - Extract, Denoise, and Enforce: Evaluating and Predicting Lexical
  Constraints for Conditional Text Generation [31.341566859483056]
We present a systematic analysis of conditional generation to study whether current PLMs are good enough for preserving important concepts in the input.
We propose a framework for automatic constraint extraction, denoising, and enforcement that is shown to perform comparably or better than unconstrained generation.
arXiv  Detail & Related papers  (2021-04-18T05:29:02Z) - Understanding and Improving Lexical Choice in Non-Autoregressive
  Translation [98.11249019844281]
We propose to expose the raw data to NAT models to restore the useful information of low-frequency words.
Our approach pushes the SOTA NAT performance on the WMT14 English-German and WMT16 Romanian-English datasets up to 27.8 and 33.8 BLEU points, respectively.
arXiv  Detail & Related papers  (2020-12-29T03:18:50Z) - Self-Supervised Contrastive Learning for Unsupervised Phoneme
  Segmentation [37.054709598792165]
The model is a convolutional neural network that operates directly on the raw waveform.
It is optimized to identify spectral changes in the signal using the Noise-Contrastive Estimation principle.
At test time, a peak detection algorithm is applied over the model outputs to produce the final boundaries.
arXiv  Detail & Related papers  (2020-07-27T12:10:21Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.