A Computational Evaluation Framework for Singable Lyric Translation
- URL: http://arxiv.org/abs/2308.13715v1
- Date: Sat, 26 Aug 2023 00:27:08 GMT
- Title: A Computational Evaluation Framework for Singable Lyric Translation
- Authors: Haven Kim, Kento Watanabe, Masataka Goto, Juhan Nam
- Abstract summary: We present a computational framework for the quantitative evaluation of singable lyric translation.
We measure syllable count distance, phoneme repetition similarity, musical structure distance, and semantic similarity.
Our framework seamlessly integrates musical, linguistic, and cultural dimensions of lyrics.
- Score: 17.492053233802135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Lyric translation plays a pivotal role in amplifying the global resonance of
music, bridging cultural divides, and fostering universal connections.
Translating lyrics, unlike conventional translation tasks, requires a delicate
balance between singability and semantics. In this paper, we present a
computational framework for the quantitative evaluation of singable lyric
translation, which seamlessly integrates musical, linguistic, and cultural
dimensions of lyrics. Our comprehensive framework consists of four metrics that
measure syllable count distance, phoneme repetition similarity, musical
structure distance, and semantic similarity. To substantiate the efficacy of
our framework, we collected a singable lyrics dataset, which precisely aligns
English, Japanese, and Korean lyrics on a line-by-line and section-by-section
basis, and conducted a comparative analysis between singable and non-singable
lyrics. Our multidisciplinary approach provides insights into the key
components that underlie the art of lyric translation and establishes a solid
groundwork for the future of computational lyric translation assessment.
Related papers
- Towards Estimating Personal Values in Song Lyrics [5.170818712089796]
Most music widely consumed in Western Countries contains song lyrics, with U.S. samples reporting almost all of their song libraries contain lyrics.
In this project, we take a perspectivist approach, guided by social science theory, to gathering annotations, estimating their quality, and aggregating them.
We then compare aggregated ratings to estimates based on pre-trained sentence/word embedding models by employing a validated value dictionary.
arXiv Detail & Related papers (2024-08-22T19:22:55Z) - Lyrics Transcription for Humans: A Readability-Aware Benchmark [1.2499537119440243]
We introduce Jam-ALT, a comprehensive lyrics transcription benchmark.
The benchmark features a complete revision of the JamendoLyrics dataset, along with evaluation metrics designed to capture and assess the lyric-specific nuances.
We apply the benchmark to recent transcription systems and present additional error analysis, as well as an experimental comparison with a classical music dataset.
arXiv Detail & Related papers (2024-07-30T14:20:09Z) - SongComposer: A Large Language Model for Lyric and Melody Composition in
Song Generation [88.33522730306674]
SongComposer could understand and generate melodies and lyrics in symbolic song representations.
We resort to symbolic song representation, the mature and efficient way humans designed for music.
With extensive experiments, SongComposer demonstrates superior performance in lyric-to-melody generation, melody-to-lyric generation, song continuation, and text-to-song creation.
arXiv Detail & Related papers (2024-02-27T16:15:28Z) - Bilingual Corpus Mining and Multistage Fine-Tuning for Improving Machine
Translation of Lecture Transcripts [50.00305136008848]
We propose a framework for parallel corpus mining, which provides a quick and effective way to mine a parallel corpus from publicly available lectures on Coursera.
For both English--Japanese and English--Chinese lecture translations, we extracted parallel corpora of approximately 50,000 lines and created development and test sets.
This study also suggests guidelines for gathering and cleaning corpora, mining parallel sentences, cleaning noise in the mined data, and creating high-quality evaluation splits.
arXiv Detail & Related papers (2023-11-07T03:50:25Z) - LyricSIM: A novel Dataset and Benchmark for Similarity Detection in
Spanish Song LyricS [52.77024349608834]
We present a new dataset and benchmark tailored to the task of semantic similarity in song lyrics.
Our dataset, originally consisting of 2775 pairs of Spanish songs, was annotated in a collective annotation experiment by 63 native annotators.
arXiv Detail & Related papers (2023-06-02T07:48:20Z) - Unsupervised Melody-to-Lyric Generation [91.29447272400826]
We propose a method for generating high-quality lyrics without training on any aligned melody-lyric data.
We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints.
Our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines.
arXiv Detail & Related papers (2023-05-30T17:20:25Z) - SenteCon: Leveraging Lexicons to Learn Human-Interpretable Language
Representations [51.08119762844217]
SenteCon is a method for introducing human interpretability in deep language representations.
We show that SenteCon provides high-level interpretability at little to no cost to predictive performance on downstream tasks.
arXiv Detail & Related papers (2023-05-24T05:06:28Z) - Unsupervised Melody-Guided Lyrics Generation [84.22469652275714]
We propose to generate pleasantly listenable lyrics without training on melody-lyric aligned data.
We leverage the crucial alignments between melody and lyrics and compile the given melody into constraints to guide the generation process.
arXiv Detail & Related papers (2023-05-12T20:57:20Z) - Translate the Beauty in Songs: Jointly Learning to Align Melody and
Translate Lyrics [38.35809268026605]
We propose Lyrics-Melody Translation with Adaptive Grouping (LTAG) as a holistic solution to automatic song translation.
It is a novel encoder-decoder framework that can simultaneously translate the source lyrics and determine the number of aligned notes at each decoding step.
Experiments conducted on an English-Chinese song translation data set show the effectiveness of our model in both automatic and human evaluation.
arXiv Detail & Related papers (2023-03-28T03:17:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.