Leveraging Knowledge in Multilingual Commonsense Reasoning
- URL: http://arxiv.org/abs/2110.08462v1
- Date: Sat, 16 Oct 2021 03:51:53 GMT
- Title: Leveraging Knowledge in Multilingual Commonsense Reasoning
- Authors: Yuwei Fang, Shuohang Wang, Yichong Xu, Ruochen Xu, Siqi Sun, Chenguang
Zhu, Michael Zeng
- Abstract summary: We propose to utilize English knowledge sources via a translate-retrieve-translate (TRT) strategy.
For multilingual commonsense questions and choices, we collect related knowledge via translation and retrieval from the knowledge sources.
The retrieved knowledge is then translated into the target language and integrated into a pre-trained multilingual language model.
- Score: 25.155987513306854
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Commonsense reasoning (CSR) requires the model to be equipped with general
world knowledge. While CSR is a language-agnostic process, most comprehensive
knowledge sources are in few popular languages, especially English. Thus, it
remains unclear how to effectively conduct multilingual commonsense reasoning
(XCSR) for various languages. In this work, we propose to utilize English
knowledge sources via a translate-retrieve-translate (TRT) strategy. For
multilingual commonsense questions and choices, we collect related knowledge
via translation and retrieval from the knowledge sources. The retrieved
knowledge is then translated into the target language and integrated into a
pre-trained multilingual language model via visible knowledge attention. Then
we utilize a diverse of 4 English knowledge sources to provide more
comprehensive coverage of knowledge in different formats. Extensive results on
the XCSR benchmark demonstrate that TRT with external knowledge can
significantly improve multilingual commonsense reasoning in both zero-shot and
translate-train settings, outperforming 3.3 and 3.6 points over the previous
state-of-the-art on XCSR benchmark datasets (X-CSQA and X-CODAH).
Related papers
- How and Where to Translate? The Impact of Translation Strategies in Cross-lingual LLM Prompting [15.388822834013599]
In multilingual retrieval-augmented generation (RAG)-based systems, knowledge bases (KB) are often shared from high-resource languages (such as English) to low-resource ones.<n>Two common practices are pre-translation to create a mono-lingual prompt and cross-lingual prompting for direct inference.<n>We show that an optimized prompting strategy can significantly improve knowledge sharing across languages, therefore improve the performance on the downstream classification task.
arXiv Detail & Related papers (2025-07-21T19:37:15Z) - Multilingual Information Retrieval with a Monolingual Knowledge Base [2.419638771866955]
We propose a novel strategy to fine-tune multilingual embedding models with weighted sampling for contrastive learning.<n>We demonstrate that the weighted sampling strategy produces performance gains compared to standard ones by up to 31.03% in MRR and up to 33.98% in Recall@3.
arXiv Detail & Related papers (2025-06-03T07:05:49Z) - Multilingual Retrieval-Augmented Generation for Knowledge-Intensive Task [73.35882908048423]
Retrieval-augmented generation (RAG) has become a cornerstone of contemporary NLP.
This paper investigates the effectiveness of RAG across multiple languages by proposing novel approaches for multilingual open-domain question-answering.
arXiv Detail & Related papers (2025-04-04T17:35:43Z) - Not All Languages are Equal: Insights into Multilingual Retrieval-Augmented Generation [38.631934251052485]
We evaluate six multilingual RALMs using our benchmark to explore the challenges of multilingual RALMs.
High-resource languages stand out in Monolingual Knowledge Extraction.
Indo-European languages lead RALMs to provide answers directly from documents.
English benefits from RALMs' selection bias and speaks louder in multilingual knowledge selection.
arXiv Detail & Related papers (2024-10-29T11:53:19Z) - Cross-Lingual Multi-Hop Knowledge Editing -- Benchmarks, Analysis and a Simple Contrastive Learning based Approach [53.028586843468915]
We propose the Cross-Lingual Multi-Hop Knowledge Editing paradigm, for measuring and analyzing the performance of various SoTA knowledge editing techniques in a cross-lingual setup.
Specifically, we create a parallel cross-lingual benchmark, CROLIN-MQUAKE for measuring the knowledge editing capabilities.
Following this, we propose a significantly improved system for cross-lingual multi-hop knowledge editing, CLEVER-CKE.
arXiv Detail & Related papers (2024-07-14T17:18:16Z) - Large Language Models Are Cross-Lingual Knowledge-Free Reasoners [43.99097308487008]
We decompose the process of reasoning tasks into two separated components: knowledge retrieval and knowledge-free reasoning.
We show that the knowledge-free reasoning capability can be nearly perfectly transferred across various source-target language directions.
We hypothesize that knowledge-free reasoning shares similar neurons in different languages for reasoning, while knowledge is stored separately in different languages.
arXiv Detail & Related papers (2024-06-24T14:03:04Z) - MLaKE: Multilingual Knowledge Editing Benchmark for Large Language Models [65.10456412127405]
MLaKE is a benchmark for the adaptability of knowledge editing methods across five languages.
MLaKE aggregates fact chains from Wikipedia across languages and generates questions in both free-form and multiple-choice.
We evaluate the multilingual knowledge editing generalization capabilities of existing methods on MLaKE.
arXiv Detail & Related papers (2024-04-07T15:23:28Z) - CLICKER: Attention-Based Cross-Lingual Commonsense Knowledge Transfer [5.375217612596619]
We propose the attention-based Cross-LIngual Commonsense Knowledge transfER framework.
CLICKER minimizes the performance gaps between English and non-English languages in commonsense question-answering tasks.
CLICKER achieves remarkable improvements in the cross-lingual task for languages other than English.
arXiv Detail & Related papers (2023-02-26T00:57:29Z) - Overcoming Language Disparity in Online Content Classification with
Multimodal Learning [22.73281502531998]
Large language models are now the standard to develop state-of-the-art solutions for text detection and classification tasks.
The development of advanced computational techniques and resources is disproportionately focused on the English language.
We explore the promise of incorporating the information contained in images via multimodal machine learning.
arXiv Detail & Related papers (2022-05-19T17:56:02Z) - Prix-LM: Pretraining for Multilingual Knowledge Base Construction [59.02868906044296]
We propose a unified framework, Prix-LM, for multilingual knowledge construction and completion.
We leverage two types of knowledge, monolingual triples and cross-lingual links, extracted from existing multilingual KBs.
Experiments on standard entity-related tasks, such as link prediction in multiple languages, cross-lingual entity linking and bilingual lexicon induction, demonstrate its effectiveness.
arXiv Detail & Related papers (2021-10-16T02:08:46Z) - Generated Knowledge Prompting for Commonsense Reasoning [53.88983683513114]
We propose generating knowledge statements directly from a language model with a generic prompt format.
This approach improves performance of both off-the-shelf and finetuned language models on four commonsense reasoning tasks.
Notably, we find that a model's predictions can improve when using its own generated knowledge.
arXiv Detail & Related papers (2021-10-15T21:58:03Z) - X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained
Language Models [103.75890012041366]
Language models (LMs) have proven surprisingly successful at capturing factual knowledge.
However, studies on LMs' factual representation ability have almost invariably been performed on English.
We create a benchmark of cloze-style probes for 23 typologically diverse languages.
arXiv Detail & Related papers (2020-10-13T05:29:56Z) - Unsupervised Commonsense Question Answering with Self-Talk [71.63983121558843]
We propose an unsupervised framework based on self-talk as a novel alternative to commonsense tasks.
Inspired by inquiry-based discovery learning, our approach inquires language models with a number of information seeking questions.
Empirical results demonstrate that the self-talk procedure substantially improves the performance of zero-shot language model baselines.
arXiv Detail & Related papers (2020-04-11T20:43:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.