GeoMLAMA: Geo-Diverse Commonsense Probing on Multilingual Pre-Trained
Language Models
- URL: http://arxiv.org/abs/2205.12247v1
- Date: Tue, 24 May 2022 17:54:50 GMT
- Title: GeoMLAMA: Geo-Diverse Commonsense Probing on Multilingual Pre-Trained
Language Models
- Authors: Da Yin, Hritik Bansal, Masoud Monajatipoor, Liunian Harold Li, Kai-Wei
Chang
- Abstract summary: We introduce a framework for geo-diverse commonsense probing on multilingual Language Models (mPLMs)
We benchmark 11 standard mPLMs which include variants of mBERT, XLM, mT5, and XGLM on GeoMLAMA dataset.
We find that 1) larger mPLM variants do not necessarily store geo-diverse concepts better than its smaller variant; 2) mPLMs are not intrinsically biased towards knowledge from the Western countries; and 3) a language may better probe knowledge about a non-native country than its native country.
- Score: 68.50584946761813
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent work has shown that Pre-trained Language Models (PLMs) have the
ability to store the relational knowledge from pre-training data in their model
parameters. However, it is not clear up to what extent do PLMs store
geo-diverse commonsense knowledge, the knowledge associated with a culture and
only shared locally. For instance, the color of bridal dress is white in
American weddings whereas it is red in Chinese weddings. Here, we wish to probe
if PLMs can predict red and white as the color of the bridal dress when queried
for American and Chinese weddings, respectively. To this end, we introduce a
framework for geo-diverse commonsense probing on multilingual PLMs (mPLMs) and
introduce a corresponding benchmark Geo-diverse Commonsense Multilingual
Language Model Analysis (GeoMLAMA) dataset. GeoMLAMA contains 3125 prompts in
English, Chinese, Hindi, Persian, and Swahili, with a wide coverage of concepts
shared by people from American, Chinese, Indian, Iranian and Kenyan cultures.
We benchmark 11 standard mPLMs which include variants of mBERT, XLM, mT5, and
XGLM on GeoMLAMA. Interestingly, we find that 1) larger mPLM variants do not
necessarily store geo-diverse concepts better than its smaller variant; 2)
mPLMs are not intrinsically biased towards knowledge from the Western countries
(the United States); 3) the native language of a country may not be the best
language to probe its knowledge and 4) a language may better probe knowledge
about a non-native country than its native country.
Related papers
- Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Models [62.91524967852552]
Large language models (LLMs) are typically multilingual due to pretraining on diverse multilingual corpora.
But can these models relate corresponding concepts across languages, effectively being crosslingual?
This study evaluates six state-of-the-art LLMs on inherently crosslingual tasks.
arXiv Detail & Related papers (2024-06-23T15:15:17Z) - OMGEval: An Open Multilingual Generative Evaluation Benchmark for Large
Language Models [59.54423478596468]
We introduce OMGEval, the first Open-source Multilingual Generative test set that can assess the capability of LLMs in different languages.
For each language, OMGEval provides 804 open-ended questions, covering a wide range of important capabilities of LLMs.
Specifically, the current version of OMGEval includes 5 languages (i.e., Zh, Ru, Fr, Es, Ar)
arXiv Detail & Related papers (2024-02-21T04:42:41Z) - GeoLM: Empowering Language Models for Geospatially Grounded Language
Understanding [45.36562604939258]
This paper introduces GeoLM, a language model that enhances the understanding of geo-entities in natural language.
We demonstrate that GeoLM exhibits promising capabilities in supporting toponym recognition, toponym linking, relation extraction, and geo-entity typing.
arXiv Detail & Related papers (2023-10-23T01:20:01Z) - GeoLLM: Extracting Geospatial Knowledge from Large Language Models [49.20315582673223]
We present GeoLLM, a novel method that can effectively extract geospatial knowledge from large language models.
We demonstrate the utility of our approach across multiple tasks of central interest to the international community, including the measurement of population density and economic livelihoods.
Our experiments reveal that LLMs are remarkably sample-efficient, rich in geospatial information, and robust across the globe.
arXiv Detail & Related papers (2023-10-10T00:03:23Z) - This Land is {Your, My} Land: Evaluating Geopolitical Biases in Language Models [40.61046400448044]
We show that large language models (LLM) recall certain geographical knowledge inconsistently when queried in different languages.
As a targeted case study, we consider territorial disputes, an inherently controversial and multilingual task.
We propose a suite of evaluation metrics to precisely quantify bias and consistency in responses across different languages.
arXiv Detail & Related papers (2023-05-24T01:16:17Z) - Language Models are Multilingual Chain-of-Thought Reasoners [83.37148309771378]
We introduce the Multilingual Grade School Math benchmark, by manually translating 250 grade-school math problems into ten typologically diverse languages.
We find that the ability to solve MGSM problems via chain-of-thought prompting emerges with increasing model scale.
We show that the multilingual reasoning abilities of language models extend to other tasks such as commonsense reasoning and word-in-context semantic judgment.
arXiv Detail & Related papers (2022-10-06T17:03:34Z) - Geographic Adaptation of Pretrained Language Models [29.81557992080902]
We introduce geoadaptation, an intermediate training step that couples language modeling with geolocation prediction in a multi-task learning setup.
We show that the effectiveness of geoadaptation stems from its ability to geographically retrofit the representation space of the pretrained language models.
arXiv Detail & Related papers (2022-03-16T11:55:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.