NorQuAD: Norwegian Question Answering Dataset
- URL: http://arxiv.org/abs/2305.01957v1
- Date: Wed, 3 May 2023 08:17:07 GMT
- Title: NorQuAD: Norwegian Question Answering Dataset
- Authors: Sardana Ivanova, Fredrik Aas Andreassen, Matias Jentoft, Sondre Wold,
Lilja {\O}vrelid
- Abstract summary: The dataset consists of 4,752 manually created question-answer pairs.
We benchmark several multilingual and Norwegian monolingual language models on the dataset and compare them against human performance.
The dataset will be made freely available.
- Score: 0.03281128493853064
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper we present NorQuAD: the first Norwegian question answering
dataset for machine reading comprehension. The dataset consists of 4,752
manually created question-answer pairs. We here detail the data collection
procedure and present statistics of the dataset. We also benchmark several
multilingual and Norwegian monolingual language models on the dataset and
compare them against human performance. The dataset will be made freely
available.
Related papers
- A Collection of Question Answering Datasets for Norwegian [6.149436325733799]
The data covers a wide range of skills and knowledge domains, including world knowledge, commonsense reasoning, truthfulness, and knowledge about Norway.
Our datasets comprise over 10k question-answer pairs, created by native speakers.
Most LMs perform better in Bokmaal than Nynorsk, struggle most with commonsense reasoning, and are often untruthful in generating answers to questions.
arXiv Detail & Related papers (2025-01-19T17:42:48Z) - Benchmarking Abstractive Summarisation: A Dataset of Human-authored Summaries of Norwegian News Articles [8.083472758337559]
We introduce a dataset of high-quality human-authored summaries of news articles in Norwegian.
The dataset is intended for benchmarking the abstractive summarisation capabilities of generative language models.
arXiv Detail & Related papers (2025-01-13T22:08:29Z) - Aya Dataset: An Open-Access Collection for Multilingual Instruction
Tuning [49.79783940841352]
Existing datasets are almost all in the English language.
We work with fluent speakers of languages from around the world to collect natural instances of instructions and completions.
We create the most extensive multilingual collection to date, comprising 513 million instances through templating and translating existing datasets across 114 languages.
arXiv Detail & Related papers (2024-02-09T18:51:49Z) - NLEBench+NorGLM: A Comprehensive Empirical Analysis and Benchmark Dataset for Generative Language Models in Norwegian [4.062031248854444]
Norwegian, spoken by only 5 million population, is under-representative within the most impressive breakthroughs in NLP tasks.
To fill this gap, we compiled the existing Norwegian dataset and pre-trained 4 Norwegian Open Language Models.
We find that the mainstream, English-dominated LM GPT-3.5 has limited capability in understanding the Norwegian context.
arXiv Detail & Related papers (2023-12-03T08:09:45Z) - The Belebele Benchmark: a Parallel Reading Comprehension Dataset in 122 Language Variants [80.4837840962273]
We present Belebele, a dataset spanning 122 language variants.
This dataset enables the evaluation of text models in high-, medium-, and low-resource languages.
arXiv Detail & Related papers (2023-08-31T17:43:08Z) - The Nordic Pile: A 1.2TB Nordic Dataset for Language Modeling [5.687459576800633]
We curate a high-quality dataset consisting of 1.2TB of text in all of the major North Germanic languages.
This paper details our considerations and processes for collecting, cleaning, and filtering the dataset.
arXiv Detail & Related papers (2023-03-30T06:42:22Z) - Neural Label Search for Zero-Shot Multi-Lingual Extractive Summarization [80.94424037751243]
In zero-shot multilingual extractive text summarization, a model is typically trained on English dataset and then applied on summarization datasets of other languages.
We propose NLS (Neural Label Search for Summarization), which jointly learns hierarchical weights for different sets of labels together with our summarization model.
We conduct multilingual zero-shot summarization experiments on MLSUM and WikiLingua datasets, and we achieve state-of-the-art results using both human and automatic evaluations.
arXiv Detail & Related papers (2022-04-28T14:02:16Z) - NorDiaChange: Diachronic Semantic Change Dataset for Norwegian [63.65426535861836]
NorDiaChange is the first diachronic semantic change dataset for Norwegian.
It covers about 80 Norwegian nouns manually annotated with graded semantic change over time.
arXiv Detail & Related papers (2022-01-13T18:27:33Z) - MTVR: Multilingual Moment Retrieval in Videos [89.24431389933703]
We introduce mTVR, a large-scale multilingual video moment retrieval dataset, containing 218K English and Chinese queries from 21.8K TV show video clips.
The dataset is collected by extending the popular TVR dataset (in English) with paired Chinese queries and subtitles.
We propose mXML, a multilingual moment retrieval model that learns and operates on data from both languages.
arXiv Detail & Related papers (2021-07-30T20:01:03Z) - Large-Scale Contextualised Language Modelling for Norwegian [7.5722195869569]
This paper introduces the first large-scale monolingual language models for Norwegian, based on both the ELMo and BERT frameworks.
In addition to detailing the training process, we present contrastive benchmark results on a suite of NLP tasks for Norwegian.
arXiv Detail & Related papers (2021-04-13T23:18:04Z) - TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data [113.29476656550342]
We present TaBERT, a pretrained LM that jointly learns representations for NL sentences and tables.
TaBERT is trained on a large corpus of 26 million tables and their English contexts.
Implementation of the model will be available at http://fburl.com/TaBERT.
arXiv Detail & Related papers (2020-05-17T17:26:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.