NLAS-multi: A Multilingual Corpus of Automatically Generated Natural
Language Argumentation Schemes
- URL: http://arxiv.org/abs/2402.14458v1
- Date: Thu, 22 Feb 2024 11:31:50 GMT
- Title: NLAS-multi: A Multilingual Corpus of Automatically Generated Natural
Language Argumentation Schemes
- Authors: Ramon Ruiz-Dolz, Joaquin Taverner, John Lawrence and Chris Reed
- Abstract summary: We present an effective methodology for the automatic generation of natural language arguments in different topics and languages.
We also present a set of solid baselines and fine-tuned models for the automatic identification of argumentation schemes.
- Score: 4.015890309289342
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Some of the major limitations identified in the areas of argument mining,
argument generation, and natural language argument analysis are related to the
complexity of annotating argumentatively rich data, the limited size of these
corpora, and the constraints that represent the different languages and domains
in which these data is annotated. To address these limitations, in this paper
we present the following contributions: (i) an effective methodology for the
automatic generation of natural language arguments in different topics and
languages, (ii) the largest publicly available corpus of natural language
argumentation schemes, and (iii) a set of solid baselines and fine-tuned models
for the automatic identification of argumentation schemes.
Related papers
- Exploring the Potential of Large Language Models in Computational Argumentation [54.85665903448207]
Large language models (LLMs) have demonstrated impressive capabilities in understanding context and generating natural language.
This work aims to embark on an assessment of LLMs, such as ChatGPT, Flan models, and LLaMA2 models, in both zero-shot and few-shot settings.
arXiv Detail & Related papers (2023-11-15T15:12:15Z) - Soft Language Clustering for Multilingual Model Pre-training [57.18058739931463]
We propose XLM-P, which contextually retrieves prompts as flexible guidance for encoding instances conditionally.
Our XLM-P enables (1) lightweight modeling of language-invariant and language-specific knowledge across languages, and (2) easy integration with other multilingual pre-training methods.
arXiv Detail & Related papers (2023-06-13T08:08:08Z) - Using Large Language Models for (De-)Formalization and Natural Argumentation Exercises for Beginner's Students [0.0]
We describe two systems currently being developed that use large language models for the automatized correction of (i) exercises in translating back and forth between natural language and the languages of propositional logic and first-order predicate logic.
arXiv Detail & Related papers (2023-04-12T23:05:02Z) - Argument Mining using BERT and Self-Attention based Embeddings [1.6114012813668934]
In this paper, a novel methodology for argument mining is proposed.
It employs attention-based embeddings for link prediction to model the causational hierarchies in typical argument structures prevalent in online discourse.
arXiv Detail & Related papers (2023-02-27T15:52:31Z) - Can Transformers Reason in Fragments of Natural Language? [2.1485350418225244]
State-of-the-art deep-learning-based approaches to Natural Language Processing (NLP) are credited with various capabilities that involve reasoning with natural language texts.
We study the detection of formally valid inferences in controlled fragments of natural language for which the satisfiability problem becomes increasingly complex.
arXiv Detail & Related papers (2022-11-10T08:46:53Z) - CLSE: Corpus of Linguistically Significant Entities [58.29901964387952]
We release a Corpus of Linguistically Significant Entities (CLSE) annotated by experts.
CLSE covers 74 different semantic types to support various applications from airline ticketing to video games.
We create a linguistically representative NLG evaluation benchmark in three languages: French, Marathi, and Russian.
arXiv Detail & Related papers (2022-11-04T12:56:12Z) - Multilingual Generative Language Models for Zero-Shot Cross-Lingual
Event Argument Extraction [80.61458287741131]
We present a study on leveraging multilingual pre-trained generative language models for zero-shot cross-lingual event argument extraction (EAE)
By formulating EAE as a language generation task, our method effectively encodes event structures and captures the dependencies between arguments.
Our proposed model finetunes multilingual pre-trained generative language models to generate sentences that fill in the language-agnostic template with arguments extracted from the input passage.
arXiv Detail & Related papers (2022-03-15T23:00:32Z) - Characterizing the Program Expressive Power of Existential Rule
Languages [4.38078043834754]
Existential rule languages have been widely used in in-mediated query answering (OMQA)
The expressive power of representing domain knowledge for OMQA, known as the program expressive power, is not well-understood yet.
In this paper, we establish a number of novel characterizations for the program expressive power of several important existential rule languages.
arXiv Detail & Related papers (2021-12-15T14:08:38Z) - SML: a new Semantic Embedding Alignment Transformer for efficient
cross-lingual Natural Language Inference [71.57324258813674]
The ability of Transformers to perform with precision a variety of tasks such as question answering, Natural Language Inference (NLI) or summarising, have enable them to be ranked as one of the best paradigms to address this kind of tasks at present.
NLI is one of the best scenarios to test these architectures, due to the knowledge required to understand complex sentences and established a relation between a hypothesis and a premise.
In this paper, we propose a new architecture, siamese multilingual transformer, to efficiently align multilingual embeddings for Natural Language Inference.
arXiv Detail & Related papers (2021-03-17T13:23:53Z) - Limits of Detecting Text Generated by Large-Scale Language Models [65.46403462928319]
Some consider large-scale language models that can generate long and coherent pieces of text as dangerous, since they may be used in misinformation campaigns.
Here we formulate large-scale language model output detection as a hypothesis testing problem to classify text as genuine or generated.
arXiv Detail & Related papers (2020-02-09T19:53:23Z) - Model-theoretic Characterizations of Existential Rule Languages [9.845164265154832]
Existential rules, a.k.a. dependencies in databases, are a family of important logical languages widely used in computer science and artificial intelligence.
We establish model-theoretic characterizations for a number of existential rule languages such as (disjunctive) embedded dependencies,generating dependencies (TGDs), (frontier-)guarded TGDs and linear TGDs.
As a natural application of these characterizations, complexity bounds for the rewritability of above languages are also identified.
arXiv Detail & Related papers (2020-01-23T17:29:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.