Multilingual Irony Detection with Dependency Syntax and Neural Models
- URL: http://arxiv.org/abs/2011.05706v1
- Date: Wed, 11 Nov 2020 11:22:05 GMT
- Title: Multilingual Irony Detection with Dependency Syntax and Neural Models
- Authors: Alessandra Teresa Cignarella, Valerio Basile, Manuela Sanguinetti,
Cristina Bosco, Paolo Rosso and Farah Benamara
- Abstract summary: It focuses on the contribution from syntactic knowledge, exploiting linguistic resources where syntax is annotated according to the Universal Dependencies scheme.
The results suggest that fine-grained dependency-based syntactic information is informative for the detection of irony.
- Score: 61.32653485523036
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents an in-depth investigation of the effectiveness of
dependency-based syntactic features on the irony detection task in a
multilingual perspective (English, Spanish, French and Italian). It focuses on
the contribution from syntactic knowledge, exploiting linguistic resources
where syntax is annotated according to the Universal Dependencies scheme. Three
distinct experimental settings are provided. In the first, a variety of
syntactic dependency-based features combined with classical machine learning
classifiers are explored. In the second scenario, two well-known types of word
embeddings are trained on parsed data and tested against gold standard
datasets. In the third setting, dependency-based syntactic features are
combined into the Multilingual BERT architecture. The results suggest that
fine-grained dependency-based syntactic information is informative for the
detection of irony.
Related papers
- Exploring syntactic information in sentence embeddings through multilingual subject-verb agreement [1.4335183427838039]
We take the approach of developing curated synthetic data on a large scale, with specific properties.
We use a new multiple-choice task and datasets, Blackbird Language Matrices, to focus on a specific grammatical structural phenomenon.
We show that despite having been trained on multilingual texts in a consistent manner, multilingual pretrained language models have language-specific differences.
arXiv Detail & Related papers (2024-09-10T14:58:55Z) - Syntax and Semantics Meet in the "Middle": Probing the Syntax-Semantics
Interface of LMs Through Agentivity [68.8204255655161]
We present the semantic notion of agentivity as a case study for probing such interactions.
This suggests LMs may potentially serve as more useful tools for linguistic annotation, theory testing, and discovery.
arXiv Detail & Related papers (2023-05-29T16:24:01Z) - A Pilot Study on Dialogue-Level Dependency Parsing for Chinese [21.698966896156087]
We develop a high-quality human-annotated corpus, which contains 850 dialogues and 199,803 dependencies.
Considering that such tasks suffer from high annotation costs, we investigate zero-shot and few-shot scenarios.
Based on an existing syntactic treebank, we adopt a signal-based method to transform seen syntactic dependencies into unseen ones.
arXiv Detail & Related papers (2023-05-21T12:20:13Z) - Beyond Contrastive Learning: A Variational Generative Model for
Multilingual Retrieval [109.62363167257664]
We propose a generative model for learning multilingual text embeddings.
Our model operates on parallel data in $N$ languages.
We evaluate this method on a suite of tasks including semantic similarity, bitext mining, and cross-lingual question retrieval.
arXiv Detail & Related papers (2022-12-21T02:41:40Z) - Multi-grained Label Refinement Network with Dependency Structures for
Joint Intent Detection and Slot Filling [13.963083174197164]
intent and semantic components of a utterance are dependent on the syntactic elements of a sentence.
In this paper, we investigate a multi-grained label refinement network, which utilizes dependency structures and label semantic embeddings.
Considering to enhance syntactic representations, we introduce the dependency structures of sentences into our model by graph attention layer.
arXiv Detail & Related papers (2022-09-09T07:27:38Z) - Are Multilingual Models Effective in Code-Switching? [57.78477547424949]
We study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting.
Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching.
arXiv Detail & Related papers (2021-03-24T16:20:02Z) - GATE: Graph Attention Transformer Encoder for Cross-lingual Relation and
Event Extraction [107.8262586956778]
We introduce graph convolutional networks (GCNs) with universal dependency parses to learn language-agnostic sentence representations.
GCNs struggle to model words with long-range dependencies or are not directly connected in the dependency tree.
We propose to utilize the self-attention mechanism to learn the dependencies between words with different syntactic distances.
arXiv Detail & Related papers (2020-10-06T20:30:35Z) - Linguistic Typology Features from Text: Inferring the Sparse Features of
World Atlas of Language Structures [73.06435180872293]
We construct a recurrent neural network predictor based on byte embeddings and convolutional layers.
We show that some features from various linguistic types can be predicted reliably.
arXiv Detail & Related papers (2020-04-30T21:00:53Z) - Mutlitask Learning for Cross-Lingual Transfer of Semantic Dependencies [21.503766432869437]
We develop broad-coverage semantic dependencys for languages with no semantically annotated resource.
We leverage a multitask learning framework coupled with an annotation projection method.
We show that our best multitask model improves the labeled F1 score over the single-task baseline by 1.8 in the in-domain SemEval data.
arXiv Detail & Related papers (2020-04-30T17:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.