A hybrid entity-centric approach to Persian pronoun resolution
- URL: http://arxiv.org/abs/2211.06257v1
- Date: Fri, 11 Nov 2022 14:59:58 GMT
- Title: A hybrid entity-centric approach to Persian pronoun resolution
- Authors: Hassan Haji Mohammadi, Alireza Talebpour, Ahmad Mahmoudi Aznaveh,
Samaneh Yazdani
- Abstract summary: This paper presents a hybrid model combining multiple rulebased sieves with a machine-learning sieve for pronouns.
For this purpose, seven high-precision rule-based sieves are designed for the Persian language.
The presented method demonstrates exemplary performance using pipeline design and combining the advantages of machine learning and rulebased methods.
- Score: 5.419608513284392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pronoun resolution is a challenging subset of an essential field in natural
language processing called coreference resolution. Coreference resolution is
about finding all entities in the text that refers to the same real-world
entity. This paper presents a hybrid model combining multiple rulebased sieves
with a machine-learning sieve for pronouns. For this purpose, seven
high-precision rule-based sieves are designed for the Persian language. Then, a
random forest classifier links pronouns to the previous partial clusters. The
presented method demonstrates exemplary performance using pipeline design and
combining the advantages of machine learning and rulebased methods. This method
has solved some challenges in end-to-end models. In this paper, the authors
develop a Persian coreference corpus called Mehr in the form of 400 documents.
This corpus fixes some weaknesses of the previous corpora in the Persian
language. Finally, the efficiency of the presented system compared to the
earlier model in Persian is reported by evaluating the proposed method on the
Mehr and Uppsala test sets.
Related papers
- Persian Pronoun Resolution: Leveraging Neural Networks and Language Models [8.604145658574689]
This study proposes the first end-to-end neural network system for Persian pronoun resolution, leveraging pre-trained Transformer models like ParsBERT.
Our system jointly optimize both mention detection and antecedent linking, achieving a 3.37 F1 score improvement over the previous state-of-the-art system.
arXiv Detail & Related papers (2024-05-17T11:56:00Z) - Language Model Decoding as Direct Metrics Optimization [87.68281625776282]
Current decoding methods struggle to generate texts that align with human texts across different aspects.
In this work, we frame decoding from a language model as an optimization problem with the goal of strictly matching the expected performance with human texts.
We prove that this induced distribution is guaranteed to improve the perplexity on human texts, which suggests a better approximation to the underlying distribution of human texts.
arXiv Detail & Related papers (2023-10-02T09:35:27Z) - One model to rule them all: ranking Slovene summarizers [0.0]
We propose a system that recommends the most suitable summarization model for a given text.
The proposed system employs a fully connected neural network that analyzes the input content.
We evaluate the proposed SloMetaSum model performance automatically and parts of it manually.
arXiv Detail & Related papers (2023-06-20T13:12:58Z) - Conjunct Resolution in the Face of Verbal Omissions [51.220650412095665]
We propose a conjunct resolution task that operates directly on the text and makes use of a split-and-rephrase paradigm in order to recover the missing elements in the coordination structure.
We curate a large dataset, containing over 10K examples of naturally-occurring verbal omissions with crowd-sourced annotations.
We train various neural baselines for this task, and show that while our best method obtains decent performance, it leaves ample space for improvement.
arXiv Detail & Related papers (2023-05-26T08:44:02Z) - A Sequence-to-Sequence Approach for Arabic Pronoun Resolution [0.0]
This paper proposes a sequence-to-sequence learning approach for Arabic pronoun resolution.
The proposed approach is evaluated on the AnATAr dataset.
arXiv Detail & Related papers (2023-05-19T08:53:41Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - Beyond Contrastive Learning: A Variational Generative Model for
Multilingual Retrieval [109.62363167257664]
We propose a generative model for learning multilingual text embeddings.
Our model operates on parallel data in $N$ languages.
We evaluate this method on a suite of tasks including semantic similarity, bitext mining, and cross-lingual question retrieval.
arXiv Detail & Related papers (2022-12-21T02:41:40Z) - ARMAN: Pre-training with Semantically Selecting and Reordering of
Sentences for Persian Abstractive Summarization [7.16879432974126]
We propose ARMAN, a Transformer-based encoder-decoder model pre-trained with three novel objectives to address this issue.
In ARMAN, salient sentences from a document are selected according to a modified semantic score to be masked and form a pseudo summary.
We show that our proposed model achieves state-of-the-art performance on all six summarization tasks measured by ROUGE and BERTScore.
arXiv Detail & Related papers (2021-09-09T08:35:39Z) - LexSubCon: Integrating Knowledge from Lexical Resources into Contextual
Embeddings for Lexical Substitution [76.615287796753]
We introduce LexSubCon, an end-to-end lexical substitution framework based on contextual embedding models.
This is achieved by combining contextual information with knowledge from structured lexical resources.
Our experiments show that LexSubCon outperforms previous state-of-the-art methods on LS07 and CoInCo benchmark datasets.
arXiv Detail & Related papers (2021-07-11T21:25:56Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.