Paraphrasing vs Coreferring: Two Sides of the Same Coin
- URL: http://arxiv.org/abs/2004.14979v2
- Date: Fri, 9 Oct 2020 16:48:36 GMT
- Title: Paraphrasing vs Coreferring: Two Sides of the Same Coin
- Authors: Yehudit Meged, Avi Caciularu, Vered Shwartz, Ido Dagan
- Abstract summary: We study the potential synergy between two different NLP tasks.
We use annotations from an event coreference dataset as distant supervision to re-scoreally-extracted predicate paraphrases.
We also use the same re-ranking features as additional inputs to a state-of-the-art event coreference resolution model.
- Score: 28.80553558538015
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the potential synergy between two different NLP tasks, both
confronting predicate lexical variability: identifying predicate paraphrases,
and event coreference resolution. First, we used annotations from an event
coreference dataset as distant supervision to re-score heuristically-extracted
predicate paraphrases. The new scoring gained more than 18 points in average
precision upon their ranking by the original scoring method. Then, we used the
same re-ranking features as additional inputs to a state-of-the-art event
coreference resolution model, which yielded modest but consistent improvements
to the model's performance. The results suggest a promising direction to
leverage data and models for each of the tasks to the benefit of the other.
Related papers
- SPLICE: A Singleton-Enhanced PipeLIne for Coreference REsolution [11.062090350704617]
Singleton mentions, i.e.entities mentioned only once in a text, are important to how humans understand discourse from a theoretical perspective.
Previous attempts to incorporate their detection in end-to-end neural coreference resolution for English have been hampered by the lack of singleton mention spans in the OntoNotes benchmark.
This paper addresses this limitation by combining predicted mentions from existing nested NER systems and features derived from OntoNotes syntax trees.
arXiv Detail & Related papers (2024-03-25T22:46:16Z) - Preserving Knowledge Invariance: Rethinking Robustness Evaluation of
Open Information Extraction [50.62245481416744]
We present the first benchmark that simulates the evaluation of open information extraction models in the real world.
We design and annotate a large-scale testbed in which each example is a knowledge-invariant clique.
By further elaborating the robustness metric, a model is judged to be robust if its performance is consistently accurate on the overall cliques.
arXiv Detail & Related papers (2023-05-23T12:05:09Z) - Mention Annotations Alone Enable Efficient Domain Adaptation for
Coreference Resolution [8.08448832546021]
We show that annotating mentions alone is nearly twice as fast as annotating full coreference chains.
Our approach facilitates annotation-efficient transfer and results in a 7-14% improvement in average F1 without increasing annotator time.
arXiv Detail & Related papers (2022-10-14T07:57:27Z) - Instance-Level Relative Saliency Ranking with Graph Reasoning [126.09138829920627]
We present a novel unified model to segment salient instances and infer relative saliency rank order.
A novel loss function is also proposed to effectively train the saliency ranking branch.
experimental results demonstrate that our proposed model is more effective than previous methods.
arXiv Detail & Related papers (2021-07-08T13:10:42Z) - NLP-CIC @ DIACR-Ita: POS and Neighbor Based Distributional Models for
Lexical Semantic Change in Diachronic Italian Corpora [62.997667081978825]
We present our systems and findings on unsupervised lexical semantic change for the Italian language.
The task is to determine whether a target word has evolved its meaning with time, only relying on raw-text from two time-specific datasets.
We propose two models representing the target words across the periods to predict the changing words using threshold and voting schemes.
arXiv Detail & Related papers (2020-11-07T11:27:18Z) - Keyphrase Extraction with Dynamic Graph Convolutional Networks and
Diversified Inference [50.768682650658384]
Keyphrase extraction (KE) aims to summarize a set of phrases that accurately express a concept or a topic covered in a given document.
Recent Sequence-to-Sequence (Seq2Seq) based generative framework is widely used in KE task, and it has obtained competitive performance on various benchmarks.
In this paper, we propose to adopt the Dynamic Graph Convolutional Networks (DGCN) to solve the above two problems simultaneously.
arXiv Detail & Related papers (2020-10-24T08:11:23Z) - Pointwise Paraphrase Appraisal is Potentially Problematic [21.06607915149245]
We show that the standard way of fine-tuning BERT for paraphrase identification by pairing two sentences as one sequence results in a model with state-of-the-art performance.
We also show that these models may even predict a pair of randomly-selected sentences with higher paraphrase score than a pair of identical ones.
arXiv Detail & Related papers (2020-05-25T09:27:31Z) - Active Learning for Coreference Resolution using Discrete Annotation [76.36423696634584]
We improve upon pairwise annotation for active learning in coreference resolution.
We ask annotators to identify mention antecedents if a presented mention pair is deemed not coreferent.
In experiments with existing benchmark coreference datasets, we show that the signal from this additional question leads to significant performance gains per human-annotation hour.
arXiv Detail & Related papers (2020-04-28T17:17:11Z) - Pseudo-Convolutional Policy Gradient for Sequence-to-Sequence
Lip-Reading [96.48553941812366]
Lip-reading aims to infer the speech content from the lip movement sequence.
Traditional learning process of seq2seq models suffers from two problems.
We propose a novel pseudo-convolutional policy gradient (PCPG) based method to address these two problems.
arXiv Detail & Related papers (2020-03-09T09:12:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.