A Benchmark of Rule-Based and Neural Coreference Resolution in Dutch
Novels and News
- URL: http://arxiv.org/abs/2011.01615v1
- Date: Tue, 3 Nov 2020 10:52:00 GMT
- Title: A Benchmark of Rule-Based and Neural Coreference Resolution in Dutch
Novels and News
- Authors: Corb\`en Poot, Andreas van Cranenburgh
- Abstract summary: The results provide insight into the relative strengths of data-driven and knowledge-driven systems.
The neural system performs best on news/Wikipedia text, while the rule-based system performs best on literature.
- Score: 4.695687634290403
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We evaluate a rule-based (Lee et al., 2013) and neural (Lee et al., 2018)
coreference system on Dutch datasets of two domains: literary novels and
news/Wikipedia text. The results provide insight into the relative strengths of
data-driven and knowledge-driven systems, as well as the influence of domain,
document length, and annotation schemes. The neural system performs best on
news/Wikipedia text, while the rule-based system performs best on literature.
The neural system shows weaknesses with limited training data and long
documents, while the rule-based system is affected by annotation differences.
The code and models used in this paper are available at
https://github.com/andreasvc/crac2020
Related papers
- Neuro-symbolic Rule Learning in Real-world Classification Tasks [75.0907310059298]
We extend pix2rule's neural DNF module to support rule learning in real-world multi-class and multi-label classification tasks.
We propose a novel extended model called neural DNF-EO (Exactly One) which enforces mutual exclusivity in multi-class classification.
arXiv Detail & Related papers (2023-03-29T13:27:14Z) - Like a Good Nearest Neighbor: Practical Content Moderation and Text
Classification [66.02091763340094]
Like a Good Nearest Neighbor (LaGoNN) is a modification to SetFit that introduces no learnable parameters but alters input text with information from its nearest neighbor.
LaGoNN is effective at flagging undesirable content and text classification, and improves the performance of SetFit.
arXiv Detail & Related papers (2023-02-17T15:43:29Z) - SLCNN: Sentence-Level Convolutional Neural Network for Text
Classification [0.0]
Convolutional neural network (CNN) has shown remarkable success in the task of text classification.
New baseline models have been studied for text classification using CNN.
Results have shown that the proposed models have better performance, particularly in the longer documents.
arXiv Detail & Related papers (2023-01-27T13:16:02Z) - Non-neural Models Matter: A Re-evaluation of Neural Referring Expression
Generation Systems [6.651864489482537]
In recent years, neural models have often outperformed rule-based and classic Machine Learning approaches in NLG.
We argue that they should not be overlooked, since, for some tasks, well-designed non-neural approaches achieve better performance than neural ones.
arXiv Detail & Related papers (2022-03-15T21:47:25Z) - LNN-EL: A Neuro-Symbolic Approach to Short-text Entity Linking [62.634516517844496]
We propose LNN-EL, a neuro-symbolic approach that combines the advantages of using interpretable rules with the performance of neural learning.
Even though constrained to using rules, LNN-EL performs competitively against SotA black-box neural approaches.
arXiv Detail & Related papers (2021-06-17T20:22:45Z) - NSL: Hybrid Interpretable Learning From Noisy Raw Data [66.15862011405882]
This paper introduces a hybrid neural-symbolic learning framework, called NSL, that learns interpretable rules from labelled unstructured data.
NSL combines pre-trained neural networks for feature extraction with FastLAS, a state-of-the-art ILP system for rule learning under the answer set semantics.
We demonstrate that NSL is able to learn robust rules from MNIST data and achieve comparable or superior accuracy when compared to neural network and random forest baselines.
arXiv Detail & Related papers (2020-12-09T13:02:44Z) - Neuro-Symbolic Representations for Video Captioning: A Case for
Leveraging Inductive Biases for Vision and Language [148.0843278195794]
We propose a new model architecture for learning multi-modal neuro-symbolic representations for video captioning.
Our approach uses a dictionary learning-based method of learning relations between videos and their paired text descriptions.
arXiv Detail & Related papers (2020-11-18T20:21:19Z) - Coreference Resolution System for Indonesian Text with Mention Pair
Method and Singleton Exclusion using Convolutional Neural Network [0.0]
We propose a new coreference resolution system for Indonesian text with mention pair method.
In addition to lexical and syntactic features, in order to learn the representation of the mentions words and context, we use word embeddings and feed them to CNN.
Our proposed system outperforms the state-of-the-art system.
arXiv Detail & Related papers (2020-09-11T22:21:19Z) - Learning Compositional Rules via Neural Program Synthesis [67.62112086708859]
We present a neuro-symbolic model which learns entire rule systems from a small set of examples.
Instead of directly predicting outputs from inputs, we train our model to induce the explicit system of rules governing a set of previously seen examples.
arXiv Detail & Related papers (2020-03-12T01:06:48Z) - Improving Neural Named Entity Recognition with Gazetteers [6.292153194561472]
This article describes how to generate gazetteers from the Wikidata knowledge graph as well as how to integrate the information into a neural NER system.
Experiments reveal that the approach yields performance gains in two distinct languages.
arXiv Detail & Related papers (2020-03-06T08:29:37Z) - SeMemNN: A Semantic Matrix-Based Memory Neural Network for Text
Classification [15.111940377403252]
We propose 5 different configurations for the semantic matrix-based memory neural network with end-to-end learning manner.
We evaluate our proposed method on two corpora of news articles (AG news, Sogou news)
arXiv Detail & Related papers (2020-03-04T02:00:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.