Non-neural Models Matter: A Re-evaluation of Neural Referring Expression
Generation Systems
- URL: http://arxiv.org/abs/2203.08274v1
- Date: Tue, 15 Mar 2022 21:47:25 GMT
- Title: Non-neural Models Matter: A Re-evaluation of Neural Referring Expression
Generation Systems
- Authors: Fahime Same, Guanyi Chen, Kees van Deemter
- Abstract summary: In recent years, neural models have often outperformed rule-based and classic Machine Learning approaches in NLG.
We argue that they should not be overlooked, since, for some tasks, well-designed non-neural approaches achieve better performance than neural ones.
- Score: 6.651864489482537
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, neural models have often outperformed rule-based and classic
Machine Learning approaches in NLG. These classic approaches are now often
disregarded, for example when new neural models are evaluated. We argue that
they should not be overlooked, since, for some tasks, well-designed non-neural
approaches achieve better performance than neural ones. In this paper, the task
of generating referring expressions in linguistic context is used as an
example. We examined two very different English datasets (WEBNLG and WSJ), and
evaluated each algorithm using both automatic and human evaluations. Overall,
the results of these evaluations suggest that rule-based systems with simple
rule sets achieve on-par or better performance on both datasets compared to
state-of-the-art neural REG systems. In the case of the more realistic dataset,
WSJ, a machine learning-based system with well-designed linguistic features
performed best. We hope that our work can encourage researchers to consider
non-neural models in future.
Related papers
- Revisiting N-Gram Models: Their Impact in Modern Neural Networks for Handwritten Text Recognition [4.059708117119894]
This study addresses whether explicit language models, specifically n-gram models, still contribute to the performance of state-of-the-art deep learning architectures in the field of handwriting recognition.
We evaluate two prominent neural network architectures, PyLaia and DAN, with and without the integration of explicit n-gram language models.
The results show that incorporating character or subword n-gram models significantly improves the performance of ATR models on all datasets.
arXiv Detail & Related papers (2024-04-30T07:37:48Z) - When Newer is Not Better: Does Deep Learning Really Benefit
Recommendation From Implicit Feedback? [34.04060791716633]
We compare recent neural recommendation models against traditional ones in top-n recommendation from implicit data.
Our work illuminates the relative advantages and disadvantages of neural models in recommendation.
arXiv Detail & Related papers (2023-05-02T22:03:49Z) - Mitigating Data Scarcity for Large Language Models [7.259279261659759]
In recent years, pretrained neural language models (PNLMs) have taken the field of natural language processing by storm.
Data scarcity are commonly found in specialized domains, such as medical, or in low-resource languages that are underexplored by AI research.
In this dissertation, we focus on mitigating data scarcity using data augmentation and neural ensemble learning techniques.
arXiv Detail & Related papers (2023-02-03T15:17:53Z) - Neural-Symbolic Recursive Machine for Systematic Generalization [113.22455566135757]
We introduce the Neural-Symbolic Recursive Machine (NSR), whose core is a Grounded Symbol System (GSS)
NSR integrates neural perception, syntactic parsing, and semantic reasoning.
We evaluate NSR's efficacy across four challenging benchmarks designed to probe systematic generalization capabilities.
arXiv Detail & Related papers (2022-10-04T13:27:38Z) - Dependency-based Mixture Language Models [53.152011258252315]
We introduce the Dependency-based Mixture Language Models.
In detail, we first train neural language models with a novel dependency modeling objective.
We then formulate the next-token probability by mixing the previous dependency modeling probability distributions with self-attention.
arXiv Detail & Related papers (2022-03-19T06:28:30Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z) - Learning from Mistakes: Combining Ontologies via Self-Training for
Dialogue Generation [6.221019624345408]
Natural language generators (NLGs) for task-oriented dialogue typically take a meaning representation (MR) as input.
We create a new, larger combined ontology, and then train an NLG to produce utterances covering it.
For example, if one dataset has attributes for family-friendly and rating information, and the other has attributes for decor and service, our aim is an NLG for the combined ontology that can produce utterances that realize values for family-friendly, rating, decor and service.
arXiv Detail & Related papers (2020-09-30T23:54:38Z) - DyNODE: Neural Ordinary Differential Equations for Dynamics Modeling in
Continuous Control [0.0]
We present a novel approach that captures the underlying dynamics of a system by incorporating control in a neural ordinary differential equation framework.
Results indicate that a simple DyNODE architecture when combined with an actor-critic reinforcement learning algorithm outperforms canonical neural networks.
arXiv Detail & Related papers (2020-09-09T12:56:58Z) - Learning Compositional Rules via Neural Program Synthesis [67.62112086708859]
We present a neuro-symbolic model which learns entire rule systems from a small set of examples.
Instead of directly predicting outputs from inputs, we train our model to induce the explicit system of rules governing a set of previously seen examples.
arXiv Detail & Related papers (2020-03-12T01:06:48Z) - Rethinking Generalization of Neural Models: A Named Entity Recognition
Case Study [81.11161697133095]
We take the NER task as a testbed to analyze the generalization behavior of existing models from different perspectives.
Experiments with in-depth analyses diagnose the bottleneck of existing neural NER models.
As a by-product of this paper, we have open-sourced a project that involves a comprehensive summary of recent NER papers.
arXiv Detail & Related papers (2020-01-12T04:33:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.