What is Wrong with Language Models that Can Not Tell a Story?
- URL: http://arxiv.org/abs/2211.05044v2
- Date: Thu, 10 Nov 2022 16:14:35 GMT
- Title: What is Wrong with Language Models that Can Not Tell a Story?
- Authors: Ivan P. Yamshchikov and Alexey Tikhonov
- Abstract summary: This paper argues that a deeper understanding of narrative and the successful generation of longer subjectively interesting texts is a vital bottleneck that hinders the progress in modern Natural Language Processing (NLP)
We demonstrate that there are no adequate datasets, evaluation methods, and even operational concepts that could be used to start working on narrative processing.
- Score: 20.737171876839238
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper argues that a deeper understanding of narrative and the successful
generation of longer subjectively interesting texts is a vital bottleneck that
hinders the progress in modern Natural Language Processing (NLP) and may even
be in the whole field of Artificial Intelligence. We demonstrate that there are
no adequate datasets, evaluation methods, and even operational concepts that
could be used to start working on narrative processing.
Related papers
- Evolution of Natural Language Processing Technology: Not Just Language
Processing Towards General Purpose AI [0.0]
This report provides a technological explanation of how cutting-edge NLP has made it possible to realize the "practice makes perfect" principle.
Achievements exceeding the initial predictions have been reported from the results of learning vast amounts of textual data using deep learning.
It is an accurate example of the learner embodying the concept of "practice makes perfect" by using vast amounts of textual data.
arXiv Detail & Related papers (2023-10-10T00:41:38Z) - Putting Natural in Natural Language Processing [11.746833714322156]
The field of NLP has overwhelmingly focused on processing written rather than spoken language.
Recent advances in deep learning have led to a fortuitous convergence in methods between speech processing and mainstream NLP.
Truly natural language processing could lead to better integration with the rest of language science.
arXiv Detail & Related papers (2023-05-08T09:29:31Z) - Robust Preference Learning for Storytelling via Contrastive
Reinforcement Learning [53.92465205531759]
Controlled automated story generation seeks to generate natural language stories satisfying constraints from natural language critiques or preferences.
We train a contrastive bi-encoder model to align stories with human critiques, building a general purpose preference model.
We further fine-tune the contrastive reward model using a prompt-learning technique to increase story generation robustness.
arXiv Detail & Related papers (2022-10-14T13:21:33Z) - Transparency Helps Reveal When Language Models Learn Meaning [71.96920839263457]
Our systematic experiments with synthetic data reveal that, with languages where all expressions have context-independent denotations, both autoregressive and masked language models learn to emulate semantic relations between expressions.
Turning to natural language, our experiments with a specific phenomenon -- referential opacity -- add to the growing body of evidence that current language models do not well-represent natural language semantics.
arXiv Detail & Related papers (2022-10-14T02:35:19Z) - Visualizing and Explaining Language Models [0.0]
Natural Language Processing has become, after Computer Vision, the second field of Artificial Intelligence.
This paper showcases the techniques used in some of the most popular Deep Learning for NLP visualizations, with a special focus on interpretability and explainability.
arXiv Detail & Related papers (2022-04-30T17:23:33Z) - Testing the Ability of Language Models to Interpret Figurative Language [69.59943454934799]
Figurative and metaphorical language are commonplace in discourse.
It remains an open question to what extent modern language models can interpret nonliteral phrases.
We introduce Fig-QA, a Winograd-style nonliteral language understanding task.
arXiv Detail & Related papers (2022-04-26T23:42:22Z) - Cut the CARP: Fishing for zero-shot story evaluation [0.0]
Contrastive Authoring and Reviewing Pairing is a scalable, efficient method for performing superior, zero-shot evaluation of stories.
We show a strong correlation between human evaluation of stories and those of CARP.
We also present and analyze the Story-Critique dataset, a new corpora composed of 1.3 million aligned story-critique pairs derived from over 80,000 stories.
arXiv Detail & Related papers (2021-10-06T23:50:46Z) - Deep Learning for Text Style Transfer: A Survey [71.8870854396927]
Text style transfer is an important task in natural language generation, which aims to control certain attributes in the generated text.
We present a systematic survey of the research on neural text style transfer, spanning over 100 representative articles since the first neural text style transfer work in 2017.
We discuss the task formulation, existing datasets and subtasks, evaluation, as well as the rich methodologies in the presence of parallel and non-parallel data.
arXiv Detail & Related papers (2020-11-01T04:04:43Z) - Controlled Hallucinations: Learning to Generate Faithfully from Noisy
Data [1.0914300987810126]
We present a technique to treat such hallucinations as a controllable aspect of the generated text.
On the WikiBio corpus, a particularly noisy dataset, we demonstrate the efficacy of the technique both in an automatic and in a human evaluation.
arXiv Detail & Related papers (2020-10-12T17:25:02Z) - Knowledge-Aware Procedural Text Understanding with Multi-Stage Training [110.93934567725826]
We focus on the task of procedural text understanding, which aims to comprehend such documents and track entities' states and locations during a process.
Two challenges, the difficulty of commonsense reasoning and data insufficiency, still remain unsolved.
We propose a novel KnOwledge-Aware proceduraL text understAnding (KOALA) model, which effectively leverages multiple forms of external knowledge.
arXiv Detail & Related papers (2020-09-28T10:28:40Z) - A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation [98.25464306634758]
We propose to utilize commonsense knowledge from external knowledge bases to generate reasonable stories.
We employ multi-task learning which combines a discriminative objective to distinguish true and fake stories.
Our model can generate more reasonable stories than state-of-the-art baselines, particularly in terms of logic and global coherence.
arXiv Detail & Related papers (2020-01-15T05:42:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.