Unsupervised Learning of Explainable Parse Trees for Improved
Generalisation
- URL: http://arxiv.org/abs/2104.04998v1
- Date: Sun, 11 Apr 2021 12:10:03 GMT
- Title: Unsupervised Learning of Explainable Parse Trees for Improved
Generalisation
- Authors: Atul Sahay, Ayush Maheshwari, Ritesh Kumar, Ganesh Ramakrishnan,
Manjesh Kumar Hanawal, Kavi Arya
- Abstract summary: We propose an attention mechanism over Tree-LSTMs to learn more meaningful and explainable parse tree structures.
We also demonstrate the superior performance of our proposed model on natural language inference, semantic relatedness, and sentiment analysis tasks.
- Score: 15.576061447736057
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recursive neural networks (RvNN) have been shown useful for learning sentence
representations and helped achieve competitive performance on several natural
language inference tasks. However, recent RvNN-based models fail to learn
simple grammar and meaningful semantics in their intermediate tree
representation. In this work, we propose an attention mechanism over Tree-LSTMs
to learn more meaningful and explainable parse tree structures. We also
demonstrate the superior performance of our proposed model on natural language
inference, semantic relatedness, and sentiment analysis tasks and compare them
with other state-of-the-art RvNN based methods. Further, we present a detailed
qualitative and quantitative analysis of the learned parse trees and show that
the discovered linguistic structures are more explainable, semantically
meaningful, and grammatically correct than recent approaches. The source code
of the paper is available at
https://github.com/atul04/Explainable-Latent-Structures-Using-Attention.
Related papers
- Incorporating Constituent Syntax for Coreference Resolution [50.71868417008133]
We propose a graph-based method to incorporate constituent syntactic structures.
We also explore to utilise higher-order neighbourhood information to encode rich structures in constituent trees.
Experiments on the English and Chinese portions of OntoNotes 5.0 benchmark show that our proposed model either beats a strong baseline or achieves new state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T07:40:42Z) - Learning Gradual Argumentation Frameworks using Genetic Algorithms [5.953590600890214]
We propose a genetic algorithm to simultaneously learn the structure of argumentative classification models.
Our prototype learns argumentative classification models that are comparable to decision trees in terms of learning performance and interpretability.
arXiv Detail & Related papers (2021-06-25T12:33:31Z) - Learning compositional structures for semantic graph parsing [81.41592892863979]
We show how AM dependency parsing can be trained directly on a neural latent-variable model.
Our model picks up on several linguistic phenomena on its own and achieves comparable accuracy to supervised training.
arXiv Detail & Related papers (2021-06-08T14:20:07Z) - Can RNNs learn Recursive Nested Subject-Verb Agreements? [4.094098809740732]
Language processing requires the ability to extract nested tree structures.
Recent advances in Recurrent Neural Networks (RNNs) achieve near-human performance in some language tasks.
arXiv Detail & Related papers (2021-01-06T20:47:02Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Towards Interpretable Natural Language Understanding with Explanations
as Latent Variables [146.83882632854485]
We develop a framework for interpretable natural language understanding that requires only a small set of human annotated explanations for training.
Our framework treats natural language explanations as latent variables that model the underlying reasoning process of a neural model.
arXiv Detail & Related papers (2020-10-24T02:05:56Z) - Recursive Top-Down Production for Sentence Generation with Latent Trees [77.56794870399288]
We model the production property of context-free grammars for natural and synthetic languages.
We present a dynamic programming algorithm that marginalises over latent binary tree structures with $N$ leaves.
We also present experimental results on German-English translation on the Multi30k dataset.
arXiv Detail & Related papers (2020-10-09T17:47:16Z) - Exploiting Syntactic Structure for Better Language Modeling: A Syntactic
Distance Approach [78.77265671634454]
We make use of a multi-task objective, i.e., the models simultaneously predict words as well as ground truth parse trees in a form called "syntactic distances"
Experimental results on the Penn Treebank and Chinese Treebank datasets show that when ground truth parse trees are provided as additional training signals, the model is able to achieve lower perplexity and induce trees with better quality.
arXiv Detail & Related papers (2020-05-12T15:35:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.