Combining Improvements for Exploiting Dependency Trees in Neural
Semantic Parsing
- URL: http://arxiv.org/abs/2112.13179v1
- Date: Sat, 25 Dec 2021 03:41:42 GMT
- Title: Combining Improvements for Exploiting Dependency Trees in Neural
Semantic Parsing
- Authors: Defeng Xie and Jianmin Ji and Jiafei Xu and Ran Ji
- Abstract summary: In this paper, we examine three methods to incorporate such dependency information in a Transformer based semantic parsing system.
We first replace standard self-attention heads in the encoder with parent-scaled self-attention (PASCAL) heads.
Later, we insert the constituent attention (CA) to the encoder, which adds an extra constraint to attention heads that can better capture the inherent dependency structure of input sentences.
- Score: 1.0437764544103274
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The dependency tree of a natural language sentence can capture the
interactions between semantics and words. However, it is unclear whether those
methods which exploit such dependency information for semantic parsing can be
combined to achieve further improvement and the relationship of those methods
when they combine. In this paper, we examine three methods to incorporate such
dependency information in a Transformer based semantic parser and empirically
study their combinations. We first replace standard self-attention heads in the
encoder with parent-scaled self-attention (PASCAL) heads, i.e., the ones that
can attend to the dependency parent of each token. Then we concatenate
syntax-aware word representations (SAWRs), i.e., the intermediate hidden
representations of a neural dependency parser, with ordinary word embedding to
enhance the encoder. Later, we insert the constituent attention (CA) module to
the encoder, which adds an extra constraint to attention heads that can better
capture the inherent dependency structure of input sentences. Transductive
ensemble learning (TEL) is used for model aggregation, and an ablation study is
conducted to show the contribution of each method. Our experiments show that CA
is complementary to PASCAL or SAWRs, and PASCAL + CA provides state-of-the-art
performance among neural approaches on ATIS, GEO, and JOBS.
Related papers
- Dynamic Syntax Mapping: A New Approach to Unsupervised Syntax Parsing [0.0]
This study investigates the premise that language models, specifically their attention distributions, can encapsulate syntactic dependencies.
We introduce Dynamic Syntax Mapping (DSM), an innovative approach for the induction of these structures.
Our findings reveal that the use of an increasing array of substitutions notably enhances parsing precision on natural language data.
arXiv Detail & Related papers (2023-12-18T10:34:29Z) - mCL-NER: Cross-Lingual Named Entity Recognition via Multi-view
Contrastive Learning [54.523172171533645]
Cross-lingual named entity recognition (CrossNER) faces challenges stemming from uneven performance due to the scarcity of multilingual corpora.
We propose Multi-view Contrastive Learning for Cross-lingual Named Entity Recognition (mCL-NER)
Our experiments on the XTREME benchmark, spanning 40 languages, demonstrate the superiority of mCL-NER over prior data-driven and model-based approaches.
arXiv Detail & Related papers (2023-08-17T16:02:29Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - Syntactic Structure Processing in the Brain while Listening [3.735055636181383]
There are two popular syntactic parsing methods: constituency and dependency parsing.
Recent works have used syntactic embeddings based on constituency trees, incremental top-down parsing, and other word syntactic features for brain activity prediction given the text stimuli to study how the syntax structure is represented in the brain's language network.
We investigate the predictive power of the brain encoding models in three settings: (i) individual performance of the constituency and dependency syntactic parsing based embedding methods, (ii) efficacy of these syntactic parsing based embedding methods when controlling for basic syntactic signals, and (iii) relative effectiveness of each of the syntactic embedding methods when controlling for
arXiv Detail & Related papers (2023-02-16T21:28:11Z) - FECANet: Boosting Few-Shot Semantic Segmentation with Feature-Enhanced
Context-Aware Network [48.912196729711624]
Few-shot semantic segmentation is the task of learning to locate each pixel of a novel class in a query image with only a few annotated support images.
We propose a Feature-Enhanced Context-Aware Network (FECANet) to suppress the matching noise caused by inter-class local similarity.
In addition, we propose a novel correlation reconstruction module that encodes extra correspondence relations between foreground and background and multi-scale context semantic features.
arXiv Detail & Related papers (2023-01-19T16:31:13Z) - Integrating Dependency Tree Into Self-attention for Sentence
Representation [9.676884071651205]
We propose Dependency-Transformer, which applies a relation-attention mechanism that works in concert with the self-attention mechanism.
By a score-based method, we successfully inject the syntax information without affecting Transformer's parallelizability.
Our model outperforms or is comparable to the state-of-the-art methods on four tasks for sentence representation.
arXiv Detail & Related papers (2022-03-11T13:44:41Z) - CogAlign: Learning to Align Textual Neural Representations to Cognitive
Language Processing Signals [60.921888445317705]
We propose a CogAlign approach to integrate cognitive language processing signals into natural language processing models.
We show that CogAlign achieves significant improvements with multiple cognitive features over state-of-the-art models on public datasets.
arXiv Detail & Related papers (2021-06-10T07:10:25Z) - Dependency Parsing based Semantic Representation Learning with Graph
Neural Network for Enhancing Expressiveness of Text-to-Speech [49.05471750563229]
We propose a semantic representation learning method based on graph neural network, considering dependency relations of a sentence.
We show that our proposed method outperforms the baseline using vanilla BERT features both in LJSpeech and Bilzzard Challenge 2013 datasets.
arXiv Detail & Related papers (2021-04-14T13:09:51Z) - Syntactic representation learning for neural network based TTS with
syntactic parse tree traversal [49.05471750563229]
We propose a syntactic representation learning method based on syntactic parse tree to automatically utilize the syntactic structure information.
Experimental results demonstrate the effectiveness of our proposed approach.
For sentences with multiple syntactic parse trees, prosodic differences can be clearly perceived from the synthesized speeches.
arXiv Detail & Related papers (2020-12-13T05:52:07Z) - Comparison by Conversion: Reverse-Engineering UCCA from Syntax and
Lexical Semantics [29.971739294416714]
Building robust natural language understanding systems will require a clear characterization of whether and how various linguistic meaning representations complement each other.
We evaluate the mapping between meaning representations from different frameworks using two complementary methods: (i) a rule-based converter, and (ii) a supervised delexicalized that parses to one framework using only information from the other as features.
arXiv Detail & Related papers (2020-11-02T09:03:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.