Context-Aware Parse Trees
- URL: http://arxiv.org/abs/2003.11118v1
- Date: Tue, 24 Mar 2020 21:19:14 GMT
- Title: Context-Aware Parse Trees
- Authors: Fangke Ye, Shengtian Zhou, Anand Venkat, Ryan Marcus, Paul Petersen,
Jesmin Jahan Tithi, Tim Mattson, Tim Kraska, Pradeep Dubey, Vivek Sarkar,
Justin Gottschlich
- Abstract summary: We present a new tree structure, heavily influenced by Aroma's SPT, called a emphcontext-aware parse tree (CAPT)
CAPT enhances SPT by providing a richer level of semantic representation.
Our research quantitatively demonstrates the value of our proposed semantically-salient features, enabling a specific CAPT configuration to be 39% more accurate than SPT across the 48,610 programs we analyzed.
- Score: 18.77504064534521
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The simplified parse tree (SPT) presented in Aroma, a state-of-the-art code
recommendation system, is a tree-structured representation used to infer code
semantics by capturing program \emph{structure} rather than program
\emph{syntax}. This is a departure from the classical abstract syntax tree,
which is principally driven by programming language syntax. While we believe a
semantics-driven representation is desirable, the specifics of an SPT's
construction can impact its performance. We analyze these nuances and present a
new tree structure, heavily influenced by Aroma's SPT, called a
\emph{context-aware parse tree} (CAPT). CAPT enhances SPT by providing a richer
level of semantic representation. Specifically, CAPT provides additional
binding support for language-specific techniques for adding
semantically-salient features, and language-agnostic techniques for removing
syntactically-present but semantically-irrelevant features. Our research
quantitatively demonstrates the value of our proposed semantically-salient
features, enabling a specific CAPT configuration to be 39\% more accurate than
SPT across the 48,610 programs we analyzed.
Related papers
- LATex: Leveraging Attribute-based Text Knowledge for Aerial-Ground Person Re-Identification [63.07563443280147]
We propose a novel framework named LATex for AG-ReID.
It adopts prompt-tuning strategies to leverage attribute-based text knowledge.
Our framework can fully leverage attribute-based text knowledge to improve the AG-ReID.
arXiv Detail & Related papers (2025-03-31T04:47:05Z) - Textless Dependency Parsing by Labeled Sequence Prediction [18.32371054754222]
"textless" methods process speech representations without automatic speech recognition systems.
Our proposed method predicts a dependency tree from a speech signal without transcribing, representing the tree as a labeled sequence.
Our findings highlight the importance of fusing word-level representations and sentence-level prosody for enhanced parsing performance.
arXiv Detail & Related papers (2024-07-14T08:38:14Z) - A General and Flexible Multi-concept Parsing Framework for Multilingual Semantic Matching [60.51839859852572]
We propose to resolve the text into multi concepts for multilingual semantic matching to liberate the model from the reliance on NER models.
We conduct comprehensive experiments on English datasets QQP and MRPC, and Chinese dataset Medical-SM.
arXiv Detail & Related papers (2024-03-05T13:55:16Z) - Incorporating Constituent Syntax for Coreference Resolution [50.71868417008133]
We propose a graph-based method to incorporate constituent syntactic structures.
We also explore to utilise higher-order neighbourhood information to encode rich structures in constituent trees.
Experiments on the English and Chinese portions of OntoNotes 5.0 benchmark show that our proposed model either beats a strong baseline or achieves new state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T07:40:42Z) - Self-Attentive Constituency Parsing for UCCA-based Semantic Parsing [0.0]
Graph-based representation is one of the semantic representation approaches to express the semantic structure of a text.
In this paper, we focus primarily on UCCA graph-based semantic representation.
We present the results for both single-lingual and cross-lingual tasks using zero-shot and few-shot learning for low-resource languages.
arXiv Detail & Related papers (2021-10-01T19:10:18Z) - KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization
for Relation Extraction [111.74812895391672]
We propose a Knowledge-aware Prompt-tuning approach with synergistic optimization (KnowPrompt)
We inject latent knowledge contained in relation labels into prompt construction with learnable virtual type words and answer words.
arXiv Detail & Related papers (2021-04-15T17:57:43Z) - Dependency Parsing based Semantic Representation Learning with Graph
Neural Network for Enhancing Expressiveness of Text-to-Speech [49.05471750563229]
We propose a semantic representation learning method based on graph neural network, considering dependency relations of a sentence.
We show that our proposed method outperforms the baseline using vanilla BERT features both in LJSpeech and Bilzzard Challenge 2013 datasets.
arXiv Detail & Related papers (2021-04-14T13:09:51Z) - Syntactic representation learning for neural network based TTS with
syntactic parse tree traversal [49.05471750563229]
We propose a syntactic representation learning method based on syntactic parse tree to automatically utilize the syntactic structure information.
Experimental results demonstrate the effectiveness of our proposed approach.
For sentences with multiple syntactic parse trees, prosodic differences can be clearly perceived from the synthesized speeches.
arXiv Detail & Related papers (2020-12-13T05:52:07Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Generating Synthetic Data for Task-Oriented Semantic Parsing with
Hierarchical Representations [0.8203855808943658]
In this work, we explore the possibility of generating synthetic data for neural semantic parsing.
Specifically, we first extract masked templates from the existing labeled utterances, and then fine-tune BART to generate synthetic utterances conditioning.
We show the potential of our approach when evaluating on the Facebook TOP dataset for navigation domain.
arXiv Detail & Related papers (2020-11-03T22:55:40Z) - Software Language Comprehension using a Program-Derived Semantics Graph [29.098303489400394]
We present the program-derived semantics graph, a new structure to capture semantics of code.
The PSG is designed to provide a single structure for capturing program semantics at multiple levels of abstraction.
Although our exploration into the PSG is in its infancy, our early results and architectural analysis indicate it is a promising new research direction to automatically extract program semantics.
arXiv Detail & Related papers (2020-04-02T01:37:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.