SDCUP: Schema Dependency-Enhanced Curriculum Pre-Training for Table
Semantic Parsing
- URL: http://arxiv.org/abs/2111.09486v1
- Date: Thu, 18 Nov 2021 02:51:04 GMT
- Title: SDCUP: Schema Dependency-Enhanced Curriculum Pre-Training for Table
Semantic Parsing
- Authors: Bowen Qin, Lihan Wang, Binyuan Hui, Ruiying Geng, Zheng Cao, Min Yang,
Jian Sun, Yongbin Li
- Abstract summary: This paper designs two novel pre-training objectives to impose the desired inductive bias into the learned representations for table pre-training.
We propose a schema-aware curriculum learning approach to mitigate the impact of noise and learn effectively from the pre-training data in an easy-to-hard manner.
- Score: 19.779493883522072
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently pre-training models have significantly improved the performance of
various NLP tasks by leveraging large-scale text corpora to improve the
contextual representation ability of the neural network. The large pre-training
language model has also been applied in the area of table semantic parsing.
However, existing pre-training approaches have not carefully explored explicit
interaction relationships between a question and the corresponding database
schema, which is a key ingredient for uncovering their semantic and structural
correspondence. Furthermore, the question-aware representation learning in the
schema grounding context has received less attention in pre-training
objective.To alleviate these issues, this paper designs two novel pre-training
objectives to impose the desired inductive bias into the learned
representations for table pre-training. We further propose a schema-aware
curriculum learning approach to mitigate the impact of noise and learn
effectively from the pre-training data in an easy-to-hard manner. We evaluate
our pre-trained framework by fine-tuning it on two benchmarks, Spider and
SQUALL. The results demonstrate the effectiveness of our pre-training objective
and curriculum compared to a variety of baselines.
Related papers
- Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate [118.37653302885607]
We present the Modality Integration Rate (MIR), an effective, robust, and generalized metric to indicate the multi-modal pre-training quality of Large Vision Language Models (LVLMs)
MIR is indicative about training data selection, training strategy schedule, and model architecture design to get better pre-training results.
arXiv Detail & Related papers (2024-10-09T17:59:04Z) - Pointer-Guided Pre-Training: Infusing Large Language Models with Paragraph-Level Contextual Awareness [3.2925222641796554]
"pointer-guided segment ordering" (SO) is a novel pre-training technique aimed at enhancing the contextual understanding of paragraph-level text representations.
Our experiments show that pointer-guided pre-training significantly enhances the model's ability to understand complex document structures.
arXiv Detail & Related papers (2024-06-06T15:17:51Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Continual Contrastive Finetuning Improves Low-Resource Relation
Extraction [34.76128090845668]
Relation extraction has been particularly challenging in low-resource scenarios and domains.
Recent literature has tackled low-resource RE by self-supervised learning.
We propose to pretrain and finetune the RE model using consistent objectives of contrastive learning.
arXiv Detail & Related papers (2022-12-21T07:30:22Z) - Using Context-to-Vector with Graph Retrofitting to Improve Word
Embeddings [39.30342855873457]
We aim to improve word embeddings by incorporating more contextual information into the Skip-gram framework.
Our methods are well proven to outperform the baselines by a large margin.
arXiv Detail & Related papers (2022-10-30T14:15:43Z) - Forging Multiple Training Objectives for Pre-trained Language Models via
Meta-Learning [97.28779163988833]
Multiple pre-training objectives fill the vacancy of the understanding capability of single-objective language modeling.
We propose textitMOMETAS, a novel adaptive sampler based on meta-learning, which learns the latent sampling pattern on arbitrary pre-training objectives.
arXiv Detail & Related papers (2022-10-19T04:38:26Z) - The Inductive Bias of In-Context Learning: Rethinking Pretraining
Example Design [34.900425311720795]
We show that pretrained NLM can model stronger dependencies between text segments that appeared in the same training example, than it can between text segments that appeared in different training examples.
We propose "kNN-Pretraining": we show that including semantically related non-neighboring sentences in the same pretraining example yields improved sentence representations and open domain question answering abilities.
arXiv Detail & Related papers (2021-10-09T11:05:16Z) - Dense Contrastive Visual-Linguistic Pretraining [53.61233531733243]
Several multimodal representation learning approaches have been proposed that jointly represent image and text.
These approaches achieve superior performance by capturing high-level semantic information from large-scale multimodal pretraining.
We propose unbiased Dense Contrastive Visual-Linguistic Pretraining to replace the region regression and classification with cross-modality region contrastive learning.
arXiv Detail & Related papers (2021-09-24T07:20:13Z) - SLM: Learning a Discourse Language Representation with Sentence
Unshuffling [53.42814722621715]
We introduce Sentence-level Language Modeling, a new pre-training objective for learning a discourse language representation.
We show that this feature of our model improves the performance of the original BERT by large margins.
arXiv Detail & Related papers (2020-10-30T13:33:41Z) - Pre-training Text Representations as Meta Learning [113.3361289756749]
We introduce a learning algorithm which directly optimize model's ability to learn text representations for effective learning of downstream tasks.
We show that there is an intrinsic connection between multi-task pre-training and model-agnostic meta-learning with a sequence of meta-train steps.
arXiv Detail & Related papers (2020-04-12T09:05:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.