InfeRE: Step-by-Step Regex Generation via Chain of Inference
- URL: http://arxiv.org/abs/2308.04041v1
- Date: Tue, 8 Aug 2023 04:37:41 GMT
- Title: InfeRE: Step-by-Step Regex Generation via Chain of Inference
- Authors: Shuai Zhang, Xiaodong Gu, Yuting Chen, Beijun Shen
- Abstract summary: In this paper, we propose a new paradigm called InfeRE, which decomposes the generation of expressions into chains of step-by-step inference.
We evaluate InfeRE on two publicly available datasets, NL-RX-Turk and KB13, and compare the results with state-of-the-art approaches and the popular tree-based generation approach TRANX.
- Score: 15.276963928784047
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatically generating regular expressions (abbrev. regexes) from natural
language description (NL2RE) has been an emerging research area. Prior studies
treat regex as a linear sequence of tokens and generate the final expressions
autoregressively in a single pass. They did not take into account the
step-by-step internal text-matching processes behind the final results. This
significantly hinders the efficacy and interpretability of regex generation by
neural language models. In this paper, we propose a new paradigm called InfeRE,
which decomposes the generation of regexes into chains of step-by-step
inference. To enhance the robustness, we introduce a self-consistency decoding
mechanism that ensembles multiple outputs sampled from different models. We
evaluate InfeRE on two publicly available datasets, NL-RX-Turk and KB13, and
compare the results with state-of-the-art approaches and the popular tree-based
generation approach TRANX. Experimental results show that InfeRE substantially
outperforms previous baselines, yielding 16.3% and 14.7% improvement in DFA@5
accuracy on two datasets, respectively. Particularly, InfeRE outperforms the
popular tree-based generation approach by 18.1% and 11.3% on both datasets,
respectively, in terms of DFA@5 accuracy.
Related papers
- SRFormer: Text Detection Transformer with Incorporated Segmentation and
Regression [6.74412860849373]
We propose SRFormer, a unified DETR-based model with amalgamated and Regression.
Our empirical analysis indicates that favorable segmentation predictions can be obtained at the initial decoder layers.
Our method's exceptional robustness, superior training and data efficiency, as well as its state-of-the-art performance.
arXiv Detail & Related papers (2023-08-21T07:34:31Z) - SelfSeg: A Self-supervised Sub-word Segmentation Method for Neural
Machine Translation [51.881877192924414]
Sub-word segmentation is an essential pre-processing step for Neural Machine Translation (NMT)
This paper introduces SelfSeg, a self-supervised neural sub-word segmentation method.
SelfSeg is much faster to train/decode and requires only monolingual dictionaries instead of parallel corpora.
arXiv Detail & Related papers (2023-07-31T04:38:47Z) - Copy Is All You Need [66.00852205068327]
We formulate text generation as progressively copying text segments from an existing text collection.
Our approach achieves better generation quality according to both automatic and human evaluations.
Our approach attains additional performance gains by simply scaling up to larger text collections.
arXiv Detail & Related papers (2023-07-13T05:03:26Z) - Sequence Generation with Label Augmentation for Relation Extraction [17.38986046630852]
We propose Relation Extraction with Label Augmentation (RELA), a Seq2Seq model with automatic label augmentation for relation extraction.
Experimental results show RELA achieves competitive results compared with previous methods on four RE datasets.
arXiv Detail & Related papers (2022-12-29T11:28:05Z) - DORE: Document Ordered Relation Extraction based on Generative Framework [56.537386636819626]
This paper investigates the root cause of the underwhelming performance of the existing generative DocRE models.
We propose to generate a symbolic and ordered sequence from the relation matrix which is deterministic and easier for model to learn.
Experimental results on four datasets show that our proposed method can improve the performance of the generative DocRE models.
arXiv Detail & Related papers (2022-10-28T11:18:10Z) - A Simple and Strong Baseline for End-to-End Neural RST-style Discourse
Parsing [44.72809363746258]
This paper explores a strong baseline by integrating existing simple parsing strategies, top-down and bottom-up, with various transformer-based pre-trained language models.
The experimental results obtained from two benchmark datasets demonstrate that the parsing performance relies on the pretrained language models rather than the parsing strategies.
arXiv Detail & Related papers (2022-10-15T18:38:08Z) - HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text
Extractive Summarization [57.798070356553936]
HETFORMER is a Transformer-based pre-trained model with multi-granularity sparse attentions for extractive summarization.
Experiments on both single- and multi-document summarization tasks show that HETFORMER achieves state-of-the-art performance in Rouge F1.
arXiv Detail & Related papers (2021-10-12T22:42:31Z) - Exploiting Language Model for Efficient Linguistic Steganalysis: An
Empirical Study [23.311007481830647]
We present two methods to efficient linguistic steganalysis.
One is to pre-train a language model based on RNN, and the other is to pre-train a sequence autoencoder.
arXiv Detail & Related papers (2021-07-26T12:37:18Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z) - AMR Parsing via Graph-Sequence Iterative Inference [62.85003739964878]
We propose a new end-to-end model that treats AMR parsing as a series of dual decisions on the input sequence and the incrementally constructed graph.
We show that the answers to these two questions are mutually causalities.
We design a model based on iterative inference that helps achieve better answers in both perspectives, leading to greatly improved parsing accuracy.
arXiv Detail & Related papers (2020-04-12T09:15:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.