One2Set: Generating Diverse Keyphrases as a Set
- URL: http://arxiv.org/abs/2105.11134v1
- Date: Mon, 24 May 2021 07:29:47 GMT
- Title: One2Set: Generating Diverse Keyphrases as a Set
- Authors: Jiacheng Ye, Tao Gui, Yichao Luo, Yige Xu, Qi Zhang
- Abstract summary: We propose a new training paradigm One2Set without predefining an order to the keyphrases.
We propose a $K$-step target assignment mechanism via bipartite matching, which greatly increases the diversity and reduces the duplication ratio of generated keyphrases.
- Score: 12.670421834049668
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recently, the sequence-to-sequence models have made remarkable progress on
the task of keyphrase generation (KG) by concatenating multiple keyphrases in a
predefined order as a target sequence during training. However, the keyphrases
are inherently an unordered set rather than an ordered sequence. Imposing a
predefined order will introduce wrong bias during training, which can highly
penalize shifts in the order between keyphrases. In this work, we propose a new
training paradigm One2Set without predefining an order to concatenate the
keyphrases. To fit this paradigm, we propose a novel model that utilizes a
fixed set of learned control codes as conditions to generate a set of
keyphrases in parallel. To solve the problem that there is no correspondence
between each prediction and target during training, we propose a $K$-step
target assignment mechanism via bipartite matching, which greatly increases the
diversity and reduces the duplication ratio of generated keyphrases. The
experimental results on multiple benchmarks demonstrate that our approach
significantly outperforms the state-of-the-art methods.
Related papers
- Representation Learning of Tangled Key-Value Sequence Data for Early Classification [19.943311002522154]
Key-value sequence data has become ubiquitous and naturally appears in a variety of real-world applications.
Classifying these key-value sequences is important in many scenarios such as user profiling and malicious applications identification.
In many time-sensitive scenarios, besides the requirement of classifying a key-value sequence accurately, it is also desired to classify a key-value sequence early.
arXiv Detail & Related papers (2024-04-11T03:23:15Z) - SimCKP: Simple Contrastive Learning of Keyphrase Representations [36.88517357720033]
We propose SimCKP, a simple contrastive learning framework that consists of two stages: 1) An extractor-generator that extracts keyphrases by learning context-aware phrase-level representations in a contrastive manner while also generating keyphrases that do not appear in the document; and 2) A reranker that adapts scores for each generated phrase by likewise aligning their representations with the corresponding document.
arXiv Detail & Related papers (2023-10-12T11:11:54Z) - Instruction Position Matters in Sequence Generation with Large Language
Models [67.87516654892343]
Large language models (LLMs) are capable of performing conditional sequence generation tasks, such as translation or summarization.
We propose enhancing the instruction-following capability of LLMs by shifting the position of task instructions after the input sentences.
arXiv Detail & Related papers (2023-08-23T12:36:57Z) - Discovering Non-monotonic Autoregressive Orderings with Variational
Inference [67.27561153666211]
We develop an unsupervised parallelizable learner that discovers high-quality generation orders purely from training data.
We implement the encoder as a Transformer with non-causal attention that outputs permutations in one forward pass.
Empirical results in language modeling tasks demonstrate that our method is context-aware and discovers orderings that are competitive with or even better than fixed orders.
arXiv Detail & Related papers (2021-10-27T16:08:09Z) - Keyphrase Extraction with Dynamic Graph Convolutional Networks and
Diversified Inference [50.768682650658384]
Keyphrase extraction (KE) aims to summarize a set of phrases that accurately express a concept or a topic covered in a given document.
Recent Sequence-to-Sequence (Seq2Seq) based generative framework is widely used in KE task, and it has obtained competitive performance on various benchmarks.
In this paper, we propose to adopt the Dynamic Graph Convolutional Networks (DGCN) to solve the above two problems simultaneously.
arXiv Detail & Related papers (2020-10-24T08:11:23Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z) - Keyphrase Prediction With Pre-trained Language Model [16.06425973336514]
We propose to divide the keyphrase prediction into two subtasks, i.e., present keyphrase extraction (PKE) and absent keyphrase generation (AKG)
For PKE, we tackle this task as a sequence labeling problem with the pre-trained language model BERT.
For AKG, we introduce a Transformer-based architecture, which fully integrates the present keyphrase knowledge learned from PKE by the fine-tuned BERT.
arXiv Detail & Related papers (2020-04-22T09:35:02Z) - Exclusive Hierarchical Decoding for Deep Keyphrase Generation [63.357895318562214]
Keyphrase generation (KG) aims to summarize the main ideas of a document into a set of keyphrases.
Previous work in this setting employs a sequential decoding process to generate keyphrases.
We propose an exclusive hierarchical decoding framework that includes a hierarchical decoding process and either a soft or a hard exclusion mechanism.
arXiv Detail & Related papers (2020-04-18T02:58:00Z) - Multi-level Head-wise Match and Aggregation in Transformer for Textual
Sequence Matching [87.97265483696613]
We propose a new approach to sequence pair matching with Transformer, by learning head-wise matching representations on multiple levels.
Experiments show that our proposed approach can achieve new state-of-the-art performance on multiple tasks.
arXiv Detail & Related papers (2020-01-20T20:02:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.