Domain Controlled Title Generation with Human Evaluation
- URL: http://arxiv.org/abs/2103.05069v1
- Date: Mon, 8 Mar 2021 20:55:55 GMT
- Title: Domain Controlled Title Generation with Human Evaluation
- Authors: Abdul Waheed, Muskan Goyal, Nimisha Mittal, Deepak Gupta
- Abstract summary: A good title allows you to get the attention that your research deserves.
For domain-controlled titles, we used the pre-trained text-to-text transformer model and the additional token technique.
Title tokens are sampled from a local distribution (which is a subset of global vocabulary) of the domain-specific vocabulary and not global vocabulary.
- Score: 2.5505887482902287
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study automatic title generation and present a method for generating
domain-controlled titles for scientific articles. A good title allows you to
get the attention that your research deserves. A title can be interpreted as a
high-compression description of a document containing information on the
implemented process. For domain-controlled titles, we used the pre-trained
text-to-text transformer model and the additional token technique. Title tokens
are sampled from a local distribution (which is a subset of global vocabulary)
of the domain-specific vocabulary and not global vocabulary, thereby generating
a catchy title and closely linking it to its corresponding abstract. Generated
titles looked realistic, convincing, and very close to the ground truth. We
have performed automated evaluation using ROUGE metric and human evaluation
using five parameters to make a comparison between human and machine-generated
titles. The titles produced were considered acceptable with higher metric
ratings in contrast to the original titles. Thus we concluded that our research
proposes a promising method for domain-controlled title generation.
Related papers
- Can pre-trained language models generate titles for research papers? [3.3489486000815765]
In this paper, we fine-tune pre-trained language models to generate titles of papers from their abstracts.
We also use GPT-3.5-turbo in a zero-shot setting to generate paper titles.
Our observations suggest that AI-generated paper titles are generally accurate and appropriate.
arXiv Detail & Related papers (2024-09-22T21:34:49Z) - VacancySBERT: the approach for representation of titles and skills for
semantic similarity search in the recruitment domain [0.0]
The paper focuses on deep learning semantic search algorithms applied in the HR domain.
The aim of the article is developing a novel approach to training a Siamese network to link the skills mentioned in the job ad with the title.
arXiv Detail & Related papers (2023-07-31T13:21:15Z) - Multimodal Prompt Learning for Product Title Generation with Extremely
Limited Labels [66.54691023795097]
We propose a prompt-based approach, i.e., the Multimodal Prompt Learning framework, to generate titles for novel products with limited labels.
We build a set of multimodal prompts from different modalities to preserve the corresponding characteristics and writing styles of novel products.
With the full labelled data for training, our method achieves state-of-the-art results.
arXiv Detail & Related papers (2023-07-05T00:40:40Z) - So Different Yet So Alike! Constrained Unsupervised Text Style Transfer [54.4773992696361]
We introduce a method for constrained unsupervised text style transfer by introducing two complementary losses to the generative adversarial network (GAN) family of models.
Unlike the competing losses used in GANs, we introduce cooperative losses where the discriminator and the generator cooperate and reduce the same loss.
We show that the complementary cooperative losses improve text quality, according to both automated and human evaluation measures.
arXiv Detail & Related papers (2022-05-09T07:46:40Z) - Keyphrase Generation Beyond the Boundaries of Title and Abstract [28.56508031460787]
Keyphrase generation aims at generating phrases (keyphrases) that best describe a given document.
In this work, we explore whether the integration of additional data from semantically similar articles or from the full text of the given article can be helpful for a neural keyphrase generation model.
We discover that adding sentences from the full text particularly in the form of summary of the article can significantly improve the generation of both types of keyphrases.
arXiv Detail & Related papers (2021-12-13T16:33:01Z) - Development of an Extractive Title Generation System Using Titles of
Papers of Top Conferences for Intermediate English Students [0.0]
This study develops an extractive title generation system that formulates titles from keywords extracted from an abstract.
We also realize a title evaluation model that can evaluate the appropriateness of paper titles.
The results show that our evaluation model can identify top-conference titles more effectively than intermediate English and beginner students.
arXiv Detail & Related papers (2021-10-08T15:41:27Z) - DeepTitle -- Leveraging BERT to generate Search Engine Optimized
Headlines [0.0]
We showcase how a pre-trained language model can be leveraged to create an abstractive news headline generator for German language.
We incorporate state of the art fine-tuning techniques for abstractive text summarization, i.e. we use different baits for the encoder and decoder.
We conduct experiments on a German news data set and achieve a ROUGE-L-gram F-score of 40.02.
arXiv Detail & Related papers (2021-07-22T21:32:54Z) - Accelerating Text Mining Using Domain-Specific Stop Word Lists [57.76576681191192]
We present a novel approach for the automatic extraction of domain-specific words called the hyperplane-based approach.
The hyperplane-based approach can significantly reduce text dimensionality by eliminating irrelevant features.
Results indicate that the hyperplane-based approach can reduce the dimensionality of the corpus by 90% and outperforms mutual information.
arXiv Detail & Related papers (2020-11-18T17:42:32Z) - CMT in TREC-COVID Round 2: Mitigating the Generalization Gaps from Web
to Special Domain Search [89.48123965553098]
This paper presents a search system to alleviate the special domain adaption problem.
The system utilizes the domain-adaptive pretraining and few-shot learning technologies to help neural rankers mitigate the domain discrepancy.
Our system performs the best among the non-manual runs in Round 2 of the TREC-COVID task.
arXiv Detail & Related papers (2020-11-03T09:10:48Z) - Addressing Zero-Resource Domains Using Document-Level Context in Neural
Machine Translation [80.40677540516616]
We show that when in-domain parallel data is not available, access to document-level context enables better capturing of domain generalities.
We present two document-level Transformer models which are capable of using large context sizes.
arXiv Detail & Related papers (2020-04-30T16:28:19Z) - Structured Domain Adaptation with Online Relation Regularization for
Unsupervised Person Re-ID [62.90727103061876]
Unsupervised domain adaptation (UDA) aims at adapting the model trained on a labeled source-domain dataset to an unlabeled target-domain dataset.
We propose an end-to-end structured domain adaptation framework with an online relation-consistency regularization term.
Our proposed framework is shown to achieve state-of-the-art performance on multiple UDA tasks of person re-ID.
arXiv Detail & Related papers (2020-03-14T14:45:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.