Development of an Extractive Title Generation System Using Titles of
Papers of Top Conferences for Intermediate English Students
- URL: http://arxiv.org/abs/2110.04204v1
- Date: Fri, 8 Oct 2021 15:41:27 GMT
- Title: Development of an Extractive Title Generation System Using Titles of
Papers of Top Conferences for Intermediate English Students
- Authors: Kento Kaku, Masato Kikuchi, Tadachika Ozono, Toramatsu Shintani
- Abstract summary: This study develops an extractive title generation system that formulates titles from keywords extracted from an abstract.
We also realize a title evaluation model that can evaluate the appropriateness of paper titles.
The results show that our evaluation model can identify top-conference titles more effectively than intermediate English and beginner students.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The formulation of good academic paper titles in English is challenging for
intermediate English authors (particularly students). This is because such
authors are not aware of the type of titles that are generally in use. We aim
to realize a support system for formulating more effective English titles for
intermediate English and beginner authors. This study develops an extractive
title generation system that formulates titles from keywords extracted from an
abstract. Moreover, we realize a title evaluation model that can evaluate the
appropriateness of paper titles. We train the model with titles of
top-conference papers by using BERT. This paper describes the training data,
implementation, and experimental results. The results show that our evaluation
model can identify top-conference titles more effectively than intermediate
English and beginner students.
Related papers
- Can pre-trained language models generate titles for research papers? [3.3489486000815765]
In this paper, we fine-tune pre-trained language models to generate titles of papers from their abstracts.
We also use GPT-3.5-turbo in a zero-shot setting to generate paper titles.
Our observations suggest that AI-generated paper titles are generally accurate and appropriate.
arXiv Detail & Related papers (2024-09-22T21:34:49Z) - Good things come in three: Generating SO Post Titles with Pre-Trained Models, Self Improvement and Post Ranking [5.874782446136913]
Stack Overflow is a prominent Q and A forum, supporting developers in seeking suitable resources on programming-related matters.
Having high-quality question titles is an effective means to attract developers' attention.
Research has been conducted, predominantly leveraging pre-trained models to generate titles from code snippets and problem descriptions.
We present FILLER as a solution to generating Stack Overflow post titles using a fine-tuned language model with self-improvement and post ranking.
arXiv Detail & Related papers (2024-06-21T20:18:34Z) - Chain-of-Factors Paper-Reviewer Matching [32.86512592730291]
We propose a unified model for paper-reviewer matching that jointly considers semantic, topic, and citation factors.
We demonstrate the effectiveness of our proposed Chain-of-Factors model in comparison with state-of-the-art paper-reviewer matching methods and scientific pre-trained language models.
arXiv Detail & Related papers (2023-10-23T01:29:18Z) - Generating Natural Language Queries for More Effective Systematic Review
Screening Prioritisation [53.77226503675752]
The current state of the art uses the final title of the review as a query to rank the documents using BERT-based neural rankers.
In this paper, we explore alternative sources of queries for prioritising screening, such as the Boolean query used to retrieve the documents to be screened and queries generated by instruction-based large-scale language models such as ChatGPT and Alpaca.
Our best approach is not only viable based on the information available at the time of screening, but also has similar effectiveness to the final title.
arXiv Detail & Related papers (2023-09-11T05:12:14Z) - VacancySBERT: the approach for representation of titles and skills for
semantic similarity search in the recruitment domain [0.0]
The paper focuses on deep learning semantic search algorithms applied in the HR domain.
The aim of the article is developing a novel approach to training a Siamese network to link the skills mentioned in the job ad with the title.
arXiv Detail & Related papers (2023-07-31T13:21:15Z) - Multimodal Prompt Learning for Product Title Generation with Extremely
Limited Labels [66.54691023795097]
We propose a prompt-based approach, i.e., the Multimodal Prompt Learning framework, to generate titles for novel products with limited labels.
We build a set of multimodal prompts from different modalities to preserve the corresponding characteristics and writing styles of novel products.
With the full labelled data for training, our method achieves state-of-the-art results.
arXiv Detail & Related papers (2023-07-05T00:40:40Z) - MIReAD: Simple Method for Learning High-quality Representations from
Scientific Documents [77.34726150561087]
We propose MIReAD, a simple method that learns high-quality representations of scientific papers.
We train MIReAD on more than 500,000 PubMed and arXiv abstracts across over 2,000 journal classes.
arXiv Detail & Related papers (2023-05-07T03:29:55Z) - Disambiguation of Company names via Deep Recurrent Networks [101.90357454833845]
We propose a Siamese LSTM Network approach to extract -- via supervised learning -- an embedding of company name strings.
We analyse how an Active Learning approach to prioritise the samples to be labelled leads to a more efficient overall learning pipeline.
arXiv Detail & Related papers (2023-03-07T15:07:57Z) - Tag-Aware Document Representation for Research Paper Recommendation [68.8204255655161]
We propose a hybrid approach that leverages deep semantic representation of research papers based on social tags assigned by users.
The proposed model is effective in recommending research papers even when the rating data is very sparse.
arXiv Detail & Related papers (2022-09-08T09:13:07Z) - Domain Controlled Title Generation with Human Evaluation [2.5505887482902287]
A good title allows you to get the attention that your research deserves.
For domain-controlled titles, we used the pre-trained text-to-text transformer model and the additional token technique.
Title tokens are sampled from a local distribution (which is a subset of global vocabulary) of the domain-specific vocabulary and not global vocabulary.
arXiv Detail & Related papers (2021-03-08T20:55:55Z) - What's New? Summarizing Contributions in Scientific Literature [85.95906677964815]
We introduce a new task of disentangled paper summarization, which seeks to generate separate summaries for the paper contributions and the context of the work.
We extend the S2ORC corpus of academic articles by adding disentangled "contribution" and "context" reference labels.
We propose a comprehensive automatic evaluation protocol which reports the relevance, novelty, and disentanglement of generated outputs.
arXiv Detail & Related papers (2020-11-06T02:23:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.