Show, Write, and Retrieve: Entity-aware Article Generation and Retrieval
- URL: http://arxiv.org/abs/2112.05917v3
- Date: Fri, 20 Oct 2023 20:44:27 GMT
- Title: Show, Write, and Retrieve: Entity-aware Article Generation and Retrieval
- Authors: Zhongping Zhang, Yiwen Gu, Bryan A. Plummer
- Abstract summary: We propose an ENtity-aware article GeneratIoN and rEtrieval framework, to explicitly incorporate named entities into language models.
We conducted experiments on three public datasets: GoodNews, VisualNews, and WikiText.
Our results demonstrate that our model can boost both article generation and article retrieval performance, with a 4-5 perplexity improvement in article generation and a 3-4% boost in recall@1 in article retrieval.
- Score: 18.270878909735256
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Article comprehension is an important challenge in natural language
processing with many applications such as article generation or
image-to-article retrieval. Prior work typically encodes all tokens in articles
uniformly using pretrained language models. However, in many applications, such
as understanding news stories, these articles are based on real-world events
and may reference many named entities that are difficult to accurately
recognize and predict by language models. To address this challenge, we propose
an ENtity-aware article GeneratIoN and rEtrieval (ENGINE) framework, to
explicitly incorporate named entities into language models. ENGINE has two main
components: a named-entity extraction module to extract named entities from
both metadata and embedded images associated with articles, and an entity-aware
mechanism that enhances the model's ability to recognize and predict entity
names. We conducted experiments on three public datasets: GoodNews, VisualNews,
and WikiText, where our results demonstrate that our model can boost both
article generation and article retrieval performance, with a 4-5 perplexity
improvement in article generation and a 3-4% boost in recall@1 in article
retrieval. We release our implementation at
https://github.com/Zhongping-Zhang/ENGINE .
Related papers
- Seed-Guided Fine-Grained Entity Typing in Science and Engineering
Domains [51.02035914828596]
We study the task of seed-guided fine-grained entity typing in science and engineering domains.
We propose SEType which first enriches the weak supervision by finding more entities for each seen type from an unlabeled corpus.
It then matches the enriched entities to unlabeled text to get pseudo-labeled samples and trains a textual entailment model that can make inferences for both seen and unseen types.
arXiv Detail & Related papers (2024-01-23T22:36:03Z) - Informed Named Entity Recognition Decoding for Generative Language
Models [3.5323691899538128]
We propose Informed Named Entity Recognition Decoding (iNERD), which treats named entity recognition as a generative process.
We coarse-tune our model on a merged named entity corpus to strengthen its performance, evaluate five generative language models on eight named entity recognition datasets, and achieve remarkable results.
arXiv Detail & Related papers (2023-08-15T14:16:29Z) - DesCo: Learning Object Recognition with Rich Language Descriptions [93.8177229428617]
Recent development in vision-language approaches has instigated a paradigm shift in learning visual recognition models from language supervision.
We propose a new description-conditioned (DesCo) paradigm of learning object recognition models with rich language descriptions.
arXiv Detail & Related papers (2023-06-24T21:05:02Z) - Multi-Modal Classifiers for Open-Vocabulary Object Detection [104.77331131447541]
The goal of this paper is open-vocabulary object detection (OVOD)
We adopt a standard two-stage object detector architecture.
We explore three ways via: language descriptions, image exemplars, or a combination of the two.
arXiv Detail & Related papers (2023-06-08T18:31:56Z) - Focus! Relevant and Sufficient Context Selection for News Image
Captioning [69.36678144800936]
News Image Captioning requires describing an image by leveraging additional context from a news article.
We propose to use the pre-trained vision and language retrieval model CLIP to localize the visually grounded entities in the news article.
Our experiments demonstrate that by simply selecting a better context from the article, we can significantly improve the performance of existing models.
arXiv Detail & Related papers (2022-12-01T20:00:27Z) - BenchCLAMP: A Benchmark for Evaluating Language Models on Syntactic and
Semantic Parsing [55.058258437125524]
We introduce BenchCLAMP, a Benchmark to evaluate Constrained LAnguage Model Parsing.
We benchmark eight language models, including two GPT-3 variants available only through an API.
Our experiments show that encoder-decoder pretrained language models can achieve similar performance or surpass state-of-the-art methods for syntactic and semantic parsing when the model output is constrained to be valid.
arXiv Detail & Related papers (2022-06-21T18:34:11Z) - LUKE: Deep Contextualized Entity Representations with Entity-aware
Self-attention [37.111204321059084]
We propose new pretrained contextualized representations of words and entities based on the bidirectional transformer.
Our model is trained using a new pretraining task based on the masked language model of BERT.
We also propose an entity-aware self-attention mechanism that is an extension of the self-attention mechanism of the transformer.
arXiv Detail & Related papers (2020-10-02T15:38:03Z) - Transform and Tell: Entity-Aware News Image Captioning [77.4898875082832]
We propose an end-to-end model which generates captions for images embedded in news articles.
We address the first challenge by associating words in the caption with faces and objects in the image, via a multi-modal, multi-head attention mechanism.
We tackle the second challenge with a state-of-the-art transformer language model that uses byte-pair-encoding to generate captions as a sequence of word parts.
arXiv Detail & Related papers (2020-04-17T05:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.