VisualTextRank: Unsupervised Graph-based Content Extraction for
Automating Ad Text to Image Search
- URL: http://arxiv.org/abs/2108.02725v1
- Date: Thu, 5 Aug 2021 16:47:21 GMT
- Title: VisualTextRank: Unsupervised Graph-based Content Extraction for
Automating Ad Text to Image Search
- Authors: Shaunak Mishra, Mikhail Kuznetsov, Gaurav Srivastava, and Maxim
Sviridenko
- Abstract summary: We propose VisualTextRank as an unsupervised method to augment input ad text using semantically similar ads.
VisualTextRank builds on prior work on graph based context extraction.
Online tests with a simplified version led to a 28.7% increase in the usage of stock image search.
- Score: 6.107273836558503
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Numerous online stock image libraries offer high quality yet copyright free
images for use in marketing campaigns. To assist advertisers in navigating such
third party libraries, we study the problem of automatically fetching relevant
ad images given the ad text (via a short textual query for images). Motivated
by our observations in logged data on ad image search queries (given ad text),
we formulate a keyword extraction problem, where a keyword extracted from the
ad text (or its augmented version) serves as the ad image query. In this
context, we propose VisualTextRank: an unsupervised method to (i) augment input
ad text using semantically similar ads, and (ii) extract the image query from
the augmented ad text. VisualTextRank builds on prior work on graph based
context extraction (biased TextRank in particular) by leveraging both the text
and image of similar ads for better keyword extraction, and using advertiser
category specific biasing with sentence-BERT embeddings. Using data collected
from the Verizon Media Native (Yahoo Gemini) ad platform's stock image search
feature for onboarding advertisers, we demonstrate the superiority of
VisualTextRank compared to competitive keyword extraction baselines (including
an $11\%$ accuracy lift over biased TextRank). For the case when the stock
image library is restricted to English queries, we show the effectiveness of
VisualTextRank on multilingual ads (translated to English) while leveraging
semantically similar English ads. Online tests with a simplified version of
VisualTextRank led to a 28.7% increase in the usage of stock image search, and
a 41.6% increase in the advertiser onboarding rate in the Verizon Media Native
ad platform.
Related papers
- AdSEE: Investigating the Impact of Image Style Editing on Advertisement
Attractiveness [25.531489722164178]
We propose Advertisement Style Editing and Attractiveness Enhancement (AdSEE), which explores whether semantic editing to ads images can affect or alter the popularity of online advertisements.
We introduce StyleGAN-based facial semantic editing and inversion to ads images and train a click rate predictor attributing GAN-based face latent representations to click rates.
Online A/B tests performed over a period of 5 days have verified the increased click-through rates of AdSEE-edited samples as compared to a control group of original ads.
arXiv Detail & Related papers (2023-09-15T04:52:49Z) - Orientation-Independent Chinese Text Recognition in Scene Images [61.34060587461462]
We take the first attempt to extract orientation-independent visual features by disentangling content and orientation information of text images.
Specifically, we introduce a Character Image Reconstruction Network (CIRN) to recover corresponding printed character images with disentangled content and orientation information.
arXiv Detail & Related papers (2023-09-03T05:30:21Z) - Improving Multimodal Datasets with Image Captioning [65.74736570293622]
We study how generated captions can increase the utility of web-scraped datapoints with nondescript text.
Our experiments with using generated captions at DataComp's large scale (1.28B image-text pairs) offer insights into the limitations of synthetic text.
arXiv Detail & Related papers (2023-07-19T17:47:12Z) - Text-based Person Search without Parallel Image-Text Data [52.63433741872629]
Text-based person search (TBPS) aims to retrieve the images of the target person from a large image gallery based on a given natural language description.
Existing methods are dominated by training models with parallel image-text pairs, which are very costly to collect.
In this paper, we make the first attempt to explore TBPS without parallel image-text data.
arXiv Detail & Related papers (2023-05-22T12:13:08Z) - Efficient Image-Text Retrieval via Keyword-Guided Pre-Screening [53.1711708318581]
Current image-text retrieval methods suffer from $N$-related time complexity.
This paper presents a simple and effective keyword-guided pre-screening framework for the image-text retrieval.
arXiv Detail & Related papers (2023-03-14T09:36:42Z) - SpaText: Spatio-Textual Representation for Controllable Image Generation [61.89548017729586]
SpaText is a new method for text-to-image generation using open-vocabulary scene control.
In addition to a global text prompt that describes the entire scene, the user provides a segmentation map.
We show its effectiveness on two state-of-the-art diffusion models: pixel-based and latent-conditional-based.
arXiv Detail & Related papers (2022-11-25T18:59:10Z) - Revising Image-Text Retrieval via Multi-Modal Entailment [25.988058843564335]
Many-to-many matching phenomenon is quite common in the widely-used image-text retrieval datasets.
We propose a multi-modal entailment classifier to determine whether a sentence is entailed by an image plus its linked captions.
arXiv Detail & Related papers (2022-08-22T07:58:54Z) - NewsStories: Illustrating articles with visual summaries [49.924916589209374]
We introduce a large-scale multimodal dataset containing over 31M articles, 22M images and 1M videos.
We show that state-of-the-art image-text alignment methods are not robust to longer narratives with multiple images.
We introduce an intuitive baseline that outperforms these methods on zero-shot image-set retrieval by 10% on the GoodNews dataset.
arXiv Detail & Related papers (2022-07-26T17:34:11Z) - TSI: an Ad Text Strength Indicator using Text-to-CTR and
Semantic-Ad-Similarity [16.10904771281746]
We propose an ad text strength indicator (TSI) which: (i) predicts the click-through-rate (CTR) for an input ad text, (ii) fetches similar existing ads to create a neighborhood around the input ad, and compares the predicted CTRs in the neighborhood to declare whether the input ad is strong or weak.
As suggestions for ad text improvement, TSI shows anonymized versions of superior ads (higher predicted CTR) in the neighborhood.
arXiv Detail & Related papers (2021-08-18T16:24:40Z) - Learning to Create Better Ads: Generation and Ranking Approaches for Ad
Creative Refinement [26.70647666598025]
We study approaches to refine the given ad text and image by: (i) generating new ad text, (ii) recommending keyphrases for new ad text, and (iii) recommending image tags (objects in image)
Based on A/B tests conducted by multiple advertisers, we form pairwise examples of inferior and superior ad creatives.
We also share broadly applicable insights from our experiments using data from the Yahoo Gemini ad platform.
arXiv Detail & Related papers (2020-08-17T16:46:28Z) - Query-Variant Advertisement Text Generation with Association Knowledge [21.18443320935013]
Traditional text generation methods tend to focus on the general searching needs with high frequency.
We propose a query-variant advertisement text generation task that aims to generate candidate advertisement texts for different web search queries.
Our model can make use of various personalized needs in queries and generate query-variant advertisement texts.
arXiv Detail & Related papers (2020-04-14T12:04:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.