Incorporating Vision Bias into Click Models for Image-oriented Search
Engine
- URL: http://arxiv.org/abs/2101.02459v1
- Date: Thu, 7 Jan 2021 10:01:31 GMT
- Title: Incorporating Vision Bias into Click Models for Image-oriented Search
Engine
- Authors: Ningxin Xu, Cheng Yang, Yixin Zhu, Xiaowei Hu, Changhu Wang
- Abstract summary: In this paper, we assume that vision bias exists in an image-oriented search engine as another crucial factor affecting the examination probability aside from position.
We use regression-based EM algorithm to predict the vision bias given the visual features extracted from candidate documents.
- Score: 51.192784793764176
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most typical click models assume that the probability of a document to be
examined by users only depends on position, such as PBM and UBM. It works well
in various kinds of search engines. However, in a search engine where massive
candidate documents display images as responses to the query, the examination
probability should not only depend on position. The visual appearance of an
image-oriented document also plays an important role in its opportunity to be
examined. In this paper, we assume that vision bias exists in an image-oriented
search engine as another crucial factor affecting the examination probability
aside from position. Specifically, we apply this assumption to classical click
models and propose an extended model, to better capture the examination
probabilities of documents. We use regression-based EM algorithm to predict the
vision bias given the visual features extracted from candidate documents.
Empirically, we evaluate our model on a dataset developed from a real-world
online image-oriented search engine, and demonstrate that our proposed model
can achieve significant improvements over its baseline model in data fitness
and sparsity handling.
Related papers
- Measuring Agreeableness Bias in Multimodal Models [0.3529736140137004]
This paper examines a phenomenon in multimodal language models where pre-marked options in question images can influence model responses.
We present models with images of multiple-choice questions, which they initially answer correctly, then expose the same model to versions with pre-marked options.
Our findings reveal a significant shift in the models' responses towards the pre-marked option, even when it contradicts their answers in the neutral settings.
arXiv Detail & Related papers (2024-08-17T06:25:36Z) - Text-to-Image Diffusion Models are Great Sketch-Photo Matchmakers [120.49126407479717]
This paper explores text-to-image diffusion models for Zero-Shot Sketch-based Image Retrieval (ZS-SBIR)
We highlight a pivotal discovery: the capacity of text-to-image diffusion models to seamlessly bridge the gap between sketches and photos.
arXiv Detail & Related papers (2024-03-12T00:02:03Z) - Debiasing Vision-Language Models via Biased Prompts [79.04467131711775]
We propose a general approach for debiasing vision-language foundation models by projecting out biased directions in the text embedding.
We show that debiasing only the text embedding with a calibrated projection matrix suffices to yield robust classifiers and fair generative models.
arXiv Detail & Related papers (2023-01-31T20:09:33Z) - Pattern Spotting and Image Retrieval in Historical Documents using Deep
Hashing [60.67014034968582]
This paper presents a deep learning approach for image retrieval and pattern spotting in digital collections of historical documents.
Deep learning models are used for feature extraction, considering two distinct variants, which provide either real-valued or binary code representations.
The proposed approach also reduces the search time by up to 200x and the storage cost up to 6,000x when compared to related works.
arXiv Detail & Related papers (2022-08-04T01:39:37Z) - A Graph-Enhanced Click Model for Web Search [67.27218481132185]
We propose a novel graph-enhanced click model (GraphCM) for web search.
We exploit both intra-session and inter-session information for the sparsity and cold-start problems.
arXiv Detail & Related papers (2022-06-17T08:32:43Z) - Vision Models Are More Robust And Fair When Pretrained On Uncurated
Images Without Supervision [38.22842778742829]
Discriminative self-supervised learning allows training models on any random group of internet images.
We train models on billions of random images without any data pre-processing or prior assumptions about what we want the model to learn.
We extensively study and validate our model performance on over 50 benchmarks including fairness, to distribution shift, geographical diversity, fine grained recognition, image copy detection and many image classification datasets.
arXiv Detail & Related papers (2022-02-16T22:26:47Z) - Visual Recognition with Deep Learning from Biased Image Datasets [6.10183951877597]
We show how biasing models can be applied to remedy problems in the context of visual recognition.
Based on the (approximate) knowledge of the biasing mechanisms at work, our approach consists in reweighting the observations.
We propose to use a low dimensional image representation, shared across the image databases.
arXiv Detail & Related papers (2021-09-06T10:56:58Z) - An Intelligent Hybrid Model for Identity Document Classification [0.0]
Digitization may provide opportunities (e.g., increase in productivity, disaster recovery, and environmentally friendly solutions) and challenges for businesses.
One of the main challenges would be to accurately classify numerous scanned documents uploaded every day by customers.
There are not many studies available to address the challenge as an application of image classification.
The proposed approach has been implemented using Python and experimentally validated on synthetic and real-world datasets.
arXiv Detail & Related papers (2021-06-07T13:08:00Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.