Multi-granularity Association Learning Framework for on-the-fly
Fine-Grained Sketch-based Image Retrieval
- URL: http://arxiv.org/abs/2201.05007v1
- Date: Thu, 13 Jan 2022 14:38:50 GMT
- Title: Multi-granularity Association Learning Framework for on-the-fly
Fine-Grained Sketch-based Image Retrieval
- Authors: Dawei Dai, Xiaoyu Tang, Shuyin Xia, Yingge Liu, Guoyin Wang, Zizhong
Chen
- Abstract summary: Fine-grained sketch-based image retrieval (FG-SBIR) addresses the problem of retrieving a particular photo in a given query sketch.
In this study, we aim to retrieve the target photo with the least number of strokes possible (incomplete sketch)
We propose a multi-granularity association learning framework that further optimize the embedding space of all incomplete sketches.
- Score: 7.797006835701767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fine-grained sketch-based image retrieval (FG-SBIR) addresses the problem of
retrieving a particular photo in a given query sketch. However, its widespread
applicability is limited by the fact that it is difficult to draw a complete
sketch for most people, and the drawing process often takes time. In this
study, we aim to retrieve the target photo with the least number of strokes
possible (incomplete sketch), named on-the-fly FG-SBIR (Bhunia et al. 2020),
which starts retrieving at each stroke as soon as the drawing begins. We
consider that there is a significant correlation among these incomplete
sketches in the sketch drawing episode of each photo. To learn more efficient
joint embedding space shared between the photo and its incomplete sketches, we
propose a multi-granularity association learning framework that further
optimizes the embedding space of all incomplete sketches. Specifically, based
on the integrity of the sketch, we can divide a complete sketch episode into
several stages, each of which corresponds to a simple linear mapping layer.
Moreover, our framework guides the vector space representation of the current
sketch to approximate that of its later sketches to realize the retrieval
performance of the sketch with fewer strokes to approach that of the sketch
with more strokes. In the experiments, we proposed more realistic challenges,
and our method achieved superior early retrieval efficiency over the
state-of-the-art methods and alternative baselines on two publicly available
fine-grained sketch retrieval datasets.
Related papers
- SketchTriplet: Self-Supervised Scenarized Sketch-Text-Image Triplet Generation [6.39528707908268]
There continues to be a lack of large-scale paired datasets for scene sketches.
We propose a self-supervised method for scene sketch generation that does not rely on any existing scene sketch.
We contribute a large-scale dataset centered around scene sketches, comprising highly semantically consistent "text-sketch-image" triplets.
arXiv Detail & Related papers (2024-05-29T06:43:49Z) - Picture that Sketch: Photorealistic Image Generation from Abstract
Sketches [109.69076457732632]
Given an abstract, deformed, ordinary sketch from untrained amateurs like you and me, this paper turns it into a photorealistic image.
We do not dictate an edgemap-like sketch to start with, but aim to work with abstract free-hand human sketches.
In doing so, we essentially democratise the sketch-to-photo pipeline, "picturing" a sketch regardless of how good you sketch.
arXiv Detail & Related papers (2023-03-20T14:49:03Z) - Sketch Less Face Image Retrieval: A New Challenge [9.703239229149261]
Drawing a complete face sketch often needs skills and takes time, which hinders its widespread applicability in the practice.
In this study, we proposed sketch less face image retrieval (SLFIR), in which the retrieval was carried out at each stroke and aim to retrieve the target face photo using a partial sketch with as few strokes as possible.
Experiments indicate that the new framework can finish the retrieval using a partial or pool drawing sketch.
arXiv Detail & Related papers (2023-02-11T02:36:00Z) - I Know What You Draw: Learning Grasp Detection Conditioned on a Few
Freehand Sketches [74.63313641583602]
We propose a method to generate a potential grasp configuration relevant to the sketch-depicted objects.
Our model is trained and tested in an end-to-end manner which is easy to be implemented in real-world applications.
arXiv Detail & Related papers (2022-05-09T04:23:36Z) - FS-COCO: Towards Understanding of Freehand Sketches of Common Objects in
Context [112.07988211268612]
We advance sketch research to scenes with the first dataset of freehand scene sketches, FS-COCO.
Our dataset comprises 10,000 freehand scene vector sketches with per point space-time information by 100 non-expert individuals.
We study for the first time the problem of the fine-grained image retrieval from freehand scene sketches and sketch captions.
arXiv Detail & Related papers (2022-03-04T03:00:51Z) - Adversarial Open Domain Adaption for Sketch-to-Photo Synthesis [42.83974176146334]
We explore the open-domain sketch-to-photo translation, which aims to synthesize a realistic photo from a freehand sketch with its class label.
It is challenging due to the lack of training supervision and the large geometry distortion between the freehand sketch and photo domains.
We propose a framework that jointly learns sketch-to-photo and photo-to-sketch generation.
arXiv Detail & Related papers (2021-04-12T17:58:46Z) - Cross-Modal Hierarchical Modelling for Fine-Grained Sketch Based Image
Retrieval [147.24102408745247]
We study a further trait of sketches that has been overlooked to date, that is, they are hierarchical in terms of the levels of detail.
In this paper, we design a novel network that is capable of cultivating sketch-specific hierarchies and exploiting them to match sketch with photo at corresponding hierarchical levels.
arXiv Detail & Related papers (2020-07-29T20:50:25Z) - Sketch Less for More: On-the-Fly Fine-Grained Sketch Based Image
Retrieval [203.2520862597357]
Fine-grained sketch-based image retrieval (FG-SBIR) addresses the problem of retrieving a particular photo instance given a user's query sketch.
We reformulate the conventional FG-SBIR framework to tackle these challenges.
We propose an on-the-fly design that starts retrieving as soon as the user starts drawing.
arXiv Detail & Related papers (2020-02-24T15:36:02Z) - Deep Self-Supervised Representation Learning for Free-Hand Sketch [51.101565480583304]
We tackle the problem of self-supervised representation learning for free-hand sketches.
Key for the success of our self-supervised learning paradigm lies with our sketch-specific designs.
We show that the proposed approach outperforms the state-of-the-art unsupervised representation learning methods.
arXiv Detail & Related papers (2020-02-03T16:28:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.