FS-COCO: Towards Understanding of Freehand Sketches of Common Objects in
Context
- URL: http://arxiv.org/abs/2203.02113v1
- Date: Fri, 4 Mar 2022 03:00:51 GMT
- Title: FS-COCO: Towards Understanding of Freehand Sketches of Common Objects in
Context
- Authors: Pinaki Nath Chowdhury and Aneeshan Sain and Yulia Gryaditskaya and
Ayan Kumar Bhunia and Tao Xiang and Yi-Zhe Song
- Abstract summary: We advance sketch research to scenes with the first dataset of freehand scene sketches, FS-COCO.
Our dataset comprises 10,000 freehand scene vector sketches with per point space-time information by 100 non-expert individuals.
We study for the first time the problem of the fine-grained image retrieval from freehand scene sketches and sketch captions.
- Score: 112.07988211268612
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We advance sketch research to scenes with the first dataset of freehand scene
sketches, FS-COCO. With practical applications in mind, we collect sketches
that convey well scene content but can be sketched within a few minutes by a
person with any sketching skills. Our dataset comprises 10,000 freehand scene
vector sketches with per point space-time information by 100 non-expert
individuals, offering both object- and scene-level abstraction. Each sketch is
augmented with its text description. Using our dataset, we study for the first
time the problem of the fine-grained image retrieval from freehand scene
sketches and sketch captions. We draw insights on (i) Scene salience encoded in
sketches with strokes temporal order; (ii) The retrieval performance accuracy
from scene sketches against image captions; (iii) Complementarity of
information in sketches and image captions, as well as the potential benefit of
combining the two modalities. In addition, we propose new solutions enabled by
our dataset (i) We adopt meta-learning to show how the retrieval model can be
fine-tuned to a new user style given just a small set of sketches, (ii) We
extend a popular vector sketch LSTM-based encoder to handle sketches with
larger complexity than was supported by previous work. Namely, we propose a
hierarchical sketch decoder, which we leverage at a sketch-specific "pretext"
task. Our dataset enables for the first time research on freehand scene sketch
understanding and its practical applications.
Related papers
- SketchTriplet: Self-Supervised Scenarized Sketch-Text-Image Triplet Generation [6.39528707908268]
There continues to be a lack of large-scale paired datasets for scene sketches.
We propose a self-supervised method for scene sketch generation that does not rely on any existing scene sketch.
We contribute a large-scale dataset centered around scene sketches, comprising highly semantically consistent "text-sketch-image" triplets.
arXiv Detail & Related papers (2024-05-29T06:43:49Z) - Sketch Less Face Image Retrieval: A New Challenge [9.703239229149261]
Drawing a complete face sketch often needs skills and takes time, which hinders its widespread applicability in the practice.
In this study, we proposed sketch less face image retrieval (SLFIR), in which the retrieval was carried out at each stroke and aim to retrieve the target face photo using a partial sketch with as few strokes as possible.
Experiments indicate that the new framework can finish the retrieval using a partial or pool drawing sketch.
arXiv Detail & Related papers (2023-02-11T02:36:00Z) - Towards Practicality of Sketch-Based Visual Understanding [15.30818342202786]
Sketches have been used to conceptualise and depict visual objects from pre-historic times.
This thesis aims to progress sketch-based visual understanding towards more practicality.
arXiv Detail & Related papers (2022-10-27T03:12:57Z) - I Know What You Draw: Learning Grasp Detection Conditioned on a Few
Freehand Sketches [74.63313641583602]
We propose a method to generate a potential grasp configuration relevant to the sketch-depicted objects.
Our model is trained and tested in an end-to-end manner which is easy to be implemented in real-world applications.
arXiv Detail & Related papers (2022-05-09T04:23:36Z) - Multi-granularity Association Learning Framework for on-the-fly
Fine-Grained Sketch-based Image Retrieval [7.797006835701767]
Fine-grained sketch-based image retrieval (FG-SBIR) addresses the problem of retrieving a particular photo in a given query sketch.
In this study, we aim to retrieve the target photo with the least number of strokes possible (incomplete sketch)
We propose a multi-granularity association learning framework that further optimize the embedding space of all incomplete sketches.
arXiv Detail & Related papers (2022-01-13T14:38:50Z) - On Learning Semantic Representations for Million-Scale Free-Hand
Sketches [146.52892067335128]
We study learning semantic representations for million-scale free-hand sketches.
We propose a dual-branch CNNRNN network architecture to represent sketches.
We explore learning the sketch-oriented semantic representations in hashing retrieval and zero-shot recognition.
arXiv Detail & Related papers (2020-07-07T15:23:22Z) - Sketch-BERT: Learning Sketch Bidirectional Encoder Representation from
Transformers by Self-supervised Learning of Sketch Gestalt [125.17887147597567]
We present a model of learning Sketch BiBERT Representation from Transformer (Sketch-BERT)
We generalize BERT to sketch domain, with the novel proposed components and pre-training algorithms.
We show that the learned representation of Sketch-BERT can help and improve the performance of the downstream tasks of sketch recognition, sketch retrieval, and sketch gestalt.
arXiv Detail & Related papers (2020-05-19T01:35:44Z) - SketchyCOCO: Image Generation from Freehand Scene Sketches [71.85577739612579]
We introduce the first method for automatic image generation from scene-level freehand sketches.
Key contribution is an attribute vector bridged Geneversarative Adrial Network called EdgeGAN.
We have built a large-scale composite dataset called SketchyCOCO to support and evaluate the solution.
arXiv Detail & Related papers (2020-03-05T14:54:10Z) - Deep Self-Supervised Representation Learning for Free-Hand Sketch [51.101565480583304]
We tackle the problem of self-supervised representation learning for free-hand sketches.
Key for the success of our self-supervised learning paradigm lies with our sketch-specific designs.
We show that the proposed approach outperforms the state-of-the-art unsupervised representation learning methods.
arXiv Detail & Related papers (2020-02-03T16:28:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.