One Sketch for All: One-Shot Personalized Sketch Segmentation
- URL: http://arxiv.org/abs/2112.10838v1
- Date: Mon, 20 Dec 2021 20:10:44 GMT
- Title: One Sketch for All: One-Shot Personalized Sketch Segmentation
- Authors: Anran Qi, Yulia Gryaditskaya, Tao Xiang, Yi-Zhe Song
- Abstract summary: We present the first one-shot personalized sketch segmentation method.
We aim to segment all sketches belonging to the same category with a single sketch with a given part annotation.
We preserve the parts semantics embedded in the exemplar, and we are robust to input style and abstraction.
- Score: 84.45203849671003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the first one-shot personalized sketch segmentation method. We aim
to segment all sketches belonging to the same category provisioned with a
single sketch with a given part annotation while (i) preserving the parts
semantics embedded in the exemplar, and (ii) being robust to input style and
abstraction. We refer to this scenario as personalized. With that, we
importantly enable a much-desired personalization capability for downstream
fine-grained sketch analysis tasks. To train a robust segmentation module, we
deform the exemplar sketch to each of the available sketches of the same
category. Our method generalizes to sketches not observed during training. Our
central contribution is a sketch-specific hierarchical deformation network.
Given a multi-level sketch-strokes encoding obtained via a graph convolutional
network, our method estimates rigid-body transformation from the reference to
the exemplar, on the upper level. Finer deformation from the exemplar to the
globally warped reference sketch is further obtained through stroke-wise
deformations, on the lower level. Both levels of deformation are guided by mean
squared distances between the keypoints learned without supervision, ensuring
that the stroke semantics are preserved. We evaluate our method against the
state-of-the-art segmentation and perceptual grouping baselines re-purposed for
the one-shot setting and against two few-shot 3D shape segmentation methods. We
show that our method outperforms all the alternatives by more than 10% on
average. Ablation studies further demonstrate that our method is robust to
personalization: changes in input part semantics and style differences.
Related papers
- Class-Agnostic Visio-Temporal Scene Sketch Semantic Segmentation [0.9208007322096532]
Scene sketch semantic segmentation is a crucial task for various applications including sketch-to-image retrieval and scene understanding.
Existing sketch segmentation methods treat sketches as bitmap images, leading to the loss of temporal order among strokes.
We propose a Class-Agnostic-Temporal Network (CAVT) for scene sketch semantic segmentation.
arXiv Detail & Related papers (2024-09-30T22:34:29Z) - Open Vocabulary Semantic Scene Sketch Understanding [5.638866331696071]
We study the underexplored but fundamental vision problem of machine understanding of freehand scene sketches.
We introduce a sketch encoder that results in semantically-aware feature space, which we evaluate by testing its performance on a semantic sketch segmentation task.
Our method outperforms zero-shot CLIP pixel accuracy of segmentation results by 37 points, reaching an accuracy of $85.5%$ on the FS-COCO sketch dataset.
arXiv Detail & Related papers (2023-12-18T19:02:07Z) - Piecewise Planar Hulls for Semi-Supervised Learning of 3D Shape and Pose
from 2D Images [133.68032636906133]
We study the problem of estimating 3D shape and pose of an object in terms of keypoints, from a single 2D image.
The shape and pose are learned directly from images collected by categories and their partial 2D keypoint annotations.
arXiv Detail & Related papers (2022-11-14T16:18:11Z) - From colouring-in to pointillism: revisiting semantic segmentation
supervision [48.637031591058175]
We propose a pointillist approach for semantic segmentation annotation, where only point-wise yes/no questions are answered.
We collected and released 22.6M point labels over 4,171 classes on the Open Images dataset.
arXiv Detail & Related papers (2022-10-25T16:42:03Z) - Learn to Learn Metric Space for Few-Shot Segmentation of 3D Shapes [17.217954254022573]
We introduce a meta-learning-based method for few-shot 3D shape segmentation where only a few labeled samples are provided for the unseen classes.
We demonstrate the superior performance of our proposed on the ShapeNet part dataset under the few-shot scenario, compared with well-established baseline and state-of-the-art semi-supervised methods.
arXiv Detail & Related papers (2021-07-07T01:47:00Z) - Weakly Supervised Volumetric Segmentation via Self-taught Shape
Denoising Model [27.013224147257198]
We propose a novel weakly-supervised segmentation strategy capable of better capturing 3D shape prior in both model prediction and learning.
Our main idea is to extract a self-taught shape representation by leveraging weak labels, and then integrate this representation into segmentation prediction for shape refinement.
arXiv Detail & Related papers (2021-04-27T10:03:45Z) - On Learning Semantic Representations for Million-Scale Free-Hand
Sketches [146.52892067335128]
We study learning semantic representations for million-scale free-hand sketches.
We propose a dual-branch CNNRNN network architecture to represent sketches.
We explore learning the sketch-oriented semantic representations in hashing retrieval and zero-shot recognition.
arXiv Detail & Related papers (2020-07-07T15:23:22Z) - Self-Supervised Tuning for Few-Shot Segmentation [82.32143982269892]
Few-shot segmentation aims at assigning a category label to each image pixel with few annotated samples.
Existing meta-learning method tends to fail in generating category-specifically discriminative descriptor when the visual features extracted from support images are marginalized in embedding space.
This paper presents an adaptive framework tuning, in which the distribution of latent features across different episodes is dynamically adjusted based on a self-segmentation scheme.
arXiv Detail & Related papers (2020-04-12T03:53:53Z) - SketchDesc: Learning Local Sketch Descriptors for Multi-view
Correspondence [68.63311821718416]
We study the problem of multi-view sketch correspondence, where we take as input multiple freehand sketches with different views of the same object.
This problem is challenging since the visual features of corresponding points at different views can be very different.
We take a deep learning approach and learn a novel local sketch descriptor from data.
arXiv Detail & Related papers (2020-01-16T11:31:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.