A Public Ground-Truth Dataset for Handwritten Circuit Diagram Images
- URL: http://arxiv.org/abs/2107.10373v1
- Date: Wed, 21 Jul 2021 22:10:11 GMT
- Title: A Public Ground-Truth Dataset for Handwritten Circuit Diagram Images
- Authors: Felix Thoma, Johannes Bayer, Yakun Li
- Abstract summary: The dataset consists of 1152 images of 144 circuits by 12 drafters and 48 563 annotations.
All individual electrical components are annotated with bounding boxes and one out of 45 class labels.
The geometric and taxonomic problems arising from this task as well as the classes themselves and statistics of their appearances are stated.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The development of digitization methods for line drawings (especially in the
area of electrical engineering) relies on the availability of publicly
available training and evaluation data. This paper presents such an image set
along with annotations. The dataset consists of 1152 images of 144 circuits by
12 drafters and 48 563 annotations. Each of these images depicts an electrical
circuit diagram, taken by consumer grade cameras under varying lighting
conditions and perspectives. A variety of different pencil types and surface
materials has been used. For each image, all individual electrical components
are annotated with bounding boxes and one out of 45 class labels. In order to
simplify a graph extraction process, different helper symbols like junction
points and crossovers are introduced, while texts are annotated as well. The
geometric and taxonomic problems arising from this task as well as the classes
themselves and statistics of their appearances are stated. The performance of a
standard Faster RCNN on the dataset is provided as an object detection
baseline.
Related papers
- Modular Graph Extraction for Handwritten Circuit Diagram Images [5.780326596446099]
Hand-drawn circuit diagrams are still used today in the educational domain, where they serve as an easily accessible mean for trainees and students to learn drawing this type of diagrams.
In order to harness the capabilities of digital circuit representations, automated means for extracting the electrical graph from graphics are required.
This paper describes a modular end-to-end solution on a larger, public dataset, in which approaches for the individual sub-tasks are evaluated to form a new baseline.
arXiv Detail & Related papers (2024-02-16T21:39:28Z) - Stellar: Systematic Evaluation of Human-Centric Personalized
Text-to-Image Methods [52.806258774051216]
We focus on text-to-image systems that input a single image of an individual and ground the generation process along with text describing the desired visual context.
We introduce a standardized dataset (Stellar) that contains personalized prompts coupled with images of individuals that is an order of magnitude larger than existing relevant datasets and where rich semantic ground-truth annotations are readily available.
We derive a simple yet efficient, personalized text-to-image baseline that does not require test-time fine-tuning for each subject and which sets quantitatively and in human trials a new SoTA.
arXiv Detail & Related papers (2023-12-11T04:47:39Z) - Learning Dense Correspondences between Photos and Sketches [6.2420740599184175]
Humans effortlessly grasp the connection between sketches and real-world objects, even when these sketches are far from realistic.
We introduce a new sketch-photo correspondence benchmark, $textitPSC6k$, containing 150K annotations of 6250 sketch-photo pairs across 125 object categories.
Second, we propose a self-supervised method for learning dense correspondences between sketch-photo pairs.
arXiv Detail & Related papers (2023-07-24T17:45:40Z) - Instance Segmentation Based Graph Extraction for Handwritten Circuit
Diagram Images [4.365209337828563]
This paper describes an approach for extracting both the electrical components (including their terminals and describing texts) by the means of instance segmentation and keypoint extraction.
The resulting graph extraction process consists of a simple two-step process of model inference and trivial geometric keypoint matching.
arXiv Detail & Related papers (2023-01-09T03:00:20Z) - Using a Conditional Generative Adversarial Network to Control the
Statistical Characteristics of Generated Images for IACT Data Analysis [55.41644538483948]
We divide images into several classes according to the value of some property of the image, and then specify the required class when generating new images.
In the case of images from Imaging Atmospheric Cherenkov Telescopes (IACTs), an important property is the total brightness of all image pixels (image size)
We used a cGAN technique to generate images similar to whose obtained in the TAIGA-IACT experiment.
arXiv Detail & Related papers (2022-11-28T22:30:33Z) - I Know What You Draw: Learning Grasp Detection Conditioned on a Few
Freehand Sketches [74.63313641583602]
We propose a method to generate a potential grasp configuration relevant to the sketch-depicted objects.
Our model is trained and tested in an end-to-end manner which is easy to be implemented in real-world applications.
arXiv Detail & Related papers (2022-05-09T04:23:36Z) - Learning to generate line drawings that convey geometry and semantics [22.932131011984513]
This paper presents an unpaired method for creating line drawings from photographs.
We observe that line drawings are encodings of scene information and seek to convey 3D shape and semantic meaning.
We introduce a geometry loss which predicts depth information from the image features of a line drawing, and a semantic loss which matches the CLIP features of a line drawing with its corresponding photograph.
arXiv Detail & Related papers (2022-03-23T19:27:41Z) - Using Text to Teach Image Retrieval [47.72498265721957]
We build on the concept of image manifold to represent the feature space of images, learned via neural networks, as a graph.
We augment the manifold samples with geometrically aligned text, thereby using a plethora of sentences to teach us about images.
The experimental results show that the joint embedding manifold is a robust representation, allowing it to be a better basis to perform image retrieval.
arXiv Detail & Related papers (2020-11-19T16:09:14Z) - Semantically Tied Paired Cycle Consistency for Any-Shot Sketch-based
Image Retrieval [55.29233996427243]
Low-shot sketch-based image retrieval is an emerging task in computer vision.
In this paper, we address any-shot, i.e. zero-shot and few-shot, sketch-based image retrieval (SBIR) tasks.
For solving these tasks, we propose a semantically aligned cycle-consistent generative adversarial network (SEM-PCYC)
Our results demonstrate a significant boost in any-shot performance over the state-of-the-art on the extended version of the Sketchy, TU-Berlin and QuickDraw datasets.
arXiv Detail & Related papers (2020-06-20T22:43:53Z) - SketchDesc: Learning Local Sketch Descriptors for Multi-view
Correspondence [68.63311821718416]
We study the problem of multi-view sketch correspondence, where we take as input multiple freehand sketches with different views of the same object.
This problem is challenging since the visual features of corresponding points at different views can be very different.
We take a deep learning approach and learn a novel local sketch descriptor from data.
arXiv Detail & Related papers (2020-01-16T11:31:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.