Sketch2Stress: Sketching with Structural Stress Awareness
- URL: http://arxiv.org/abs/2306.05911v2
- Date: Mon, 11 Dec 2023 06:51:51 GMT
- Title: Sketch2Stress: Sketching with Structural Stress Awareness
- Authors: Deng Yu, Chufeng Xiao, Manfred Lau, and Hongbo Fu
- Abstract summary: We introduce Sketch2Stress that allows users to perform structural analysis of desired objects at the sketching stage.
With the specially-designed two-branch generative-adversarial framework, it automatically predicts a normal map and a corresponding structural stress map.
We demonstrate the effectiveness and practicality of our system with extensive experiments and user studies.
- Score: 15.842017507987892
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the process of product design and digital fabrication, the structural
analysis of a designed prototype is a fundamental and essential step. However,
such a step is usually invisible or inaccessible to designers at the early
sketching phase. This limits the user's ability to consider a shape's physical
properties and structural soundness. To bridge this gap, we introduce a novel
approach Sketch2Stress that allows users to perform structural analysis of
desired objects at the sketching stage. This method takes as input a 2D
freehand sketch and one or multiple locations of user-assigned external forces.
With the specially-designed two-branch generative-adversarial framework, it
automatically predicts a normal map and a corresponding structural stress map
distributed over the user-sketched underlying object. In this way, our method
empowers designers to easily examine the stress sustained everywhere and
identify potential problematic regions of their sketched object. Furthermore,
combined with the predicted normal map, users are able to conduct a region-wise
structural analysis efficiently by aggregating the stress effects of multiple
forces in the same direction. Finally, we demonstrate the effectiveness and
practicality of our system with extensive experiments and user studies.
Related papers
- CustomSketching: Sketch Concept Extraction for Sketch-based Image
Synthesis and Editing [21.12815542848095]
Personalization techniques for large text-to-image (T2I) models allow users to incorporate new concepts from reference images.
Existing methods primarily rely on textual descriptions, leading to limited control over customized images.
We identify sketches as an intuitive and versatile representation that can facilitate such control.
arXiv Detail & Related papers (2024-02-27T15:52:59Z) - SENS: Part-Aware Sketch-based Implicit Neural Shape Modeling [124.3266213819203]
We present SENS, a novel method for generating and editing 3D models from hand-drawn sketches.
S SENS analyzes the sketch and encodes its parts into ViT patch encoding.
S SENS supports refinement via part reconstruction, allowing for nuanced adjustments and artifact removal.
arXiv Detail & Related papers (2023-06-09T17:50:53Z) - Sketch2Saliency: Learning to Detect Salient Objects from Human Drawings [99.9788496281408]
We study how sketches can be used as a weak label to detect salient objects present in an image.
To accomplish this, we introduce a photo-to-sketch generation model that aims to generate sequential sketch coordinates corresponding to a given visual photo.
Tests prove our hypothesis and delineate how our sketch-based saliency detection model gives a competitive performance compared to the state-of-the-art.
arXiv Detail & Related papers (2023-03-20T23:46:46Z) - AI Art in Architecture [0.6853165736531939]
Recent diffusion-based AI art platforms are able to create impressive images from simple text descriptions.
This is also true for early stages of architectural design with multiple stages of ideation, sketching and modelling.
We research the applicability of the platforms Midjourney, DALL-E 2 and StableDiffusion to a series of common use cases in architectural design.
arXiv Detail & Related papers (2022-12-19T12:24:14Z) - Vitruvio: 3D Building Meshes via Single Perspective Sketches [0.8001739956625484]
We introduce the first deep learning method focused only on buildings that aim to convert a single sketch to a 3D building mesh: Vitruvio.
First, it accelerates the inference process by more than 26% (from 0.5s to 0.37s)
Second, it increases the reconstruction accuracy (measured by the Chamfer Distance) by 18%.
arXiv Detail & Related papers (2022-10-24T22:24:58Z) - I Know What You Draw: Learning Grasp Detection Conditioned on a Few
Freehand Sketches [74.63313641583602]
We propose a method to generate a potential grasp configuration relevant to the sketch-depicted objects.
Our model is trained and tested in an end-to-end manner which is easy to be implemented in real-world applications.
arXiv Detail & Related papers (2022-05-09T04:23:36Z) - Sketch2PQ: Freeform Planar Quadrilateral Mesh Design via a Single Sketch [36.10997511325458]
We present a novel sketch-based system to bridge the concept design and digital modeling of freeform roof-like shapes.
Our system allows the user to sketch the surface boundary and contour lines under axonometric projection.
We propose a deep neural network to infer in real-time the underlying surface shape along with a dense conjugate direction field.
arXiv Detail & Related papers (2022-01-23T21:09:59Z) - Segmentation and Analysis of a Sketched Truss Frame Using Morphological
Image Processing Techniques [0.0]
Development of computational tools to analyze and assess the building capacities has had a major impact in civil engineering.
One of the difficulties and the most time consuming steps involved in the structural modeling is defining the geometry of the structure to provide the analysis.
This paper is dedicated to the development of a methodology to automate analysis of a hand sketched or computer generated truss frame drawn on a piece of paper.
arXiv Detail & Related papers (2020-09-28T08:50:18Z) - 3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure
Prior [50.73148041205675]
The goal of the Semantic Scene Completion (SSC) task is to simultaneously predict a completed 3D voxel representation of volumetric occupancy and semantic labels of objects in the scene from a single-view observation.
We propose to devise a new geometry-based strategy to embed depth information with low-resolution voxel representation.
Our proposed geometric embedding works better than the depth feature learning from habitual SSC frameworks.
arXiv Detail & Related papers (2020-03-31T09:33:46Z) - Deep Self-Supervised Representation Learning for Free-Hand Sketch [51.101565480583304]
We tackle the problem of self-supervised representation learning for free-hand sketches.
Key for the success of our self-supervised learning paradigm lies with our sketch-specific designs.
We show that the proposed approach outperforms the state-of-the-art unsupervised representation learning methods.
arXiv Detail & Related papers (2020-02-03T16:28:29Z) - Deep Plastic Surgery: Robust and Controllable Image Editing with
Human-Drawn Sketches [133.01690754567252]
Sketch-based image editing aims to synthesize and modify photos based on the structural information provided by the human-drawn sketches.
Deep Plastic Surgery is a novel, robust and controllable image editing framework that allows users to interactively edit images using hand-drawn sketch inputs.
arXiv Detail & Related papers (2020-01-09T08:57:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.