Text2CAD: Text to 3D CAD Generation via Technical Drawings
- URL: http://arxiv.org/abs/2411.06206v1
- Date: Sat, 09 Nov 2024 15:12:06 GMT
- Title: Text2CAD: Text to 3D CAD Generation via Technical Drawings
- Authors: Mohsen Yavartanoo, Sangmin Hong, Reyhaneh Neshatavar, Kyoung Mu Lee,
- Abstract summary: Text2CAD is a novel framework that employs stable diffusion models tailored to automate the generation process.
We show that Text2CAD effectively generates technical drawings that are accurately translated into high-quality 3D CAD models.
- Score: 45.3611544056261
- License:
- Abstract: The generation of industrial Computer-Aided Design (CAD) models from user requests and specifications is crucial to enhancing efficiency in modern manufacturing. Traditional methods of CAD generation rely heavily on manual inputs and struggle with complex or non-standard designs, making them less suited for dynamic industrial needs. To overcome these challenges, we introduce Text2CAD, a novel framework that employs stable diffusion models tailored to automate the generation process and efficiently bridge the gap between user specifications in text and functional CAD models. This approach directly translates the user's textural descriptions into detailed isometric images, which are then precisely converted into orthographic views, e.g., top, front, and side, providing sufficient information to reconstruct 3D CAD models. This process not only streamlines the creation of CAD models from textual descriptions but also ensures that the resulting models uphold physical and dimensional consistency essential for practical engineering applications. Our experimental results show that Text2CAD effectively generates technical drawings that are accurately translated into high-quality 3D CAD models, showing substantial potential to revolutionize CAD automation in response to user demands.
Related papers
- CAD-Editor: A Locate-then-Infill Framework with Automated Training Data Synthesis for Text-Based CAD Editing [12.277838798842689]
We introduce emphCAD-Editor, the first framework for text-based CAD editing.
To tackle the composite nature of text-based CAD editing, we propose a locate-then-infill framework.
Experiments show that CAD-Editor achieves superior performance both quantitatively and qualitatively.
arXiv Detail & Related papers (2025-02-06T11:57:14Z) - InsTex: Indoor Scenes Stylized Texture Synthesis [81.12010726769768]
High-quality textures are crucial for 3D scenes for augmented/virtual reality (ARVR) applications.
Current methods suffer from lengthy processing times and visual artifacts.
We introduce two-stage architecture designed to generate high-quality textures for 3D scenes.
arXiv Detail & Related papers (2025-01-22T08:37:59Z) - Image2CADSeq: Computer-Aided Design Sequence and Knowledge Inference from Product Images [0.7673339435080445]
In scenarios where digital CAD files are not accessible, reverse engineering (RE) has been used to reconstruct 3D CAD models.
Recent advances have seen the rise of data-driven approaches for RE, with a primary focus on converting 3D data, such as point clouds, into 3D models in boundary representation (B-rep) format.
Our research introduces a novel data-driven approach with an Image2CADSeq neural network model.
arXiv Detail & Related papers (2025-01-09T02:36:21Z) - UrbanCAD: Towards Highly Controllable and Photorealistic 3D Vehicles for Urban Scene Simulation [46.47972242593905]
Photorealistic 3D vehicle models with high controllability are essential for autonomous driving simulation and data augmentation.
We introduce UrbanCAD, a framework that generates highly controllable and photorealistic 3D vehicle digital twins from a single urban image.
arXiv Detail & Related papers (2024-11-28T17:51:08Z) - CadVLM: Bridging Language and Vision in the Generation of Parametric CAD Sketches [24.239470848849418]
Parametric Computer-Aided Design (CAD) is central to contemporary mechanical design.
We propose CadVLM, an end-to-end vision language model for CAD generation.
arXiv Detail & Related papers (2024-09-26T01:22:29Z) - GenCAD: Image-Conditioned Computer-Aided Design Generation with
Transformer-Based Contrastive Representation and Diffusion Priors [4.485378844492069]
GenCAD is a generative model that transforms image inputs into parametric CAD command sequences.
It significantly outperforms existing state-of-the-art methods in terms of the precision and modifiability of generated 3D shapes.
arXiv Detail & Related papers (2024-09-08T23:49:11Z) - OpenECAD: An Efficient Visual Language Model for Editable 3D-CAD Design [1.481550828146527]
We fine-tuned pre-trained models to create OpenECAD models (0.55B, 0.89B, 2.4B and 3.1B)
OpenECAD models can process images of 3D designs as input and generate highly structured 2D sketches and 3D construction commands.
These outputs can be directly used with existing CAD tools' APIs to generate project files.
arXiv Detail & Related papers (2024-06-14T10:47:52Z) - Geometric Deep Learning for Computer-Aided Design: A Survey [85.79012726689511]
This survey offers a comprehensive overview of learning-based methods in computer-aided design.
It includes similarity analysis and retrieval, 2D and 3D CAD model synthesis, and CAD generation from point clouds.
It provides a complete list of benchmark datasets and their characteristics, along with open-source codes that have propelled research in this domain.
arXiv Detail & Related papers (2024-02-27T17:11:35Z) - AutoCAD: Automatically Generating Counterfactuals for Mitigating
Shortcut Learning [70.70393006697383]
We present AutoCAD, a fully automatic and task-agnostic CAD generation framework.
In this paper, we present AutoCAD, a fully automatic and task-agnostic CAD generation framework.
arXiv Detail & Related papers (2022-11-29T13:39:53Z) - GET3D: A Generative Model of High Quality 3D Textured Shapes Learned
from Images [72.15855070133425]
We introduce GET3D, a Generative model that directly generates Explicit Textured 3D meshes with complex topology, rich geometric details, and high-fidelity textures.
GET3D is able to generate high-quality 3D textured meshes, ranging from cars, chairs, animals, motorbikes and human characters to buildings.
arXiv Detail & Related papers (2022-09-22T17:16:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.