Digital Modeling for Everyone: Exploring How Novices Approach
Voice-Based 3D Modeling
- URL: http://arxiv.org/abs/2307.04481v2
- Date: Wed, 30 Aug 2023 11:17:03 GMT
- Title: Digital Modeling for Everyone: Exploring How Novices Approach
Voice-Based 3D Modeling
- Authors: Giuseppe Desolda (1), Andrea Esposito (1), Florian M\"uller (2),
Sebastian Feger (2) ((1) University of Bari Aldo Moro, Bari, Italy, (2) LMU
Munich, Munich, Germany)
- Abstract summary: We explore novice mental models in voice-based 3D modeling by conducting a high-fidelity Wizard of Oz study with 22 participants.
We conclude with design implications for voice assistants.
For example, they have to: deal with vague, incomplete and wrong commands; provide a set of straightforward commands to shape simple and composite objects; and offer different strategies to select 3D objects.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Manufacturing tools like 3D printers have become accessible to the wider
society, making the promise of digital fabrication for everyone seemingly
reachable. While the actual manufacturing process is largely automated today,
users still require knowledge of complex design applications to produce
ready-designed objects and adapt them to their needs or design new objects from
scratch. To lower the barrier to the design and customization of personalized
3D models, we explored novice mental models in voice-based 3D modeling by
conducting a high-fidelity Wizard of Oz study with 22 participants. We
performed a thematic analysis of the collected data to understand how the
mental model of novices translates into voice-based 3D modeling. We conclude
with design implications for voice assistants. For example, they have to: deal
with vague, incomplete and wrong commands; provide a set of straightforward
commands to shape simple and composite objects; and offer different strategies
to select 3D objects.
Related papers
- Scalable Cloud-Native Pipeline for Efficient 3D Model Reconstruction from Monocular Smartphone Images [9.61065600471628]
We present a novel cloud-native pipeline that can automatically reconstruct 3D models from monocular 2D images captured using a smartphone camera.
Our solution produces a reusable 3D model, with embedded materials and textures, exportable and customizable in any external software or 3D engine.
arXiv Detail & Related papers (2024-09-28T11:15:26Z) - Chat-Edit-3D: Interactive 3D Scene Editing via Text Prompts [76.73043724587679]
We propose a dialogue-based 3D scene editing approach, termed CE3D.
Hash-Atlas represents 3D scene views, which transfers the editing of 3D scenes onto 2D atlas images.
Results demonstrate that CE3D effectively integrates multiple visual models to achieve diverse editing visual effects.
arXiv Detail & Related papers (2024-07-09T13:24:42Z) - Shaping Realities: Enhancing 3D Generative AI with Fabrication Constraints [36.65470465480772]
Generative AI tools are becoming more prevalent in 3D modeling, enabling users to manipulate or create new models with text or images as inputs.
These methods focus on the aesthetic quality of the 3D models, refining them to look similar to the prompts provided by the user.
When creating 3D models intended for fabrication, designers need to trade-off the aesthetic qualities of a 3D model with their intended physical properties.
arXiv Detail & Related papers (2024-04-15T21:22:57Z) - SUGAR: Pre-training 3D Visual Representations for Robotics [85.55534363501131]
We introduce a novel 3D pre-training framework for robotics named SUGAR.
SUGAR captures semantic, geometric and affordance properties of objects through 3D point clouds.
We show that SUGAR's 3D representation outperforms state-of-the-art 2D and 3D representations.
arXiv Detail & Related papers (2024-04-01T21:23:03Z) - En3D: An Enhanced Generative Model for Sculpting 3D Humans from 2D
Synthetic Data [36.51674664590734]
We present En3D, an enhanced izable scheme for high-qualityd 3D human avatars.
Unlike previous works that rely on scarce 3D datasets or limited 2D collections with imbalance viewing angles and pose priors, our approach aims to develop a zero-shot 3D capable of producing 3D humans.
arXiv Detail & Related papers (2024-01-02T12:06:31Z) - 3D-GPT: Procedural 3D Modeling with Large Language Models [47.72968643115063]
We introduce 3D-GPT, a framework utilizing large language models(LLMs) for instruction-driven 3D modeling.
3D-GPT positions LLMs as proficient problem solvers, dissecting the procedural 3D modeling tasks into accessible segments and appointing the apt agent for each task.
Our empirical investigations confirm that 3D-GPT not only interprets and executes instructions, delivering reliable results but also collaborates effectively with human designers.
arXiv Detail & Related papers (2023-10-19T17:41:48Z) - CLIP-Guided Vision-Language Pre-training for Question Answering in 3D
Scenes [68.61199623705096]
We design a novel 3D pre-training Vision-Language method that helps a model learn semantically meaningful and transferable 3D scene point cloud representations.
We inject the representational power of the popular CLIP model into our 3D encoder by aligning the encoded 3D scene features with the corresponding 2D image and text embeddings.
We evaluate our model's 3D world reasoning capability on the downstream task of 3D Visual Question Answering.
arXiv Detail & Related papers (2023-04-12T16:52:29Z) - 3D Neural Scene Representations for Visuomotor Control [78.79583457239836]
We learn models for dynamic 3D scenes purely from 2D visual observations.
A dynamics model, constructed over the learned representation space, enables visuomotor control for challenging manipulation tasks.
arXiv Detail & Related papers (2021-07-08T17:49:37Z) - Interactive Annotation of 3D Object Geometry using 2D Scribbles [84.51514043814066]
In this paper, we propose an interactive framework for annotating 3D object geometry from point cloud data and RGB imagery.
Our framework targets naive users without artistic or graphics expertise.
arXiv Detail & Related papers (2020-08-24T21:51:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.