An attempt to generate new bridge types from latent space of variational
autoencoder
- URL: http://arxiv.org/abs/2311.03380v2
- Date: Mon, 1 Jan 2024 09:26:56 GMT
- Title: An attempt to generate new bridge types from latent space of variational
autoencoder
- Authors: Hongjun Zhang
- Abstract summary: Variational autoencoder can combine two bridge types on the basis of the original of human into one that is a new bridge type.
Generative artificial intelligence technology can assist bridge designers in bridge-type innovation, and can be used as copilot.
- Score: 2.05750372679553
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Try to generate new bridge types using generative artificial intelligence
technology. The grayscale images of the bridge facade with the change of
component width was rendered by 3dsMax animation software, and then the OpenCV
module performed an appropriate amount of geometric transformation (rotation,
horizontal scale, vertical scale) to obtain the image dataset of three-span
beam bridge, arch bridge, cable-stayed bridge and suspension bridge. Based on
Python programming language, TensorFlow and Keras deep learning platform
framework, variational autoencoder was constructed and trained, and
low-dimensional bridge-type latent space that is convenient for vector
operations was obtained. Variational autoencoder can combine two bridge types
on the basis of the original of human into one that is a new bridge type.
Generative artificial intelligence technology can assist bridge designers in
bridge-type innovation, and can be used as copilot.
Related papers
- Instance Segmentation of Reinforced Concrete Bridges with Synthetic Point Clouds [0.0]
National Bridge Inspection Standards require detailed element-level bridge inspections.
Traditionally, inspectors manually assign ratings by rating structural components based on damage.
We propose a novel approach for generating synthetic data using three distinct methods.
arXiv Detail & Related papers (2024-09-24T18:28:41Z) - An attempt to generate new bridge types from latent space of denoising
diffusion Implicit model [2.05750372679553]
Use denoising diffusion implicit model for bridge-type innovation.
Process of adding noise and denoising to an image can be likened to the process of a corpse rotting and a detective restoring the scene of a victim being killed, to help beginners understand.
arXiv Detail & Related papers (2024-02-11T08:54:37Z) - An attempt to generate new bridge types from latent space of PixelCNN [2.05750372679553]
PixelCNN can capture the statistical structure of the images and calculate the probability distribution of the next pixel.
From the obtained latent space sampling, new bridge types different from the training dataset can be generated.
arXiv Detail & Related papers (2024-01-11T15:06:25Z) - An attempt to generate new bridge types from latent space of generative
adversarial network [2.05750372679553]
Symmetric structured image dataset of three-span beam bridge, arch bridge, cable-stayed bridge and suspension bridge are used.
Based on Python programming language, and Keras deep learning platform framework, generative adversarial network is constructed and trained.
arXiv Detail & Related papers (2024-01-01T08:46:29Z) - BridgeData V2: A Dataset for Robot Learning at Scale [73.86688388408021]
BridgeData V2 is a large and diverse dataset of robotic manipulation behaviors.
It contains 60,096 trajectories collected across 24 environments on a publicly available low-cost robot.
arXiv Detail & Related papers (2023-08-24T17:41:20Z) - SDFusion: Multimodal 3D Shape Completion, Reconstruction, and Generation [89.47132156950194]
We present a novel framework built to simplify 3D asset generation for amateur users.
Our method supports a variety of input modalities that can be easily provided by a human.
Our model can combine all these tasks into one swiss-army-knife tool.
arXiv Detail & Related papers (2022-12-08T18:59:05Z) - BridgeTower: Building Bridges Between Encoders in Vision-Language Representation Learning [91.93547262073213]
Vision-Language (VL) models with the Two-Tower architecture have dominated visual representation learning in recent years.
We propose BridgeTower, which introduces multiple bridge layers that build a connection between the top layers of uni-modal encoders and each layer of the cross-modal encoder.
BridgeTower achieves an accuracy of 78.73%, outperforming the previous state-of-the-art model METER by 1.09% with the same pre-training data.
arXiv Detail & Related papers (2022-06-17T09:42:35Z) - BANMo: Building Animatable 3D Neural Models from Many Casual Videos [135.64291166057373]
We present BANMo, a method that requires neither a specialized sensor nor a pre-defined template shape.
Banmo builds high-fidelity, articulated 3D models from many monocular casual videos in a differentiable rendering framework.
On real and synthetic datasets, BANMo shows higher-fidelity 3D reconstructions than prior works for humans and animals.
arXiv Detail & Related papers (2021-12-23T18:30:31Z) - Learning to Design and Construct Bridge without Blueprint [20.524052738716435]
We study a new challenging assembly task, designing and constructing a bridge without a blueprint.
In this task, the robot needs to first design a feasible bridge architecture for arbitrarily wide cliffs and then manipulate the blocks reliably to construct a stable bridge according to the proposed design.
At the high level, the system learns a bridge blueprint policy in a physical simulator using deep reinforcement learning and curriculum learning.
For low-level control, we implement a motion-planning-based policy for real-robot motion control, which can be directly combined with a trained blueprint policy for real-world bridge construction without tuning.
arXiv Detail & Related papers (2021-08-05T08:17:22Z) - Uformer: A General U-Shaped Transformer for Image Restoration [47.60420806106756]
We build a hierarchical encoder-decoder network using the Transformer block for image restoration.
Experiments on several image restoration tasks demonstrate the superiority of Uformer.
arXiv Detail & Related papers (2021-06-06T12:33:22Z) - Multiscale Vision Transformers [79.76412415996892]
We present Multiscale Vision Transformers (MViT) for video and image recognition, by connecting the seminal idea of multiscale feature hierarchies with transformer models.
We evaluate this fundamental architectural prior for modeling the dense nature of visual signals for a variety of video recognition tasks.
arXiv Detail & Related papers (2021-04-22T17:59:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.