Mesh Mamba: A Unified State Space Model for Saliency Prediction in Non-Textured and Textured Meshes
- URL: http://arxiv.org/abs/2504.01466v2
- Date: Wed, 09 Apr 2025 08:35:39 GMT
- Title: Mesh Mamba: A Unified State Space Model for Saliency Prediction in Non-Textured and Textured Meshes
- Authors: Kaiwei Zhang, Dandan Zhu, Xiongkuo Min, Guangtao Zhai,
- Abstract summary: Mesh saliency enhances the adaptability of 3D vision by identifying and emphasizing regions that naturally attract visual attention.<n>We introduce mesh Mamba, a unified saliency prediction model based on a state space model (SSM)<n>Mesh Mamba effectively analyzes the geometric structure of the mesh while seamlessly incorporating texture features into the topological framework.
- Score: 50.23625950905638
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mesh saliency enhances the adaptability of 3D vision by identifying and emphasizing regions that naturally attract visual attention. To investigate the interaction between geometric structure and texture in shaping visual attention, we establish a comprehensive mesh saliency dataset, which is the first to systematically capture the differences in saliency distribution under both textured and non-textured visual conditions. Furthermore, we introduce mesh Mamba, a unified saliency prediction model based on a state space model (SSM), designed to adapt across various mesh types. Mesh Mamba effectively analyzes the geometric structure of the mesh while seamlessly incorporating texture features into the topological framework, ensuring coherence throughout appearance-enhanced modeling. More importantly, by subgraph embedding and a bidirectional SSM, the model enables global context modeling for both local geometry and texture, preserving the topological structure and improving the understanding of visual details and structural complexity. Through extensive theoretical and empirical validation, our model not only improves performance across various mesh types but also demonstrates high scalability and versatility, particularly through cross validations of various visual features.
Related papers
- MARS: Mesh AutoRegressive Model for 3D Shape Detailization [85.95365919236212]
We introduce MARS, a novel approach for 3D shape detailization.<n>We propose a mesh autoregressive model capable of generating such latent representations through next-LOD token prediction.<n>Experiments conducted on the challenging 3D Shape Detailization benchmark demonstrate that our proposed MARS model achieves state-of-the-art performance.
arXiv Detail & Related papers (2025-02-17T03:12:16Z) - Textured Mesh Saliency: Bridging Geometry and Texture for Human Perception in 3D Graphics [50.23625950905638]
We present a new dataset for textured mesh saliency, created through an innovative eye-tracking experiment in a six degrees of freedom (6-DOF) VR environment.<n>Our proposed model predicts saliency maps for textured mesh surfaces by treating each triangular face as an individual unit and assigning a saliency density value to reflect the importance of each local surface region.
arXiv Detail & Related papers (2024-12-11T08:27:33Z) - HybridMQA: Exploring Geometry-Texture Interactions for Colored Mesh Quality Assessment [7.526258700061012]
Mesh quality assessment (MQA) models play a critical role in the design, optimization, and evaluation of mesh operation systems.<n>We introduce HybridMQA, a hybrid full-reference colored MQA framework that integrates model-based and projection-based approaches.<n>Our method employs graph learning to extract detailed 3D representations, which are then projected to 2D using a novel feature rendering process.
arXiv Detail & Related papers (2024-12-02T21:35:33Z) - Geometry Distributions [51.4061133324376]
We propose a novel geometric data representation that models geometry as distributions.
Our approach uses diffusion models with a novel network architecture to learn surface point distributions.
We evaluate our representation qualitatively and quantitatively across various object types, demonstrating its effectiveness in achieving high geometric fidelity.
arXiv Detail & Related papers (2024-11-25T04:06:48Z) - DreamPolish: Domain Score Distillation With Progressive Geometry Generation [66.94803919328815]
We introduce DreamPolish, a text-to-3D generation model that excels in producing refined geometry and high-quality textures.
In the geometry construction phase, our approach leverages multiple neural representations to enhance the stability of the synthesis process.
In the texture generation phase, we introduce a novel score distillation objective, namely domain score distillation (DSD), to guide neural representations toward such a domain.
arXiv Detail & Related papers (2024-11-03T15:15:01Z) - A Generative Machine Learning Model for Material Microstructure 3D
Reconstruction and Performance Evaluation [4.169915659794567]
The dimensional extension from 2D to 3D is viewed as a highly challenging inverse problem from the current technological perspective.
A novel generative model that integrates the multiscale properties of U-net with and the generative capabilities of GAN has been proposed.
The model's accuracy is further improved by combining the image regularization loss with the Wasserstein distance loss.
arXiv Detail & Related papers (2024-02-24T13:42:34Z) - Single Mesh Diffusion Models with Field Latents for Texture Generation [18.78126579775479]
We introduce a framework for intrinsic latent diffusion models operating directly on the surfaces of 3D shapes.
We consider a single-textured-mesh paradigm, where our models are trained to generate variations of a given texture on a mesh.
Our models can also be adapted for user-controlled editing tasks such as inpainting and label-guided generation.
arXiv Detail & Related papers (2023-12-14T18:59:36Z) - T1: Scaling Diffusion Probabilistic Fields to High-Resolution on Unified
Visual Modalities [69.16656086708291]
Diffusion Probabilistic Field (DPF) models the distribution of continuous functions defined over metric spaces.
We propose a new model comprising of a view-wise sampling algorithm to focus on local structure learning.
The model can be scaled to generate high-resolution data while unifying multiple modalities.
arXiv Detail & Related papers (2023-05-24T03:32:03Z) - Low-Light Image Enhancement via Structure Modeling and Guidance [13.375551436077423]
This paper proposes a new framework for low-light image enhancement by simultaneously conducting the appearance and structure modeling.
The structure modeling in our framework is implemented as the edge detection in low-light images.
To improve the appearance modeling, which is implemented with a simple U-Net, a novel structure-guided enhancement module is proposed.
arXiv Detail & Related papers (2023-05-10T02:08:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.