LoGoColor: Local-Global 3D Colorization for 360° Scenes
- URL: http://arxiv.org/abs/2512.09278v1
- Date: Wed, 10 Dec 2025 03:03:38 GMT
- Title: LoGoColor: Local-Global 3D Colorization for 360° Scenes
- Authors: Yeonjin Chang, Juhwan Cho, Seunghyeon Seo, Wonsik Shin, Nojun Kwak,
- Abstract summary: Single-channel 3D reconstruction is widely used in fields such as robotics and medical imaging.<n>Recent 3D colorization studies address this problem by distilling 2D image colorization models.<n>We propose LoGoColor, a pipeline designed to preserve color diversity by eliminating the guidance-averaging process.
- Score: 29.177641673340137
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Single-channel 3D reconstruction is widely used in fields such as robotics and medical imaging. While this line of work excels at reconstructing 3D geometry, the outputs are not colored 3D models, thus 3D colorization is required for visualization. Recent 3D colorization studies address this problem by distilling 2D image colorization models. However, these approaches suffer from an inherent inconsistency of 2D image models. This results in colors being averaged during training, leading to monotonous and oversimplified results, particularly in complex 360° scenes. In contrast, we aim to preserve color diversity by generating a new set of consistently colorized training views, thereby bypassing the averaging process. Nevertheless, eliminating the averaging process introduces a new challenge: ensuring strict multi-view consistency across these colorized views. To achieve this, we propose LoGoColor, a pipeline designed to preserve color diversity by eliminating this guidance-averaging process with a `Local-Global' approach: we partition the scene into subscenes and explicitly tackle both inter-subscene and intra-subscene consistency using a fine-tuned multi-view diffusion model. We demonstrate that our method achieves quantitatively and qualitatively more consistent and plausible 3D colorization on complex 360° scenes than existing methods, and validate its superior color diversity using a novel Color Diversity Index.
Related papers
- VIRGi: View-dependent Instant Recoloring of 3D Gaussians Splats [53.602701067430075]
We introduce VIRGi, a novel approach for rapidly editing the color of scenes modeled by 3DGS.<n>By fine-tuning the weights of a single user, the color edits are seamlessly propagated to the entire scene in just two seconds.<n>An exhaustive validation on diverse datasets demonstrates significant quantitative and qualitative advancements over competitors.
arXiv Detail & Related papers (2026-03-03T13:41:17Z) - Color3D: Controllable and Consistent 3D Colorization with Personalized Colorizer [58.94607850223466]
We present Color3D, a highly adaptable framework for colorizing both static and dynamic 3D scenes from monochromatic inputs.<n>Our approach is able to preserve color diversity and steerability while ensuring cross-view and cross-time consistency.
arXiv Detail & Related papers (2025-10-11T10:21:19Z) - Follow-Your-Color: Multi-Instance Sketch Colorization [44.72374445094054]
Follow-Your-Color is a diffusion-based framework for multi-instance sketch colorization.<n>Our model critically automates the colorization process with zero manual adjustments.
arXiv Detail & Related papers (2025-03-21T08:53:14Z) - Coherent 3D Scene Diffusion From a Single RGB Image [68.31336308924477]
We present a novel diffusion-based approach for coherent 3D scene reconstruction from a single RGB image.<n>Our method simultaneously denoises the 3D poses and geometries of all objects within the scene.<n>By framing the task of single RGB image 3D scene reconstruction as a conditional diffusion process, our approach surpasses current state-of-the-art methods.
arXiv Detail & Related papers (2024-12-13T17:26:45Z) - MVD-Fusion: Single-view 3D via Depth-consistent Multi-view Generation [54.27399121779011]
We present MVD-Fusion: a method for single-view 3D inference via generative modeling of multi-view-consistent RGB-D images.
We show that our approach can yield more accurate synthesis compared to recent state-of-the-art, including distillation-based 3D inference and prior multi-view generation methods.
arXiv Detail & Related papers (2024-04-04T17:59:57Z) - Learning Naturally Aggregated Appearance for Efficient 3D Editing [90.57414218888536]
We learn the color field as an explicit 2D appearance aggregation, also called canonical image.<n>We complement the canonical image with a projection field that maps 3D points onto 2D pixels for texture query.<n>Our approach demonstrates remarkable efficiency by being at least 20 times faster per edit compared to existing NeRF-based editing methods.
arXiv Detail & Related papers (2023-12-11T18:59:31Z) - UNeR3D: Versatile and Scalable 3D RGB Point Cloud Generation from 2D
Images in Unsupervised Reconstruction [2.7848140839111903]
UNeR3D sets a new standard for generating detailed 3D reconstructions solely from 2D views.
Our model significantly cuts down the training costs tied to supervised approaches.
UNeR3D ensures seamless color transitions, enhancing visual fidelity.
arXiv Detail & Related papers (2023-12-10T15:18:55Z) - ChromaDistill: Colorizing Monochrome Radiance Fields with Knowledge Distillation [23.968181738235266]
We present a method for colorized novel views from input grayscale multi-view images.<n>We propose a distillation-based method that transfers color from these networks trained on natural images to the target 3D representation.<n>Our method is agnostic to the underlying 3D representation and easily generalizable to NeRF and 3DGS methods.
arXiv Detail & Related papers (2023-09-14T12:30:48Z) - Cross-Modal 3D Shape Generation and Manipulation [62.50628361920725]
We propose a generic multi-modal generative model that couples the 2D modalities and implicit 3D representations through shared latent spaces.
We evaluate our framework on two representative 2D modalities of grayscale line sketches and rendered color images.
arXiv Detail & Related papers (2022-07-24T19:22:57Z) - SPSG: Self-Supervised Photometric Scene Generation from RGB-D Scans [34.397726189729994]
SPSG is a novel approach to generate high-quality, colored 3D models of scenes from RGB-D scan observations.
Our self-supervised approach learns to jointly inpaint geometry and color by correlating an incomplete RGB-D scan with a more complete version of that scan.
arXiv Detail & Related papers (2020-06-25T18:58:23Z) - 3D Photography using Context-aware Layered Depth Inpainting [50.66235795163143]
We propose a method for converting a single RGB-D input image into a 3D photo.
A learning-based inpainting model synthesizes new local color-and-depth content into the occluded region.
The resulting 3D photos can be efficiently rendered with motion parallax.
arXiv Detail & Related papers (2020-04-09T17:59:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.