RADAM: Texture Recognition through Randomized Aggregated Encoding of
Deep Activation Maps
- URL: http://arxiv.org/abs/2303.04554v1
- Date: Wed, 8 Mar 2023 13:09:03 GMT
- Title: RADAM: Texture Recognition through Randomized Aggregated Encoding of
Deep Activation Maps
- Authors: Leonardo Scabini, Kallil M. Zielinski, Lucas C. Ribas, Wesley N.
Gon\c{c}alves, Bernard De Baets, Odemir M. Bruno
- Abstract summary: We propose a new method named textbfRandom encoding of textbfAggregated textbfDeep textbfActivation textbfMaps (RADAM)
RADAM extracts rich texture representations without ever changing the backbone.
Our results suggest that pre-trained backbones may not require additional fine-tuning for texture recognition if their learned representations are better encoded.
- Score: 12.35895883721336
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Texture analysis is a classical yet challenging task in computer vision for
which deep neural networks are actively being applied. Most approaches are
based on building feature aggregation modules around a pre-trained backbone and
then fine-tuning the new architecture on specific texture recognition tasks.
Here we propose a new method named \textbf{R}andom encoding of
\textbf{A}ggregated \textbf{D}eep \textbf{A}ctivation \textbf{M}aps (RADAM)
which extracts rich texture representations without ever changing the backbone.
The technique consists of encoding the output at different depths of a
pre-trained deep convolutional network using a Randomized Autoencoder (RAE).
The RAE is trained locally to each image using a closed-form solution, and its
decoder weights are used to compose a 1-dimensional texture representation that
is fed into a linear SVM. This means that no fine-tuning or backpropagation is
needed. We explore RADAM on several texture benchmarks and achieve
state-of-the-art results with different computational budgets. Our results
suggest that pre-trained backbones may not require additional fine-tuning for
texture recognition if their learned representations are better encoded.
Related papers
- Neural Graphics Texture Compression Supporting Random Access [34.974631096947284]
We introduce a novel approach to texture set compression that integrates traditional GPU texture representation and NIC techniques.
We propose an asymmetric auto-encoder framework that employs a convolutional encoder to capture detailed information in a bottleneck-latent space.
Experimental results demonstrate that this approach provides much better results than conventional texture compression.
arXiv Detail & Related papers (2024-05-06T19:44:13Z) - SceneTex: High-Quality Texture Synthesis for Indoor Scenes via Diffusion
Priors [49.03627933561738]
SceneTex is a novel method for generating high-quality and style-consistent textures for indoor scenes using depth-to-image diffusion priors.
SceneTex enables various and accurate texture synthesis for 3D-FRONT scenes, demonstrating significant improvements in visual quality and prompt fidelity over the prior texture generation methods.
arXiv Detail & Related papers (2023-11-28T22:49:57Z) - Not All Image Regions Matter: Masked Vector Quantization for
Autoregressive Image Generation [78.13793505707952]
Existing autoregressive models follow the two-stage generation paradigm that first learns a codebook in the latent space for image reconstruction and then completes the image generation autoregressively based on the learned codebook.
We propose a novel two-stage framework, which consists of Masked Quantization VAE (MQ-VAE) Stack model from modeling redundancy.
arXiv Detail & Related papers (2023-05-23T02:15:53Z) - TMO: Textured Mesh Acquisition of Objects with a Mobile Device by using
Differentiable Rendering [54.35405028643051]
We present a new pipeline for acquiring a textured mesh in the wild with a single smartphone.
Our method first introduces an RGBD-aided structure from motion, which can yield filtered depth maps.
We adopt the neural implicit surface reconstruction method, which allows for high-quality mesh.
arXiv Detail & Related papers (2023-03-27T10:07:52Z) - PixMIM: Rethinking Pixel Reconstruction in Masked Image Modeling [83.67628239775878]
Masked Image Modeling (MIM) has achieved promising progress with the advent of Masked Autoencoders (MAE) and BEiT.
This paper undertakes a fundamental analysis of MIM from the perspective of pixel reconstruction.
We propose a remarkably simple and effective method, ourmethod, that entails two strategies.
arXiv Detail & Related papers (2023-03-04T13:38:51Z) - Joint Learning of Deep Texture and High-Frequency Features for
Computer-Generated Image Detection [24.098604827919203]
We propose a joint learning strategy with deep texture and high-frequency features for CG image detection.
A semantic segmentation map is generated to guide the affine transformation operation.
The combination of the original image and the high-frequency components of the original and rendered images are fed into a multi-branch neural network equipped with attention mechanisms.
arXiv Detail & Related papers (2022-09-07T17:30:40Z) - Reconstructing vehicles from orthographic drawings using deep neural
networks [0.0]
The paper proposes a system based on pixel-aligned implicit functions (PIFu) and develops an advanced sampling strategy to generate signed distance samples.
The trained neural network generalizes well to real-world inputs and creates plausible and detailed reconstructions.
arXiv Detail & Related papers (2022-06-14T12:32:32Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - Over-and-Under Complete Convolutional RNN for MRI Reconstruction [57.95363471940937]
Recent deep learning-based methods for MR image reconstruction usually leverage a generic auto-encoder architecture.
We propose an Over-and-Under Complete Convolu?tional Recurrent Neural Network (OUCR), which consists of an overcomplete and an undercomplete Convolutional Recurrent Neural Network(CRNN)
The proposed method achieves significant improvements over the compressed sensing and popular deep learning-based methods with less number of trainable parameters.
arXiv Detail & Related papers (2021-06-16T15:56:34Z) - On Demand Solid Texture Synthesis Using Deep 3D Networks [3.1542695050861544]
This paper describes a novel approach for on demand texture synthesis based on a deep learning framework.
A generative network is trained to synthesize coherent portions of solid textures of arbitrary sizes.
The synthesized volumes have good visual results that are at least equivalent to the state-of-the-art patch based approaches.
arXiv Detail & Related papers (2020-01-13T20:59:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.