Multi-level Data Representation For Training Deep Helmholtz Machines
- URL: http://arxiv.org/abs/2210.14855v1
- Date: Wed, 26 Oct 2022 16:55:40 GMT
- Title: Multi-level Data Representation For Training Deep Helmholtz Machines
- Authors: Jose Miguel Ramos, Luis Sa-Couto and Andreas Wichert
- Abstract summary: We guide the learning of a biologically plausible generative model called the Helmholtz Machine in complex search spaces using the Human Image Perception mechanism.
We propose to overcome this problem, by providing the network's hidden layers with visual queues at different resolutions using a Multi-level Data representation.
The results on several image datasets showed the model was able to not only obtain better overall quality but also a wider diversity in the generated images.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: A vast majority of the current research in the field of Machine Learning is
done using algorithms with strong arguments pointing to their biological
implausibility such as Backpropagation, deviating the field's focus from
understanding its original organic inspiration to a compulsive search for
optimal performance. Yet, there have been a few proposed models that respect
most of the biological constraints present in the human brain and are valid
candidates for mimicking some of its properties and mechanisms. In this paper,
we will focus on guiding the learning of a biologically plausible generative
model called the Helmholtz Machine in complex search spaces using a heuristic
based on the Human Image Perception mechanism. We hypothesize that this model's
learning algorithm is not fit for Deep Networks due to its Hebbian-like local
update rule, rendering it incapable of taking full advantage of the
compositional properties that multi-layer networks provide. We propose to
overcome this problem, by providing the network's hidden layers with visual
queues at different resolutions using a Multi-level Data representation. The
results on several image datasets showed the model was able to not only obtain
better overall quality but also a wider diversity in the generated images,
corroborating our intuition that using our proposed heuristic allows the model
to take more advantage of the network's depth growth. More importantly, they
show the unexplored possibilities underlying brain-inspired models and
techniques.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Artificial Bee Colony optimization of Deep Convolutional Neural Networks
in the context of Biomedical Imaging [9.334663477968027]
We propose a novel, hybrid neuroevolutive algorithm that integrates the Artificial Bee Colony Algorithm with Evolutionary Computation tools to generate models from scratch.
The Chimera Algorithm has been validated with two datasets of natural and medical images, producing models that surpassed the performance of those coming from Transfer Learning.
arXiv Detail & Related papers (2024-02-23T10:21:03Z) - Unveiling the Unseen: Identifiable Clusters in Trained Depthwise
Convolutional Kernels [56.69755544814834]
Recent advances in depthwise-separable convolutional neural networks (DS-CNNs) have led to novel architectures.
This paper reveals another striking property of DS-CNN architectures: discernible and explainable patterns emerge in their trained depthwise convolutional kernels in all layers.
arXiv Detail & Related papers (2024-01-25T19:05:53Z) - Neural Echos: Depthwise Convolutional Filters Replicate Biological
Receptive Fields [56.69755544814834]
We present evidence suggesting that depthwise convolutional kernels are effectively replicating the biological receptive fields observed in the mammalian retina.
We propose a scheme that draws inspiration from the biological receptive fields.
arXiv Detail & Related papers (2024-01-18T18:06:22Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Multimodal Adaptive Fusion of Face and Gait Features using Keyless
attention based Deep Neural Networks for Human Identification [67.64124512185087]
Soft biometrics such as gait are widely used with face in surveillance tasks like person recognition and re-identification.
We propose a novel adaptive multi-biometric fusion strategy for the dynamic incorporation of gait and face biometric cues by leveraging keyless attention deep neural networks.
arXiv Detail & Related papers (2023-03-24T05:28:35Z) - GM-NeRF: Learning Generalizable Model-based Neural Radiance Fields from
Multi-view Images [79.39247661907397]
We introduce an effective framework Generalizable Model-based Neural Radiance Fields to synthesize free-viewpoint images.
Specifically, we propose a geometry-guided attention mechanism to register the appearance code from multi-view 2D images to a geometry proxy.
arXiv Detail & Related papers (2023-03-24T03:32:02Z) - Interpretability of an Interaction Network for identifying $H
\rightarrow b\bar{b}$ jets [4.553120911976256]
In recent times, AI models based on deep neural networks are becoming increasingly popular for many of these applications.
We explore interpretability of AI models by examining an Interaction Network (IN) model designed to identify boosted $Hto bbarb$ jets.
We additionally illustrate the activity of hidden layers within the IN model as Neural Activation Pattern (NAP) diagrams.
arXiv Detail & Related papers (2022-11-23T08:38:52Z) - Generative Deep Learning Techniques for Password Generation [0.5249805590164902]
We study a broad collection of deep learning and probabilistic based models in the light of password guessing.
We provide novel generative deep-learning models in terms of variational autoencoders exhibiting state-of-art sampling performance.
We perform a thorough empirical analysis in a unified controlled framework over well-known datasets.
arXiv Detail & Related papers (2020-12-10T14:11:45Z) - Network Bending: Expressive Manipulation of Deep Generative Models [0.2062593640149624]
We introduce a new framework for manipulating and interacting with deep generative models that we call network bending.
We show how it allows for the direct manipulation of semantically meaningful aspects of the generative process as well as allowing for a broad range of expressive outcomes.
arXiv Detail & Related papers (2020-05-25T21:48:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.