Introducing Fuzzy Layers for Deep Learning
- URL: http://arxiv.org/abs/2003.00880v1
- Date: Fri, 21 Feb 2020 19:33:30 GMT
- Title: Introducing Fuzzy Layers for Deep Learning
- Authors: Stanton R. Price, Steven R. Price, Derek T. Anderson
- Abstract summary: We introduce a new layer to deep learning: the fuzzy layer.
Traditionally, the network architecture of neural networks is composed of an input layer, some combination of hidden layers, and an output layer.
We propose the introduction of fuzzy layers into the deep learning architecture to exploit the powerful aggregation properties expressed through fuzzy methodologies.
- Score: 5.209583609264815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many state-of-the-art technologies developed in recent years have been
influenced by machine learning to some extent. Most popular at the time of this
writing are artificial intelligence methodologies that fall under the umbrella
of deep learning. Deep learning has been shown across many applications to be
extremely powerful and capable of handling problems that possess great
complexity and difficulty. In this work, we introduce a new layer to deep
learning: the fuzzy layer. Traditionally, the network architecture of neural
networks is composed of an input layer, some combination of hidden layers, and
an output layer. We propose the introduction of fuzzy layers into the deep
learning architecture to exploit the powerful aggregation properties expressed
through fuzzy methodologies, such as the Choquet and Sugueno fuzzy integrals.
To date, fuzzy approaches taken to deep learning have been through the
application of various fusion strategies at the decision level to aggregate
outputs from state-of-the-art pre-trained models, e.g., AlexNet, VGG16,
GoogLeNet, Inception-v3, ResNet-18, etc. While these strategies have been shown
to improve accuracy performance for image classification tasks, none have
explored the use of fuzzified intermediate, or hidden, layers. Herein, we
present a new deep learning strategy that incorporates fuzzy strategies into
the deep learning architecture focused on the application of semantic
segmentation using per-pixel classification. Experiments are conducted on a
benchmark data set as well as a data set collected via an unmanned aerial
system at a U.S. Army test site for the task of automatic road segmentation,
and preliminary results are promising.
Related papers
- Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - A Generic Layer Pruning Method for Signal Modulation Recognition Deep Learning Models [17.996775444294276]
Deep neural networks are becoming the preferred method for signal classification.
They often come with high computational complexity and large model sizes.
We propose a novel layer pruning method to address this challenge.
arXiv Detail & Related papers (2024-06-12T06:46:37Z) - Generalized Label-Efficient 3D Scene Parsing via Hierarchical Feature
Aligned Pre-Training and Region-Aware Fine-tuning [55.517000360348725]
This work presents a framework for dealing with 3D scene understanding when the labeled scenes are quite limited.
To extract knowledge for novel categories from the pre-trained vision-language models, we propose a hierarchical feature-aligned pre-training and knowledge distillation strategy.
Experiments with both indoor and outdoor scenes demonstrated the effectiveness of our approach in both data-efficient learning and open-world few-shot learning.
arXiv Detail & Related papers (2023-12-01T15:47:04Z) - Understanding Deep Representation Learning via Layerwise Feature
Compression and Discrimination [33.273226655730326]
We show that each layer of a deep linear network progressively compresses within-class features at a geometric rate and discriminates between-class features at a linear rate.
This is the first quantitative characterization of feature evolution in hierarchical representations of deep linear networks.
arXiv Detail & Related papers (2023-11-06T09:00:38Z) - Layer-wise Linear Mode Connectivity [52.6945036534469]
Averaging neural network parameters is an intuitive method for the knowledge of two independent models.
It is most prominently used in federated learning.
We analyse the performance of the models that result from averaging single, or groups.
arXiv Detail & Related papers (2023-07-13T09:39:10Z) - A Law of Data Separation in Deep Learning [41.58856318262069]
We study the fundamental question of how deep neural networks process data in the intermediate layers.
Our finding is a simple and quantitative law that governs how deep neural networks separate data according to class membership.
arXiv Detail & Related papers (2022-10-31T02:25:38Z) - A Layer-Wise Information Reinforcement Approach to Improve Learning in
Deep Belief Networks [0.4893345190925178]
This paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining.
Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification.
arXiv Detail & Related papers (2021-01-17T18:53:18Z) - Neural Function Modules with Sparse Arguments: A Dynamic Approach to
Integrating Information across Layers [84.57980167400513]
Neural Function Modules (NFM) aims to introduce the same structural capability into deep learning.
Most of the work in the context of feed-forward networks combining top-down and bottom-up feedback is limited to classification problems.
The key contribution of our work is to combine attention, sparsity, top-down and bottom-up feedback, in a flexible algorithm.
arXiv Detail & Related papers (2020-10-15T20:43:17Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - Sparse Coding Driven Deep Decision Tree Ensembles for Nuclear
Segmentation in Digital Pathology Images [15.236873250912062]
We propose an easily trained yet powerful representation learning approach with performance highly competitive to deep neural networks in a digital pathology image segmentation task.
The method, called sparse coding driven deep decision tree ensembles that we abbreviate as ScD2TE, provides a new perspective on representation learning.
arXiv Detail & Related papers (2020-08-13T02:59:31Z) - Understanding Deep Architectures with Reasoning Layer [60.90906477693774]
We show that properties of the algorithm layers, such as convergence, stability, and sensitivity, are intimately related to the approximation and generalization abilities of the end-to-end model.
Our theory can provide useful guidelines for designing deep architectures with reasoning layers.
arXiv Detail & Related papers (2020-06-24T00:26:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.