Examining convolutional feature extraction using Maximum Entropy (ME)
and Signal-to-Noise Ratio (SNR) for image classification
- URL: http://arxiv.org/abs/2105.04097v1
- Date: Mon, 10 May 2021 03:58:06 GMT
- Title: Examining convolutional feature extraction using Maximum Entropy (ME)
and Signal-to-Noise Ratio (SNR) for image classification
- Authors: Nidhi Gowdra, Roopak Sinha and Stephen MacDonell
- Abstract summary: Convolutional Neural Networks (CNNs) specialize in feature extraction rather than function mapping.
In this paper, we examine the feature extraction capabilities of CNNs using Maximum Entropy (ME) and Signal-to-Noise Ratio (SNR)
We show that the classification accuracy or performance of CNNs is greatly dependent on the amount, complexity and quality of the signal information present in the input data.
- Score: 0.6875312133832078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNNs) specialize in feature extraction rather
than function mapping. In doing so they form complex internal hierarchical
feature representations, the complexity of which gradually increases with a
corresponding increment in neural network depth. In this paper, we examine the
feature extraction capabilities of CNNs using Maximum Entropy (ME) and
Signal-to-Noise Ratio (SNR) to validate the idea that, CNN models should be
tailored for a given task and complexity of the input data. SNR and ME measures
are used as they can accurately determine in the input dataset, the relative
amount of signal information to the random noise and the maximum amount of
information respectively. We use two well known benchmarking datasets, MNIST
and CIFAR-10 to examine the information extraction and abstraction capabilities
of CNNs. Through our experiments, we examine convolutional feature extraction
and abstraction capabilities in CNNs and show that the classification accuracy
or performance of CNNs is greatly dependent on the amount, complexity and
quality of the signal information present in the input data. Furthermore, we
show the effect of information overflow and underflow on CNN classification
accuracies. Our hypothesis is that the feature extraction and abstraction
capabilities of convolutional layers are limited and therefore, CNN models
should be tailored to the input data by using appropriately sized CNNs based on
the SNR and ME measures of the input dataset.
Related papers
- Fuzzy Convolution Neural Networks for Tabular Data Classification [0.0]
Convolutional neural networks (CNNs) have attracted a great deal of attention due to their remarkable performance in various domains.
In this paper, we propose a novel framework fuzzy convolution neural network (FCNN) tailored specifically for tabular data.
arXiv Detail & Related papers (2024-06-04T20:33:35Z) - Application of Tensorized Neural Networks for Cloud Classification [0.0]
Convolutional neural networks (CNNs) have gained widespread usage across various fields such as weather forecasting, computer vision, autonomous driving, and medical image analysis.
However, the practical implementation and commercialization of CNNs in these domains are hindered by challenges related to model sizes, overfitting, and computational time.
We propose a groundbreaking approach that involves tensorizing the dense layers in the CNN to reduce model size and computational time.
arXiv Detail & Related papers (2024-03-21T06:28:22Z) - Assessing Neural Network Representations During Training Using
Noise-Resilient Diffusion Spectral Entropy [55.014926694758195]
Entropy and mutual information in neural networks provide rich information on the learning process.
We leverage data geometry to access the underlying manifold and reliably compute these information-theoretic measures.
We show that they form noise-resistant measures of intrinsic dimensionality and relationship strength in high-dimensional simulated data.
arXiv Detail & Related papers (2023-12-04T01:32:42Z) - FFEINR: Flow Feature-Enhanced Implicit Neural Representation for
Spatio-temporal Super-Resolution [4.577685231084759]
This paper proposes a Feature-Enhanced Neural Implicit Representation (FFEINR) for super-resolution of flow field data.
It can take full advantage of the implicit neural representation in terms of model structure and sampling resolution.
The training process of FFEINR is facilitated by introducing feature enhancements for the input layer.
arXiv Detail & Related papers (2023-08-24T02:28:18Z) - Attention-based Feature Compression for CNN Inference Offloading in Edge
Computing [93.67044879636093]
This paper studies the computational offloading of CNN inference in device-edge co-inference systems.
We propose a novel autoencoder-based CNN architecture (AECNN) for effective feature extraction at end-device.
Experiments show that AECNN can compress the intermediate data by more than 256x with only about 4% accuracy loss.
arXiv Detail & Related papers (2022-11-24T18:10:01Z) - What Can Be Learnt With Wide Convolutional Neural Networks? [69.55323565255631]
We study infinitely-wide deep CNNs in the kernel regime.
We prove that deep CNNs adapt to the spatial scale of the target function.
We conclude by computing the generalisation error of a deep CNN trained on the output of another deep CNN.
arXiv Detail & Related papers (2022-08-01T17:19:32Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - Lost Vibration Test Data Recovery Using Convolutional Neural Network: A
Case Study [0.0]
This paper proposes a CNN algorithm for the Alamosa Canyon Bridge as a real structure.
Three different CNN models were considered to predict one and two malfunctioned sensors.
The accuracy of the model was increased by adding a convolutional layer.
arXiv Detail & Related papers (2022-04-11T23:24:03Z) - An Analysis of Complex-Valued CNNs for RF Data-Driven Wireless Device
Classification [12.810432378755904]
Recent deep neural network-based device classification studies show that complex-valued neural networks (CVNNs) yield higher classification accuracy than real-valued neural networks (RVNNs)
Our study provides a deeper understanding of this trend using real LoRa and WiFi RF datasets.
arXiv Detail & Related papers (2022-02-20T10:35:20Z) - Deep Networks for Direction-of-Arrival Estimation in Low SNR [89.45026632977456]
We introduce a Convolutional Neural Network (CNN) that is trained from mutli-channel data of the true array manifold matrix.
We train a CNN in the low-SNR regime to predict DoAs across all SNRs.
Our robust solution can be applied in several fields, ranging from wireless array sensors to acoustic microphones or sonars.
arXiv Detail & Related papers (2020-11-17T12:52:18Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.