A Novel Sleep Stage Classification Using CNN Generated by an Efficient
Neural Architecture Search with a New Data Processing Trick
- URL: http://arxiv.org/abs/2110.15277v1
- Date: Wed, 27 Oct 2021 10:36:52 GMT
- Title: A Novel Sleep Stage Classification Using CNN Generated by an Efficient
Neural Architecture Search with a New Data Processing Trick
- Authors: Yu Xue, Ziming Yuan and Adam Slowik
- Abstract summary: We propose an efficient five-sleep-consuming classification method using convolutional neural networks (CNNs) with a novel data processing trick.
We make full use of genetic algorithm (GA), NASG, to search for the best CNN architecture.
We verify convergence of our data processing trick and compare the performance of traditional CNNs before and after using our trick.
- Score: 4.365107026636095
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the development of automatic sleep stage classification (ASSC)
techniques, many classical methods such as k-means, decision tree, and SVM have
been used in automatic sleep stage classification. However, few methods explore
deep learning on ASSC. Meanwhile, most deep learning methods require extensive
expertise and suffer from a mass of handcrafted steps which are time-consuming
especially when dealing with multi-classification tasks. In this paper, we
propose an efficient five-sleep-stage classification method using convolutional
neural networks (CNNs) with a novel data processing trick and we design neural
architecture search (NAS) technique based on genetic algorithm (GA), NAS-G, to
search for the best CNN architecture. Firstly, we attach each kernel with an
adaptive coefficient to enhance the signal processing of the inputs. This can
enhance the propagation of informative features and suppress the propagation of
useless features in the early stage of the network. Then, we make full use of
GA's heuristic search and the advantage of no need for the gradient to search
for the best architecture of CNN. This can achieve a CNN with better
performance than a handcrafted one in a large search space at the minimum cost.
We verify the convergence of our data processing trick and compare the
performance of traditional CNNs before and after using our trick. Meanwhile, we
compare the performance between the CNN generated through NAS-G and the
traditional CNNs with our trick. The experiments demonstrate that the
convergence of CNNs with data processing trick is faster than without data
processing trick and the CNN with data processing trick generated by NAS-G
outperforms the handcrafted counterparts that use the data processing trick
too.
Related papers
- Dynamic Semantic Compression for CNN Inference in Multi-access Edge
Computing: A Graph Reinforcement Learning-based Autoencoder [82.8833476520429]
We propose a novel semantic compression method, autoencoder-based CNN architecture (AECNN) for effective semantic extraction and compression in partial offloading.
In the semantic encoder, we introduce a feature compression module based on the channel attention mechanism in CNNs, to compress intermediate data by selecting the most informative features.
In the semantic decoder, we design a lightweight decoder to reconstruct the intermediate data through learning from the received compressed data to improve accuracy.
arXiv Detail & Related papers (2024-01-19T15:19:47Z) - Training Convolutional Neural Networks with the Forward-Forward
algorithm [1.74440662023704]
Forward Forward (FF) algorithm has up to now only been used in fully connected networks.
We show how the FF paradigm can be extended to CNNs.
Our FF-trained CNN, featuring a novel spatially-extended labeling technique, achieves a classification accuracy of 99.16% on the MNIST hand-written digits dataset.
arXiv Detail & Related papers (2023-12-22T18:56:35Z) - An Efficient Evolutionary Deep Learning Framework Based on Multi-source
Transfer Learning to Evolve Deep Convolutional Neural Networks [8.40112153818812]
Convolutional neural networks (CNNs) have constantly achieved better performance over years by introducing more complex topology, and enlarging the capacity towards deeper and wider CNNs.
The computational cost is still the bottleneck of automatically designing CNNs.
In this paper, inspired by transfer learning, a new evolutionary computation based framework is proposed to efficiently evolve CNNs.
arXiv Detail & Related papers (2022-12-07T20:22:58Z) - Attention-based Feature Compression for CNN Inference Offloading in Edge
Computing [93.67044879636093]
This paper studies the computational offloading of CNN inference in device-edge co-inference systems.
We propose a novel autoencoder-based CNN architecture (AECNN) for effective feature extraction at end-device.
Experiments show that AECNN can compress the intermediate data by more than 256x with only about 4% accuracy loss.
arXiv Detail & Related papers (2022-11-24T18:10:01Z) - Towards a General Purpose CNN for Long Range Dependencies in
$\mathrm{N}$D [49.57261544331683]
We propose a single CNN architecture equipped with continuous convolutional kernels for tasks on arbitrary resolution, dimensionality and length without structural changes.
We show the generality of our approach by applying the same CCNN to a wide set of tasks on sequential (1$mathrmD$) and visual data (2$mathrmD$)
Our CCNN performs competitively and often outperforms the current state-of-the-art across all tasks considered.
arXiv Detail & Related papers (2022-06-07T15:48:02Z) - An Acceleration Method Based on Deep Learning and Multilinear Feature
Space [0.0]
This paper presents an alternative approach based on the Multilinear Feature Space (MFS) method resorting to transfer learning from large CNN architectures.
The proposed method uses CNNs to generate feature maps, although it does not work as complexity reduction approach.
Our method, named AMFC, uses the transfer learning from pre-trained CNN to reduce the classification time of new sample image, with minimal accuracy loss.
arXiv Detail & Related papers (2021-10-16T23:49:12Z) - CONet: Channel Optimization for Convolutional Neural Networks [33.58529066005248]
We study channel size optimization in convolutional neural networks (CNN)
We introduce an efficient dynamic scaling algorithm -- CONet -- that automatically optimize channel sizes across network layers for a given CNN.
We conduct experiments on CIFAR10/100 and ImageNet datasets and show that CONet can find efficient and accurate architectures searched in ResNet, DARTS, and DARTS+ spaces.
arXiv Detail & Related papers (2021-08-15T21:48:25Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Fusion of CNNs and statistical indicators to improve image
classification [65.51757376525798]
Convolutional Networks have dominated the field of computer vision for the last ten years.
Main strategy to prolong this trend relies on further upscaling networks in size.
We hypothesise that adding heterogeneous sources of information may be more cost-effective to a CNN than building a bigger network.
arXiv Detail & Related papers (2020-12-20T23:24:31Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z) - Human Activity Recognition using Multi-Head CNN followed by LSTM [1.8830374973687412]
This study presents a novel method to recognize human physical activities using CNN followed by LSTM.
By using the proposed method, we achieve state-of-the-art accuracy, which is comparable to traditional machine learning algorithms and other deep neural network algorithms.
arXiv Detail & Related papers (2020-02-21T14:29:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.