Omni-Scale CNNs: a simple and effective kernel size configuration for
time series classification
- URL: http://arxiv.org/abs/2002.10061v3
- Date: Fri, 17 Jun 2022 07:56:18 GMT
- Title: Omni-Scale CNNs: a simple and effective kernel size configuration for
time series classification
- Authors: Wensi Tang, Guodong Long, Lu Liu, Tianyi Zhou, Michael Blumenstein,
Jing Jiang
- Abstract summary: The Receptive Field (RF) size has been one of the most important factors for One Dimensional Convolutional Neural Networks (1D-CNNs) on time series classification tasks.
We propose an Omni-Scale block (OS-block) for 1D-CNNs, where the kernel sizes are decided by a simple and universal rule.
Experiment result shows that models with the OS-block can achieve a similar performance as models with the searched optimal RF size.
- Score: 47.423272376757204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Receptive Field (RF) size has been one of the most important factors for
One Dimensional Convolutional Neural Networks (1D-CNNs) on time series
classification tasks. Large efforts have been taken to choose the appropriate
size because it has a huge influence on the performance and differs
significantly for each dataset. In this paper, we propose an Omni-Scale block
(OS-block) for 1D-CNNs, where the kernel sizes are decided by a simple and
universal rule. Particularly, it is a set of kernel sizes that can efficiently
cover the best RF size across different datasets via consisting of multiple
prime numbers according to the length of the time series. The experiment result
shows that models with the OS-block can achieve a similar performance as models
with the searched optimal RF size and due to the strong optimal RF size capture
ability, simple 1D-CNN models with OS-block achieves the state-of-the-art
performance on four time series benchmarks, including both univariate and
multivariate data from multiple domains. Comprehensive analysis and discussions
shed light on why the OS-block can capture optimal RF sizes across different
datasets. Code available [https://github.com/Wensi-Tang/OS-CNN]
Related papers
- AdaFSNet: Time Series Classification Based on Convolutional Network with a Adaptive and Effective Kernel Size Configuration [9.070240771963544]
We propose an Adaptive and Effective Full-Scope Convolutional Neural Network (AdaFSNet) to enhance the accuracy of time series classification.
Our model surpassed baseline models in terms of classification accuracy, underscoring the AdaFSNet network's efficiency and effectiveness in handling time series classification tasks.
arXiv Detail & Related papers (2024-04-28T16:58:53Z) - D-Net: Dynamic Large Kernel with Dynamic Feature Fusion for Volumetric Medical Image Segmentation [7.894630378784007]
We propose Dynamic Large Kernel (DLK) and Dynamic Feature Fusion (DFF) modules.
D-Net is able to effectively utilize a multi-scale large receptive field and adaptively harness global contextual information.
arXiv Detail & Related papers (2024-03-15T20:49:43Z) - DISCO: Distributed Inference with Sparse Communications [4.463769269318892]
Distributed computing is a common approach to reduce single-node memory consumption.
In this paper, we explore the "within-layer model parallelism", which distributes the inference of each layer into multiple nodes.
We show the benefit of the DISCO framework on a variety of CV tasks such as image classification, object detection, semantic segmentation, and image super resolution.
arXiv Detail & Related papers (2023-02-22T07:20:34Z) - Toward Large Kernel Models [16.704246627541103]
We introduce EigenPro 3.0, an algorithm based on projected dual preconditioned SGD.
We show scaling to model and data sizes which have not been possible with existing kernel methods.
arXiv Detail & Related papers (2023-02-06T07:57:50Z) - Towards a General Purpose CNN for Long Range Dependencies in
$\mathrm{N}$D [49.57261544331683]
We propose a single CNN architecture equipped with continuous convolutional kernels for tasks on arbitrary resolution, dimensionality and length without structural changes.
We show the generality of our approach by applying the same CCNN to a wide set of tasks on sequential (1$mathrmD$) and visual data (2$mathrmD$)
Our CCNN performs competitively and often outperforms the current state-of-the-art across all tasks considered.
arXiv Detail & Related papers (2022-06-07T15:48:02Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - Solving Mixed Integer Programs Using Neural Networks [57.683491412480635]
This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one.
Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP.
We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each.
arXiv Detail & Related papers (2020-12-23T09:33:11Z) - FarSee-Net: Real-Time Semantic Segmentation by Efficient Multi-scale
Context Aggregation and Feature Space Super-resolution [14.226301825772174]
We introduce a novel and efficient module called Cascaded Factorized Atrous Spatial Pyramid Pooling (CF-ASPP)
It is a lightweight cascaded structure for Convolutional Neural Networks (CNNs) to efficiently leverage context information.
We achieve 68.4% mIoU at 84 fps on the Cityscapes test set with a single Nivida Titan X (Maxwell) GPU card.
arXiv Detail & Related papers (2020-03-09T03:53:57Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.