TransNAS-TSAD: Harnessing Transformers for Multi-Objective Neural
Architecture Search in Time Series Anomaly Detection
- URL: http://arxiv.org/abs/2311.18061v3
- Date: Mon, 4 Mar 2024 23:04:21 GMT
- Title: TransNAS-TSAD: Harnessing Transformers for Multi-Objective Neural
Architecture Search in Time Series Anomaly Detection
- Authors: Ijaz Ul Haq, Byung Suk Lee and Donna M. Rizzo
- Abstract summary: This paper introduces TransNAS-TSAD, a framework that synergizes the transformer architecture with neural architecture search (NAS)
Our evaluation reveals that TransNAS-TSAD surpasses conventional anomaly detection models due to its tailored architectural adaptability.
- Score: 3.5681028373124066
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The surge in real-time data collection across various industries has
underscored the need for advanced anomaly detection in both univariate and
multivariate time series data. This paper introduces TransNAS-TSAD, a framework
that synergizes the transformer architecture with neural architecture search
(NAS), enhanced through NSGA-II algorithm optimization. This approach
effectively tackles the complexities of time series data, balancing
computational efficiency with detection accuracy. Our evaluation reveals that
TransNAS-TSAD surpasses conventional anomaly detection models due to its
tailored architectural adaptability and the efficient exploration of complex
search spaces, leading to marked improvements in diverse data scenarios. We
also introduce the Efficiency-Accuracy-Complexity Score (EACS) as a new metric
for assessing model performance, emphasizing the balance between accuracy and
computational resources. TransNAS-TSAD sets a new benchmark in time series
anomaly detection, offering a versatile, efficient solution for complex
real-world applications. This research highlights the TransNAS-TSAD potential
across a wide range of industry applications and paves the way for future
developments in the field.
Related papers
- Learning Novel Transformer Architecture for Time-series Forecasting [9.412920379798928]
AutoFormer-TS is a novel framework that leverages a comprehensive search space for Transformer architectures tailored to time-series prediction tasks.
Our framework introduces a differentiable neural architecture search (DNAS) method, AB-DARTS, which improves upon existing DNAS approaches.
arXiv Detail & Related papers (2025-02-19T13:49:20Z) - Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - POMONAG: Pareto-Optimal Many-Objective Neural Architecture Generator [4.09225917049674]
Transferable NAS has emerged, generalizing the search process from dataset-dependent to task-dependent.
This paper introduces POMONAG, extending DiffusionNAG via a many-optimal diffusion process.
Results were validated on two search spaces -- NAS201 and MobileNetV3 -- and evaluated across 15 image classification datasets.
arXiv Detail & Related papers (2024-09-30T16:05:29Z) - SMORE: Similarity-based Hyperdimensional Domain Adaptation for
Multi-Sensor Time Series Classification [17.052624039805856]
We propose SMORE, a novel resource-efficient domain adaptation (DA) algorithm for multi-sensor time series classification.
SMORE achieves on average 1.98% higher accuracy than state-of-the-art (SOTA) DNN-based DA algorithms with 18.81x faster training and 4.63x faster inference.
arXiv Detail & Related papers (2024-02-20T18:48:49Z) - Efficient Model Adaptation for Continual Learning at the Edge [15.334881190102895]
Most machine learning (ML) systems assume stationary and matching data distributions during training and deployment.
Data distributions often shift over time due to changes in environmental factors, sensor characteristics, and task-of-interest.
This paper presents theAdaptor-Reconfigurator (EAR) framework for efficient continual learning under domain shifts.
arXiv Detail & Related papers (2023-08-03T23:55:17Z) - Joint Spatial-Temporal and Appearance Modeling with Transformer for
Multiple Object Tracking [59.79252390626194]
We propose a novel solution named TransSTAM, which leverages Transformer to model both the appearance features of each object and the spatial-temporal relationships among objects.
The proposed method is evaluated on multiple public benchmarks including MOT16, MOT17, and MOT20, and it achieves a clear performance improvement in both IDF1 and HOTA.
arXiv Detail & Related papers (2022-05-31T01:19:18Z) - $\beta$-DARTS: Beta-Decay Regularization for Differentiable Architecture
Search [85.84110365657455]
We propose a simple-but-efficient regularization method, termed as Beta-Decay, to regularize the DARTS-based NAS searching process.
Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets.
arXiv Detail & Related papers (2022-03-03T11:47:14Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - TELESTO: A Graph Neural Network Model for Anomaly Classification in
Cloud Services [77.454688257702]
Machine learning (ML) and artificial intelligence (AI) are applied on IT system operation and maintenance.
One direction aims at the recognition of re-occurring anomaly types to enable remediation automation.
We propose a method that is invariant to dimensionality changes of given data.
arXiv Detail & Related papers (2021-02-25T14:24:49Z) - Deep Cellular Recurrent Network for Efficient Analysis of Time-Series
Data with Spatial Information [52.635997570873194]
This work proposes a novel deep cellular recurrent neural network (DCRNN) architecture to process complex multi-dimensional time series data with spatial information.
The proposed architecture achieves state-of-the-art performance while utilizing substantially less trainable parameters when compared to comparable methods in the literature.
arXiv Detail & Related papers (2021-01-12T20:08:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.