Extracting Interpretable Local and Global Representations from Attention
on Time Series
- URL: http://arxiv.org/abs/2312.11466v1
- Date: Sat, 16 Sep 2023 00:51:49 GMT
- Title: Extracting Interpretable Local and Global Representations from Attention
on Time Series
- Authors: Leonid Schwenke, Martin Atzmueller
- Abstract summary: This paper targets two transformer attention based interpretability methods working with local abstraction and global representation.
We distinguish local and global contexts, and provide a comprehensive framework for both general interpretation options.
- Score: 0.135975510645475
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper targets two transformer attention based interpretability methods
working with local abstraction and global representation, in the context of
time series data. We distinguish local and global contexts, and provide a
comprehensive framework for both general interpretation options. We discuss
their specific instantiation via different methods in detail, also outlining
their respective computational implementation and abstraction variants.
Furthermore, we provide extensive experimentation demonstrating the efficacy of
the presented approaches. In particular, we perform our experiments using a
selection of univariate datasets from the UCR UEA time series repository where
we both assess the performance of the proposed approaches, as well as their
impact on explainability and interpretability/complexity. Here, with an
extensive analysis of hyperparameters, the presented approaches demonstrate an
significant improvement in interpretability/complexity, while capturing many
core decisions of and maintaining a similar performance to the baseline model.
Finally, we draw general conclusions outlining and guiding the application of
the presented methods.
Related papers
- Uniting contrastive and generative learning for event sequences models [51.547576949425604]
This study investigates the integration of two self-supervised learning techniques - instance-wise contrastive learning and a generative approach based on restoring masked events in latent space.
Experiments conducted on several public datasets, focusing on sequence classification and next-event type prediction, show that the integrated method achieves superior performance compared to individual approaches.
arXiv Detail & Related papers (2024-08-19T13:47:17Z) - Diffusion Features to Bridge Domain Gap for Semantic Segmentation [2.8616666231199424]
This paper investigates the approach that leverages the sampling and fusion techniques to harness the features of diffusion models efficiently.
By leveraging the strength of text-to-image generation capability, we introduce a new training framework designed to implicitly learn posterior knowledge from it.
arXiv Detail & Related papers (2024-06-02T15:33:46Z) - Disentangling Multi-view Representations Beyond Inductive Bias [32.15900989696017]
We propose a novel multi-view representation disentangling method that ensures both interpretability and generalizability of the resulting representations.
Our experiments on four multi-view datasets demonstrate that our proposed method outperforms 12 comparison methods in terms of clustering and classification performance.
arXiv Detail & Related papers (2023-08-03T09:09:28Z) - Prompting Diffusion Representations for Cross-Domain Semantic
Segmentation [101.04326113360342]
diffusion-pretraining achieves extraordinary domain generalization results for semantic segmentation.
We introduce a scene prompt and a prompt randomization strategy to help further disentangle the domain-invariant information when training the segmentation head.
arXiv Detail & Related papers (2023-07-05T09:28:25Z) - GLOBE-CE: A Translation-Based Approach for Global Counterfactual
Explanations [10.276136171459731]
Global & Efficient Counterfactual Explanations (GLOBE-CE) is a flexible framework that tackles the reliability and scalability issues associated with current state-of-the-art.
We provide a unique mathematical analysis of categorical feature translations, utilising it in our method.
Experimental evaluation with publicly available datasets and user studies demonstrate that GLOBE-CE performs significantly better than the current state-of-the-art.
arXiv Detail & Related papers (2023-05-26T15:26:59Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Variational Distillation for Multi-View Learning [104.17551354374821]
We design several variational information bottlenecks to exploit two key characteristics for multi-view representation learning.
Under rigorously theoretical guarantee, our approach enables IB to grasp the intrinsic correlation between observations and semantic labels.
arXiv Detail & Related papers (2022-06-20T03:09:46Z) - Decoupling Local and Global Representations of Time Series [38.73548222141307]
We propose a novel generative approach for learning representations for the global and local factors of variation in time series.
In experiments, we demonstrate successful recovery of the true local and global variability factors on simulated data.
We believe that the proposed way of defining representations is beneficial for data modelling and yields better insights into the complexity of real-world data.
arXiv Detail & Related papers (2022-02-04T17:46:04Z) - Interpretable Multi-dataset Evaluation for Named Entity Recognition [110.64368106131062]
We present a general methodology for interpretable evaluation for the named entity recognition (NER) task.
The proposed evaluation method enables us to interpret the differences in models and datasets, as well as the interplay between them.
By making our analysis tool available, we make it easy for future researchers to run similar analyses and drive progress in this area.
arXiv Detail & Related papers (2020-11-13T10:53:27Z) - Invariant Causal Prediction for Block MDPs [106.63346115341862]
Generalization across environments is critical to the successful application of reinforcement learning algorithms to real-world challenges.
We propose a method of invariant prediction to learn model-irrelevance state abstractions (MISA) that generalize to novel observations in the multi-environment setting.
arXiv Detail & Related papers (2020-03-12T21:03:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.