Explaining Concept Drift through the Evolution of Group Counterfactuals
- URL: http://arxiv.org/abs/2509.09616v1
- Date: Thu, 11 Sep 2025 16:58:34 GMT
- Title: Explaining Concept Drift through the Evolution of Group Counterfactuals
- Authors: Ignacy Stępka, Jerzy Stefanowski,
- Abstract summary: We introduce a novel methodology to explain concept drift by analyzing the temporal evolution of group-based counterfactual explanations.<n>Our approach tracks shifts in the GCEs' cluster centroids and their associated counterfactual action vectors before and after a drift.
- Score: 2.7859337708965395
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine learning models in dynamic environments often suffer from concept drift, where changes in the data distribution degrade performance. While detecting this drift is a well-studied topic, explaining how and why the model's decision-making logic changes still remains a significant challenge. In this paper, we introduce a novel methodology to explain concept drift by analyzing the temporal evolution of group-based counterfactual explanations (GCEs). Our approach tracks shifts in the GCEs' cluster centroids and their associated counterfactual action vectors before and after a drift. These evolving GCEs act as an interpretable proxy, revealing structural changes in the model's decision boundary and its underlying rationale. We operationalize this analysis within a three-layer framework that synergistically combines insights from the data layer (distributional shifts), the model layer (prediction disagreement), and our proposed explanation layer. We show that such holistic view allows for a more comprehensive diagnosis of drift, making it possible to distinguish between different root causes, such as a spatial data shift versus a re-labeling of concepts.
Related papers
- Emergent Structured Representations Support Flexible In-Context Inference in Large Language Models [77.98801218316505]
Large language models (LLMs) exhibit emergent behaviors suggestive of human-like reasoning.<n>We investigate the internal processing of LLMs during in-context concept inference.
arXiv Detail & Related papers (2026-02-08T03:14:39Z) - Reasoning as State Transition: A Representational Analysis of Reasoning Evolution in Large Language Models [50.39102836928242]
We introduce a representational perspective to investigate the dynamics of the model's internal states.<n>We discover that post-training yields only limited improvement in static initial representation quality.
arXiv Detail & Related papers (2026-01-31T15:23:33Z) - Temporal Concept Dynamics in Diffusion Models via Prompt-Conditioned Interventions [70.87254264798341]
PCI is a training-free and model-agnostic framework for analyzing concept dynamics through diffusion time.<n>It reveals diverse temporal behaviors across diffusion models, in which certain phases of the trajectory are more favorable to specific concepts even within the same concept type.
arXiv Detail & Related papers (2025-12-09T11:05:08Z) - Analyzing Finetuning Representation Shift for Multimodal LLMs Steering [56.710375516257876]
We propose to map hidden states to interpretable visual and textual concepts.<n>This enables us to more efficiently compare certain semantic dynamics, such as the shift from an original and fine-tuned model.<n>We also demonstrate the use of shift vectors to capture these concepts changes.
arXiv Detail & Related papers (2025-01-06T13:37:13Z) - CORAL: Concept Drift Representation Learning for Co-evolving Time-series [6.4314326272535896]
Concept drift affects the reliability and accuracy of conventional analysis models.<n>This paper presents CORAL, a method that models time series as an evolving ecosystem to learn representations of concept drift.
arXiv Detail & Related papers (2025-01-02T15:09:00Z) - Sparse autoencoders reveal selective remapping of visual concepts during adaptation [54.82630842681845]
Adapting foundation models for specific purposes has become a standard approach to build machine learning systems.<n>We develop a new Sparse Autoencoder (SAE) for the CLIP vision transformer, named PatchSAE, to extract interpretable concepts.
arXiv Detail & Related papers (2024-12-06T18:59:51Z) - Unsupervised Assessment of Landscape Shifts Based on Persistent Entropy and Topological Preservation [0.0]
A drift in the input data can have negative consequences on a learning predictor and the system's stability.
In this article, we introduce a novel framework for monitoring changes in multi-dimensional data streams.
The framework operates in both unsupervised and supervised environments.
arXiv Detail & Related papers (2024-10-05T14:57:52Z) - Learning Discrete Concepts in Latent Hierarchical Models [73.01229236386148]
Learning concepts from natural high-dimensional data holds potential in building human-aligned and interpretable machine learning models.<n>We formalize concepts as discrete latent causal variables that are related via a hierarchical causal model.<n>We substantiate our theoretical claims with synthetic data experiments.
arXiv Detail & Related papers (2024-06-01T18:01:03Z) - Interpretable Imitation Learning with Dynamic Causal Relations [65.18456572421702]
We propose to expose captured knowledge in the form of a directed acyclic causal graph.
We also design this causal discovery process to be state-dependent, enabling it to model the dynamics in latent causal graphs.
The proposed framework is composed of three parts: a dynamic causal discovery module, a causality encoding module, and a prediction module, and is trained in an end-to-end manner.
arXiv Detail & Related papers (2023-09-30T20:59:42Z) - Model Based Explanations of Concept Drift [8.686667049158476]
Concept drift refers to the phenomenon that the distribution generating the observed data changes over time.
If drift is present, machine learning models can become inaccurate and need adjustment.
We present a novel technology characterizing concept drift in terms of the characteristic change of spatial features.
arXiv Detail & Related papers (2023-03-16T14:03:56Z) - From Concept Drift to Model Degradation: An Overview on
Performance-Aware Drift Detectors [1.757501664210825]
Changes in the system on which a predictive machine learning model has been trained may lead to performance degradation during the system's life cycle.
Different terms have been used in the literature to refer to the same type of concept drift and the same term for various types.
This lack of unified terminology is set out to create confusion on distinguishing between different concept drift variants.
arXiv Detail & Related papers (2022-03-21T15:48:13Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - Counterfactual Explanations of Concept Drift [11.53362411363005]
concept drift refers to the phenomenon that the distribution, which is underlying the observed data, changes over time.
We present a novel technology, which characterizes concept drift in terms of the characteristic change of spatial features represented by typical examples.
arXiv Detail & Related papers (2020-06-23T08:27:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.