OLC-WA: Drift Aware Tuning-Free Online Classification with Weighted Average
- URL: http://arxiv.org/abs/2512.12785v1
- Date: Sun, 14 Dec 2025 17:52:39 GMT
- Title: OLC-WA: Drift Aware Tuning-Free Online Classification with Weighted Average
- Authors: Mohammad Abu Shaira, Yunhe Feng, Heng Fan, Weishi Shi,
- Abstract summary: This paper introduces Online Classification with weighted average (OLC-WA)<n>OLC-WA operates by blending incoming data streams with an existing base model.<n>An integrated optimization mechanism dynamically detects concept drift, quantifies its magnitude, and adjusts the model.
- Score: 18.902790013938006
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world data sets often exhibit temporal dynamics characterized by evolving data distributions. Disregarding this phenomenon, commonly referred to as concept drift, can significantly diminish a model's predictive accuracy. Furthermore, the presence of hyperparameters in online models exacerbates this issue. These parameters are typically fixed and cannot be dynamically adjusted by the user in response to the evolving data distribution. This paper introduces Online Classification with Weighted Average (OLC-WA), an adaptive, hyperparameter-free online classification model equipped with an automated optimization mechanism. OLC-WA operates by blending incoming data streams with an existing base model. This blending is facilitated by an exponentially weighted moving average. Furthermore, an integrated optimization mechanism dynamically detects concept drift, quantifies its magnitude, and adjusts the model based on the observed data stream characteristics. This approach empowers the model to effectively adapt to evolving data distributions within streaming environments. Rigorous empirical evaluation across diverse benchmark datasets shows that OLC-WA achieves performance comparable to batch models in stationary environments, maintaining accuracy within 1-3%, and surpasses leading online baselines by 10-25% under drift, demonstrating its effectiveness in adapting to dynamic data streams.
Related papers
- Learning from Yesterday's Error: An Efficient Online Learning Method for Traffic Demand Prediction [6.104967994062357]
FORESEE (Forecasting Online with Residual Smoothing and Ensemble Experts) is a lightweight online adaptation framework that is accurate, robust, and efficient.<n>It corrects today's forecast in each region using yesterday's prediction error, through exponential smoothing guided by a mixture-of-experts mechanism.<n>Experiments on seven real-world datasets with backbone models demonstrate that FORESEE consistently improves prediction accuracy, maintains robustness even when distribution shifts are minimal.
arXiv Detail & Related papers (2026-02-25T10:19:39Z) - OLR-WAA: Adaptive and Drift-Resilient Online Regression with Dynamic Weighted Averaging [7.146027549101716]
"OLR-WAA: An Adaptive and Drift-Resilient Online Regression with Dynamic weighted average"<n>This paper introduces "OLR-WAA: An Adaptive and Drift-Resilient Online Regression with Dynamic weighted average"
arXiv Detail & Related papers (2025-12-14T17:39:51Z) - DAO-GP Drift Aware Online Non-Linear Regression Gaussian-Process [8.665001359628592]
Real-world datasets often exhibit temporal dynamics characterized by evolving data distributions.<n>Disregarding this phenomenon, commonly referred to as concept drift, can significantly diminish a model's predictive accuracy.<n>Drift-Aware Online Process (GP) is a novel, fully adaptive, hyper-free, and sparse non-linear regression model.
arXiv Detail & Related papers (2025-12-09T18:12:38Z) - Nonparametric Data Attribution for Diffusion Models [57.820618036556084]
Data attribution for generative models seeks to quantify the influence of individual training examples on model outputs.<n>We propose a nonparametric attribution method that operates entirely on data, measuring influence via patch-level similarity between generated and training images.
arXiv Detail & Related papers (2025-10-16T03:37:16Z) - Boosting Generalization Performance in Model-Heterogeneous Federated Learning Using Variational Transposed Convolution [0.27309692684728615]
Federated learning (FL) is a pioneering machine learning paradigm that enables distributed clients to process local data effectively.<n>Traditional model-homogeneous approaches mainly involve debiasing the local training procedures with regularization or dynamically adjusting client weights in aggregation.<n>We propose a model-heterogeneous FL framework that can improve clients' generalization performance over unseen data without model aggregation.
arXiv Detail & Related papers (2025-08-03T08:55:18Z) - Efficient Federated Learning with Heterogeneous Data and Adaptive Dropout [62.73150122809138]
Federated Learning (FL) is a promising distributed machine learning approach that enables collaborative training of a global model using multiple edge devices.<n>We propose the FedDHAD FL framework, which comes with two novel methods: Dynamic Heterogeneous model aggregation (FedDH) and Adaptive Dropout (FedAD)<n>The combination of these two methods makes FedDHAD significantly outperform state-of-the-art solutions in terms of accuracy (up to 6.7% higher), efficiency (up to 2.02 times faster), and cost (up to 15.0% smaller)
arXiv Detail & Related papers (2025-07-14T16:19:00Z) - Physics-guided Active Sample Reweighting for Urban Flow Prediction [75.24539704456791]
Urban flow prediction is a nuanced-temporal modeling that estimates the throughput of transportation services like buses, taxis and ride-driven models.
Some recent prediction solutions bring remedies with the notion of physics-guided machine learning (PGML)
We develop a atized physics-guided network (PN), and propose a data-aware framework Physics-guided Active Sample Reweighting (P-GASR)
arXiv Detail & Related papers (2024-07-18T15:44:23Z) - AiGAS-dEVL: An Adaptive Incremental Neural Gas Model for Drifting Data Streams under Extreme Verification Latency [6.7236795813629]
In streaming setups data flows are affected by factors that yield non-stationarities in the patterns (concept drift)
We propose a novel approach, AiGAS-dEVL, which relies on growing neural gas to characterize the distributions of all concepts detected within the stream over time.
Our approach exposes that the online analysis of the behavior of these points over time facilitates the definition of the evolution of concepts in the feature space.
arXiv Detail & Related papers (2024-07-07T14:04:57Z) - Test-Time Adaptation Induces Stronger Accuracy and Agreement-on-the-Line [65.14099135546594]
Recent test-time adaptation (TTA) methods drastically strengthen the ACL and AGL trends in models, even in shifts where models showed very weak correlations before.
Our results show that by combining TTA with AGL-based estimation methods, we can estimate the OOD performance of models with high precision for a broader set of distribution shifts.
arXiv Detail & Related papers (2023-10-07T23:21:25Z) - Kalman Filter for Online Classification of Non-Stationary Data [101.26838049872651]
In Online Continual Learning (OCL) a learning system receives a stream of data and sequentially performs prediction and training steps.
We introduce a probabilistic Bayesian online learning model by using a neural representation and a state space model over the linear predictor weights.
In experiments in multi-class classification we demonstrate the predictive ability of the model and its flexibility to capture non-stationarity.
arXiv Detail & Related papers (2023-06-14T11:41:42Z) - Employing chunk size adaptation to overcome concept drift [2.277447144331876]
We propose a new Chunk Adaptive Restoration framework that can be adapted to any block-based data stream classification algorithm.
The proposed algorithm adjusts the data chunk size in the case of concept drift detection to minimize the impact of the change on the predictive performance of the used model.
arXiv Detail & Related papers (2021-10-25T12:36:22Z) - Autoregressive Dynamics Models for Offline Policy Evaluation and
Optimization [60.73540999409032]
We show that expressive autoregressive dynamics models generate different dimensions of the next state and reward sequentially conditioned on previous dimensions.
We also show that autoregressive dynamics models are useful for offline policy optimization by serving as a way to enrich the replay buffer.
arXiv Detail & Related papers (2021-04-28T16:48:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.