Temporal Knowledge Distillation for Time-Sensitive Financial Services
Applications
- URL: http://arxiv.org/abs/2312.16799v1
- Date: Thu, 28 Dec 2023 03:04:30 GMT
- Title: Temporal Knowledge Distillation for Time-Sensitive Financial Services
Applications
- Authors: Hongda Shen and Eren Kurshan
- Abstract summary: Anomaly detection is frequently used in key compliance and risk functions such as financial crime detection fraud and cybersecurity.
Keeping up with the rapid changes by retraining the models with the latest data patterns introduces pressures in balancing the historical and current patterns.
The proposed approach provides advantages in retraining times while improving the model performance.
- Score: 7.1795069620810805
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detecting anomalies has become an increasingly critical function in the
financial service industry. Anomaly detection is frequently used in key
compliance and risk functions such as financial crime detection fraud and
cybersecurity. The dynamic nature of the underlying data patterns especially in
adversarial environments like fraud detection poses serious challenges to the
machine learning models. Keeping up with the rapid changes by retraining the
models with the latest data patterns introduces pressures in balancing the
historical and current patterns while managing the training data size.
Furthermore the model retraining times raise problems in time-sensitive and
high-volume deployment systems where the retraining period directly impacts the
models ability to respond to ongoing attacks in a timely manner. In this study
we propose a temporal knowledge distillation-based label augmentation approach
(TKD) which utilizes the learning from older models to rapidly boost the latest
model and effectively reduces the model retraining times to achieve improved
agility. Experimental results show that the proposed approach provides
advantages in retraining times while improving the model performance.
Related papers
- Temporal-Difference Variational Continual Learning [89.32940051152782]
A crucial capability of Machine Learning models in real-world applications is the ability to continuously learn new tasks.
In Continual Learning settings, models often struggle to balance learning new tasks with retaining previous knowledge.
We propose new learning objectives that integrate the regularization effects of multiple previous posterior estimations.
arXiv Detail & Related papers (2024-10-10T10:58:41Z) - A Cost-Aware Approach to Adversarial Robustness in Neural Networks [1.622320874892682]
We propose using accelerated failure time models to measure the effect of hardware choice, batch size, number of epochs, and test-set accuracy.
We evaluate several GPU types and use the Tree Parzen Estimator to maximize model robustness and minimize model run-time simultaneously.
arXiv Detail & Related papers (2024-09-11T20:43:59Z) - EsaCL: Efficient Continual Learning of Sparse Models [10.227171407348326]
Key challenge in the continual learning setting is to efficiently learn a sequence of tasks without forgetting how to perform previously learned tasks.
We propose a new method for efficient continual learning of sparse models (EsaCL) that can automatically prune redundant parameters without adversely impacting the model's predictive power.
arXiv Detail & Related papers (2024-01-11T04:59:44Z) - Learn from the Past: A Proxy Guided Adversarial Defense Framework with
Self Distillation Regularization [53.04697800214848]
Adversarial Training (AT) is pivotal in fortifying the robustness of deep learning models.
AT methods, relying on direct iterative updates for target model's defense, frequently encounter obstacles such as unstable training and catastrophic overfitting.
We present a general proxy guided defense framework, LAST' (bf Learn from the Pbf ast)
arXiv Detail & Related papers (2023-10-19T13:13:41Z) - On the Costs and Benefits of Adopting Lifelong Learning for Software
Analytics -- Empirical Study on Brown Build and Risk Prediction [17.502553991799832]
This paper evaluates the use of lifelong learning (LL) for industrial use cases at Ubisoft.
LL is used to continuously build and maintain ML-based software analytics tools using an incremental learner that progressively updates the old model using new data.
arXiv Detail & Related papers (2023-05-16T21:57:16Z) - DeepVol: Volatility Forecasting from High-Frequency Data with Dilated Causal Convolutions [53.37679435230207]
We propose DeepVol, a model based on Dilated Causal Convolutions that uses high-frequency data to forecast day-ahead volatility.
Our empirical results suggest that the proposed deep learning-based approach effectively learns global features from high-frequency data.
arXiv Detail & Related papers (2022-09-23T16:13:47Z) - Augmented Bilinear Network for Incremental Multi-Stock Time-Series
Classification [83.23129279407271]
We propose a method to efficiently retain the knowledge available in a neural network pre-trained on a set of securities.
In our method, the prior knowledge encoded in a pre-trained neural network is maintained by keeping existing connections fixed.
This knowledge is adjusted for the new securities by a set of augmented connections, which are optimized using the new data.
arXiv Detail & Related papers (2022-07-23T18:54:10Z) - CausalAgents: A Robustness Benchmark for Motion Forecasting using Causal
Relationships [8.679073301435265]
We construct a new benchmark for evaluating and improving model robustness by applying perturbations to existing data.
We use these labels to perturb the data by deleting non-causal agents from the scene.
Under non-causal perturbations, we observe a $25$-$38%$ relative change in minADE as compared to the original.
arXiv Detail & Related papers (2022-07-07T21:28:23Z) - DST: Dynamic Substitute Training for Data-free Black-box Attack [79.61601742693713]
We propose a novel dynamic substitute training attack method to encourage substitute model to learn better and faster from the target model.
We introduce a task-driven graph-based structure information learning constrain to improve the quality of generated training data.
arXiv Detail & Related papers (2022-04-03T02:29:11Z) - Leveraging the structure of dynamical systems for data-driven modeling [111.45324708884813]
We consider the impact of the training set and its structure on the quality of the long-term prediction.
We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models.
arXiv Detail & Related papers (2021-12-15T20:09:20Z) - Label Augmentation via Time-based Knowledge Distillation for Financial
Anomaly Detection [0.0]
Financial anomaly detection use cases face serious challenges due to the dynamic nature of the underlying patterns.
Keeping up with the rapid changes introduces other challenges as it moves the model away from older patterns or continuously grows the size of the training data.
We propose a label augmentation approach to utilize the learning from older models to boost the latest.
arXiv Detail & Related papers (2021-01-05T18:24:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.