Discovering Novel Customer Features with Recurrent Neural Networks for
Personality Based Financial Services
- URL: http://arxiv.org/abs/2109.11871v1
- Date: Fri, 24 Sep 2021 10:32:36 GMT
- Title: Discovering Novel Customer Features with Recurrent Neural Networks for
Personality Based Financial Services
- Authors: Charl Maree, Christian W. Omlin
- Abstract summary: The micro-segmentation of customers in the finance sector is a non-trivial task and has been an atypical omission from recent scientific literature.
Where traditional segmentation classifies customers based on coarse features such as demographics, micro-segmentation depicts more nuanced differences between individuals.
Although ubiquitous in many industries, the proliferation of AI in sensitive industries such as finance has become contingent on the imperatives of responsible AI.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The micro-segmentation of customers in the finance sector is a non-trivial
task and has been an atypical omission from recent scientific literature. Where
traditional segmentation classifies customers based on coarse features such as
demographics, micro-segmentation depicts more nuanced differences between
individuals, bringing forth several advantages including the potential for
improved personalization in financial services. AI and representation learning
offer a unique opportunity to solve the problem of micro-segmentation. Although
ubiquitous in many industries, the proliferation of AI in sensitive industries
such as finance has become contingent on the imperatives of responsible AI. We
had previously solved the micro-segmentation problem by extracting temporal
features from the state space of a recurrent neural network (RNN). However, due
to the inherent opacity of RNNs our solution lacked an explanation - one of the
imperatives of responsible AI. In this study, we address this issue by
extracting an explanation for and providing an interpretation of our temporal
features. We investigate the state space of our RNN and through a linear
regression model reconstruct the trajectories in the state space with high
fidelity. We show that our linear regression coefficients have not only learned
the rules used to create the RNN's output data but have also learned the
relationships that were not directly evident in the raw data.
Related papers
- Global Convergence and Rich Feature Learning in $L$-Layer Infinite-Width Neural Networks under $μ$P Parametrization [66.03821840425539]
In this paper, we investigate the training dynamics of $L$-layer neural networks using the tensor gradient program (SGD) framework.
We show that SGD enables these networks to learn linearly independent features that substantially deviate from their initial values.
This rich feature space captures relevant data information and ensures that any convergent point of the training process is a global minimum.
arXiv Detail & Related papers (2025-03-12T17:33:13Z) - MAPPING: Debiasing Graph Neural Networks for Fair Node Classification
with Limited Sensitive Information Leakage [1.8238848494579714]
We propose a novel model-agnostic debiasing framework named MAPPING for fair node classification.
Our results show that MAPPING can achieve better trade-offs between utility and fairness, and privacy risks of sensitive information leakage.
arXiv Detail & Related papers (2024-01-23T14:59:46Z) - Bayesian Neural Networks for Macroeconomic Analysis [0.0]
We develop Bayesian neural networks (BNNs) that are well-suited for handling datasets commonly used for macroeconomic analysis in policy institutions.
Our approach avoids extensive specification searches through a novel mixture specification for the activation function.
We show that our BNNs produce precise density forecasts, typically better than those from other machine learning methods.
arXiv Detail & Related papers (2022-11-09T09:10:57Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Recurrent Neural Networks for Learning Long-term Temporal Dependencies
with Reanalysis of Time Scale Representation [16.32068729107421]
We argue that the interpretation of a forget gate as a temporal representation is valid when the gradient of loss with respect to the state decreases exponentially as time goes back.
We propose an approach to construct new RNNs that can represent a longer time scale than conventional models.
arXiv Detail & Related papers (2021-11-05T06:22:58Z) - Clustering in Recurrent Neural Networks for Micro-Segmentation using
Spending Personality [0.0]
Fine-grained customer segments are notoriously elusive and one method of obtaining them is through feature extraction.
We consider both temporal and non-sequential models, using long short-term memory (LSTM) and feed-forward neural networks, respectively.
We show that classification using these extracted features performs at least as well as bespoke models on two common metrics, namely loan default rate and customer liquidity index.
arXiv Detail & Related papers (2021-09-20T11:06:58Z) - Wide Network Learning with Differential Privacy [7.453881927237143]
Current generation of neural networks suffers significant loss accuracy under most practically relevant privacy training regimes.
We develop a general approach towards training these models that takes advantage of the sparsity of the gradients of private Empirical Minimization (ERM)
Following the same number of parameters, we propose a novel algorithm for privately training neural networks.
arXiv Detail & Related papers (2021-03-01T20:31:50Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Lipschitz Recurrent Neural Networks [100.72827570987992]
We show that our Lipschitz recurrent unit is more robust with respect to input and parameter perturbations as compared to other continuous-time RNNs.
Our experiments demonstrate that the Lipschitz RNN can outperform existing recurrent units on a range of benchmark tasks.
arXiv Detail & Related papers (2020-06-22T08:44:52Z) - Neural Networks and Value at Risk [59.85784504799224]
We perform Monte-Carlo simulations of asset returns for Value at Risk threshold estimation.
Using equity markets and long term bonds as test assets, we investigate neural networks.
We find our networks when fed with substantially less data to perform significantly worse.
arXiv Detail & Related papers (2020-05-04T17:41:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.