KAN based Autoencoders for Factor Models
- URL: http://arxiv.org/abs/2408.02694v1
- Date: Sun, 4 Aug 2024 02:02:09 GMT
- Title: KAN based Autoencoders for Factor Models
- Authors: Tianqi Wang, Shubham Singh,
- Abstract summary: Inspired by recent advances in Kolmogorov-Arnold Networks (KANs), we introduce a novel approach to latent factor conditional asset pricing models.
Our method introduces a KAN-based autoencoder which surpasses models in both accuracy and interpretability.
Our model offers enhanced flexibility in approximating exposures as nonlinear functions of asset characteristics, while simultaneously providing users with an intuitive framework for interpreting latent factors.
- Score: 13.512750745176664
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Inspired by recent advances in Kolmogorov-Arnold Networks (KANs), we introduce a novel approach to latent factor conditional asset pricing models. While previous machine learning applications in asset pricing have predominantly used Multilayer Perceptrons with ReLU activation functions to model latent factor exposures, our method introduces a KAN-based autoencoder which surpasses MLP models in both accuracy and interpretability. Our model offers enhanced flexibility in approximating exposures as nonlinear functions of asset characteristics, while simultaneously providing users with an intuitive framework for interpreting latent factors. Empirical backtesting demonstrates our model's superior ability to explain cross-sectional risk exposures. Moreover, long-short portfolios constructed using our model's predictions achieve higher Sharpe ratios, highlighting its practical value in investment management.
Related papers
- Scalable Language Models with Posterior Inference of Latent Thought Vectors [52.63299874322121]
Latent-Thought Language Models (LTMs) incorporate explicit latent thought vectors that follow an explicit prior model in latent space.
LTMs possess additional scaling dimensions beyond traditional LLMs, yielding a structured design space.
LTMs significantly outperform conventional autoregressive models and discrete diffusion models in validation perplexity and zero-shot language modeling.
arXiv Detail & Related papers (2025-02-03T17:50:34Z) - Disentangling Length Bias In Preference Learning Via Response-Conditioned Modeling [87.17041933863041]
We introduce a Response-conditioned Bradley-Terry (Rc-BT) model that enhances the reward model's capability in length bias mitigating and length instruction following.
We also propose the Rc-DPO algorithm to leverage the Rc-BT model for direct policy optimization (DPO) of large language models.
arXiv Detail & Related papers (2025-02-02T14:50:25Z) - STORM: A Spatio-Temporal Factor Model Based on Dual Vector Quantized Variational Autoencoders for Financial Trading [55.02735046724146]
In financial trading, factor models are widely used to price assets and capture excess returns from mispricing.
We propose a Spatio-Temporal factOR Model based on dual vector quantized variational autoencoders, named STORM.
Storm extracts features of stocks from temporal and spatial perspectives, then fuses and aligns these features at the fine-grained and semantic level, and represents the factors as multi-dimensional embeddings.
arXiv Detail & Related papers (2024-12-12T17:15:49Z) - KACDP: A Highly Interpretable Credit Default Prediction Model [2.776411854233918]
This paper introduces a method based on Kolmogorov-Arnold Networks (KANs)
KANs is a new type of neural network architecture with learnable activation functions and no linear weights.
Experiments show that the KACDP model outperforms mainstream credit default prediction models in performance metrics.
arXiv Detail & Related papers (2024-11-26T12:58:03Z) - NeuralFactors: A Novel Factor Learning Approach to Generative Modeling of Equities [0.0]
We introduce NeuralFactors, a novel machine-learning based approach to factor analysis where a neural network outputs factor exposures and factor returns.
We show that this model outperforms prior approaches in terms of log-likelihood performance and computational efficiency.
arXiv Detail & Related papers (2024-08-02T18:01:09Z) - Deep Partial Least Squares for Empirical Asset Pricing [0.4511923587827302]
We use deep partial least squares (DPLS) to estimate an asset pricing model for individual stock returns.
The novel contribution is to resolve the nonlinear factor structure, thus advancing the current paradigm of deep learning in empirical asset pricing.
arXiv Detail & Related papers (2022-06-20T21:30:39Z) - Deep Sequence Modeling: Development and Applications in Asset Pricing [35.027865343844766]
We predict asset returns and measure risk premia using a prominent technique from artificial intelligence -- deep sequence modeling.
Because asset returns often exhibit sequential dependence that may not be effectively captured by conventional time series models, sequence modeling offers a promising path with its data-driven approach and superior performance.
arXiv Detail & Related papers (2021-08-20T04:40:55Z) - Closed-form Continuous-Depth Models [99.40335716948101]
Continuous-depth neural models rely on advanced numerical differential equation solvers.
We present a new family of models, termed Closed-form Continuous-depth (CfC) networks, that are simple to describe and at least one order of magnitude faster.
arXiv Detail & Related papers (2021-06-25T22:08:51Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - On the model-based stochastic value gradient for continuous
reinforcement learning [50.085645237597056]
We show that simple model-based agents can outperform state-of-the-art model-free agents in terms of both sample-efficiency and final reward.
Our findings suggest that model-based policy evaluation deserves closer attention.
arXiv Detail & Related papers (2020-08-28T17:58:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.