Transfer Ranking in Finance: Applications to Cross-Sectional Momentum
with Data Scarcity
- URL: http://arxiv.org/abs/2208.09968v2
- Date: Wed, 24 Aug 2022 10:12:45 GMT
- Title: Transfer Ranking in Finance: Applications to Cross-Sectional Momentum
with Data Scarcity
- Authors: Daniel Poh, Stephen Roberts and Stefan Zohren
- Abstract summary: We introduce Fused Networks -- a novel and hybrid parameter-sharing transfer ranking model.
The model fuses information extracted using an encoder-attention module operated on a source dataset.
It mitigates the issue of models with poor generalisability that are a consequence of training on scarce target data.
- Score: 2.3204178451683264
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cross-sectional strategies are a classical and popular trading style, with
recent high performing variants incorporating sophisticated neural
architectures. While these strategies have been applied successfully to
data-rich settings involving mature assets with long histories, deploying them
on instruments with limited samples generally produce over-fitted models with
degraded performance. In this paper, we introduce Fused Encoder Networks -- a
novel and hybrid parameter-sharing transfer ranking model. The model fuses
information extracted using an encoder-attention module operated on a source
dataset with a similar but separate module focused on a smaller target dataset
of interest. This mitigates the issue of models with poor generalisability that
are a consequence of training on scarce target data. Additionally, the
self-attention mechanism enables interactions among instruments to be accounted
for, not just at the loss level during model training, but also at inference
time. Focusing on momentum applied to the top ten cryptocurrencies by market
capitalisation as a demonstrative use-case, the Fused Encoder Networks
outperforms the reference benchmarks on most performance measures, delivering a
three-fold boost in the Sharpe ratio over classical momentum as well as an
improvement of approximately 50% against the best benchmark model without
transaction costs. It continues outperforming baselines even after accounting
for the high transaction costs associated with trading cryptocurrencies.
Related papers
- Semi-Supervised Reward Modeling via Iterative Self-Training [52.48668920483908]
We propose Semi-Supervised Reward Modeling (SSRM), an approach that enhances RM training using unlabeled data.
We demonstrate that SSRM significantly improves reward models without incurring additional labeling costs.
Overall, SSRM substantially reduces the dependency on large volumes of human-annotated data, thereby decreasing the overall cost and time involved in training effective reward models.
arXiv Detail & Related papers (2024-09-10T22:57:58Z) - Large-scale Time-Varying Portfolio Optimisation using Graph Attention Networks [4.2056926734482065]
This is the first study to incorporate risky firms and use all the firms in portfolio optimisation.
We propose and empirically test a novel method that leverages Graph Attention networks (GATs)
GATs are deep learning-based models that exploit network data to uncover nonlinear relationships.
arXiv Detail & Related papers (2024-07-22T10:50:47Z) - Towards Robust Federated Learning via Logits Calibration on Non-IID Data [49.286558007937856]
Federated learning (FL) is a privacy-preserving distributed management framework based on collaborative model training of distributed devices in edge networks.
Recent studies have shown that FL is vulnerable to adversarial examples, leading to a significant drop in its performance.
In this work, we adopt the adversarial training (AT) framework to improve the robustness of FL models against adversarial example (AE) attacks.
arXiv Detail & Related papers (2024-03-05T09:18:29Z) - Enhancing Multiple Reliability Measures via Nuisance-extended
Information Bottleneck [77.37409441129995]
In practical scenarios where training data is limited, many predictive signals in the data can be rather from some biases in data acquisition.
We consider an adversarial threat model under a mutual information constraint to cover a wider class of perturbations in training.
We propose an autoencoder-based training to implement the objective, as well as practical encoder designs to facilitate the proposed hybrid discriminative-generative training.
arXiv Detail & Related papers (2023-03-24T16:03:21Z) - Augmented Bilinear Network for Incremental Multi-Stock Time-Series
Classification [83.23129279407271]
We propose a method to efficiently retain the knowledge available in a neural network pre-trained on a set of securities.
In our method, the prior knowledge encoded in a pre-trained neural network is maintained by keeping existing connections fixed.
This knowledge is adjusted for the new securities by a set of augmented connections, which are optimized using the new data.
arXiv Detail & Related papers (2022-07-23T18:54:10Z) - Compatible deep neural network framework with financial time series
data, including data preprocessor, neural network model and trading strategy [2.347843817145202]
This research introduces a new deep neural network architecture and a novel idea of how to prepare financial data before feeding them to the model.
Three different datasets are used to evaluate this method, where results indicate that this framework can provide us with profitable and robust predictions.
arXiv Detail & Related papers (2022-05-11T20:44:08Z) - Long Short-Term Memory Neural Network for Financial Time Series [0.0]
We present an ensemble of independent and parallel long short-term memory neural networks for the prediction of stock price movement.
With a straightforward trading strategy, comparisons with a randomly chosen portfolio and a portfolio containing all the stocks in the index show that the portfolio resulting from the LSTM ensemble provides better average daily returns and higher cumulative returns over time.
arXiv Detail & Related papers (2022-01-20T15:17:26Z) - Evaluating data augmentation for financial time series classification [85.38479579398525]
We evaluate several augmentation methods applied to stocks datasets using two state-of-the-art deep learning models.
For a relatively small dataset augmentation methods achieve up to $400%$ improvement in risk adjusted return performance.
For a larger stock dataset augmentation methods achieve up to $40%$ improvement.
arXiv Detail & Related papers (2020-10-28T17:53:57Z) - Knowledge-Enriched Distributional Model Inversion Attacks [49.43828150561947]
Model inversion (MI) attacks are aimed at reconstructing training data from model parameters.
We present a novel inversion-specific GAN that can better distill knowledge useful for performing attacks on private models from public data.
Our experiments show that the combination of these techniques can significantly boost the success rate of the state-of-the-art MI attacks by 150%.
arXiv Detail & Related papers (2020-10-08T16:20:48Z) - DoubleEnsemble: A New Ensemble Method Based on Sample Reweighting and
Feature Selection for Financial Data Analysis [22.035287788330663]
We propose DoubleEnsemble, an ensemble framework leveraging learning trajectory based sample reweighting and shuffling based feature selection.
Our model is applicable to a wide range of base models, capable of extracting complex patterns, while mitigating the overfitting and instability issues for financial market prediction.
arXiv Detail & Related papers (2020-10-03T02:57:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.