DiffsFormer: A Diffusion Transformer on Stock Factor Augmentation
- URL: http://arxiv.org/abs/2402.06656v1
- Date: Mon, 5 Feb 2024 03:54:36 GMT
- Title: DiffsFormer: A Diffusion Transformer on Stock Factor Augmentation
- Authors: Yuan Gao, Haokun Chen, Xiang Wang, Zhicai Wang, Xue Wang, Jinyang Gao,
Bolin Ding
- Abstract summary: We introduce the Diffusion Model to generate stock factors with Transformer architecture (DiffsFormer)
When presented with a specific downstream task, we employ DiffsFormer to augment the training procedure by editing existing samples.
The proposed method achieves relative improvements of 7.2% and 27.8% in annualized return ratio for the respective datasets.
- Score: 36.75453713794983
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine learning models have demonstrated remarkable efficacy and efficiency
in a wide range of stock forecasting tasks. However, the inherent challenges of
data scarcity, including low signal-to-noise ratio (SNR) and data homogeneity,
pose significant obstacles to accurate forecasting. To address this issue, we
propose a novel approach that utilizes artificial intelligence-generated
samples (AIGS) to enhance the training procedures. In our work, we introduce
the Diffusion Model to generate stock factors with Transformer architecture
(DiffsFormer). DiffsFormer is initially trained on a large-scale source domain,
incorporating conditional guidance so as to capture global joint distribution.
When presented with a specific downstream task, we employ DiffsFormer to
augment the training procedure by editing existing samples. This editing step
allows us to control the strength of the editing process, determining the
extent to which the generated data deviates from the target domain. To evaluate
the effectiveness of DiffsFormer augmented training, we conduct experiments on
the CSI300 and CSI800 datasets, employing eight commonly used machine learning
models. The proposed method achieves relative improvements of 7.2% and 27.8% in
annualized return ratio for the respective datasets. Furthermore, we perform
extensive experiments to gain insights into the functionality of DiffsFormer
and its constituent components, elucidating how they address the challenges of
data scarcity and enhance the overall model performance. Our research
demonstrates the efficacy of leveraging AIGS and the DiffsFormer architecture
to mitigate data scarcity in stock forecasting tasks.
Related papers
- A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception [78.26734070960886]
Current perceptive models heavily depend on resource-intensive datasets.
We introduce perception-aware loss (P.A. loss) through segmentation, improving both quality and controllability.
Our method customizes data augmentation by extracting and utilizing perception-aware attribute (P.A. Attr) during generation.
arXiv Detail & Related papers (2024-03-20T04:58:03Z) - FedD2S: Personalized Data-Free Federated Knowledge Distillation [19.975420988169454]
We propose a novel approach named FedD2S for Personalized Federated Learning (pFL), leveraging knowledge distillation.
FedD2S incorporates a deep-to-shallow layer-dropping mechanism in the data-free knowledge distillation process to enhance local model personalization.
The proposed approach demonstrates superior performance, characterized by accelerated convergence and improved fairness among clients.
arXiv Detail & Related papers (2024-02-16T17:36:51Z) - SMaRt: Improving GANs with Score Matching Regularity [94.81046452865583]
Generative adversarial networks (GANs) usually struggle in learning from highly diverse data, whose underlying manifold is complex.
We show that score matching serves as a promising solution to this issue thanks to its capability of persistently pushing the generated data points towards the real data manifold.
We propose to improve the optimization of GANs with score matching regularity (SMaRt)
arXiv Detail & Related papers (2023-11-30T03:05:14Z) - One-Shot Federated Learning with Classifier-Guided Diffusion Models [44.604485649167216]
One-shot federated learning (OSFL) has gained attention in recent years due to its low communication cost.
In this paper, we explore the novel opportunities that diffusion models bring to OSFL and propose FedCADO.
FedCADO generates data that complies with clients' distributions and subsequently training the aggregated model on the server.
arXiv Detail & Related papers (2023-11-15T11:11:25Z) - Data-Centric Long-Tailed Image Recognition [49.90107582624604]
Long-tail models exhibit a strong demand for high-quality data.
Data-centric approaches aim to enhance both the quantity and quality of data to improve model performance.
There is currently a lack of research into the underlying mechanisms explaining the effectiveness of information augmentation.
arXiv Detail & Related papers (2023-11-03T06:34:37Z) - AST: Effective Dataset Distillation through Alignment with Smooth and
High-Quality Expert Trajectories [18.266786462036553]
We propose an effective DD framework named AST, standing for Alignment with Smooth and high-quality expert Trajectories.
We conduct extensive experiments on datasets of different scales, sizes, and resolutions.
arXiv Detail & Related papers (2023-10-16T16:13:53Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.