SuperDeConFuse: A Supervised Deep Convolutional Transform based Fusion
Framework for Financial Trading Systems
- URL: http://arxiv.org/abs/2011.04364v1
- Date: Mon, 9 Nov 2020 11:58:12 GMT
- Title: SuperDeConFuse: A Supervised Deep Convolutional Transform based Fusion
Framework for Financial Trading Systems
- Authors: Pooja Gupta, Angshul Majumdar, Emilie Chouzenoux, Giovanni Chierchia
- Abstract summary: This work proposes a supervised multi-channel time-series learning framework for financial stock trading.
Our approach consists of processing the data channels through separate 1-D convolution layers, then fusing the outputs with a series of fully-connected layers, and finally applying a softmax classification layer.
Numerical experiments confirm that the proposed model yields considerably better results than state-of-the-art deep learning techniques for real-world problem of stock trading.
- Score: 29.411173536818477
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work proposes a supervised multi-channel time-series learning framework
for financial stock trading. Although many deep learning models have recently
been proposed in this domain, most of them treat the stock trading time-series
data as 2-D image data, whereas its true nature is 1-D time-series data. Since
the stock trading systems are multi-channel data, many existing techniques
treating them as 1-D time-series data are not suggestive of any technique to
effectively fusion the information carried by the multiple channels. To
contribute towards both of these shortcomings, we propose an end-to-end
supervised learning framework inspired by the previously established
(unsupervised) convolution transform learning framework. Our approach consists
of processing the data channels through separate 1-D convolution layers, then
fusing the outputs with a series of fully-connected layers, and finally
applying a softmax classification layer. The peculiarity of our framework -
SuperDeConFuse (SDCF), is that we remove the nonlinear activation located
between the multi-channel convolution layers and the fully-connected layers, as
well as the one located between the latter and the output layer. We compensate
for this removal by introducing a suitable regularization on the aforementioned
layer outputs and filters during the training phase. Specifically, we apply a
logarithm determinant regularization on the layer filters to break symmetry and
force diversity in the learnt transforms, whereas we enforce the non-negativity
constraint on the layer outputs to mitigate the issue of dead neurons. This
results in the effective learning of a richer set of features and filters with
respect to a standard convolutional neural network. Numerical experiments
confirm that the proposed model yields considerably better results than
state-of-the-art deep learning techniques for real-world problem of stock
trading.
Related papers
- A Framework for Fine-Tuning LLMs using Heterogeneous Feedback [69.51729152929413]
We present a framework for fine-tuning large language models (LLMs) using heterogeneous feedback.
First, we combine the heterogeneous feedback data into a single supervision format, compatible with methods like SFT and RLHF.
Next, given this unified feedback dataset, we extract a high-quality and diverse subset to obtain performance increases.
arXiv Detail & Related papers (2024-08-05T23:20:32Z) - Multi-Epoch learning with Data Augmentation for Deep Click-Through Rate Prediction [53.88231294380083]
We introduce a novel Multi-Epoch learning with Data Augmentation (MEDA) framework, suitable for both non-continual and continual learning scenarios.
MEDA minimizes overfitting by reducing the dependency of the embedding layer on subsequent training data.
Our findings confirm that pre-trained layers can adapt to new embedding spaces, enhancing performance without overfitting.
arXiv Detail & Related papers (2024-06-27T04:00:15Z) - Entropy Guided Extrapolative Decoding to Improve Factuality in Large Language Models [55.45444773200529]
Large language models (LLMs) exhibit impressive natural language capabilities but suffer from hallucination.
Recent work has focused on decoding techniques to improve factuality during inference.
arXiv Detail & Related papers (2024-04-14T19:45:35Z) - GIFD: A Generative Gradient Inversion Method with Feature Domain
Optimization [52.55628139825667]
Federated Learning (FL) has emerged as a promising distributed machine learning framework to preserve clients' privacy.
Recent studies find that an attacker can invert the shared gradients and recover sensitive data against an FL system by leveraging pre-trained generative adversarial networks (GAN) as prior knowledge.
We propose textbfGradient textbfInversion over textbfFeature textbfDomains (GIFD), which disassembles the GAN model and searches the feature domains of the intermediate layers.
arXiv Detail & Related papers (2023-08-09T04:34:21Z) - Layer-wise Linear Mode Connectivity [52.6945036534469]
Averaging neural network parameters is an intuitive method for the knowledge of two independent models.
It is most prominently used in federated learning.
We analyse the performance of the models that result from averaging single, or groups.
arXiv Detail & Related papers (2023-07-13T09:39:10Z) - Large Scale Time-Series Representation Learning via Simultaneous Low and
High Frequency Feature Bootstrapping [7.0064929761691745]
We propose a non-contrastive self-supervised learning approach efficiently captures low and high-frequency time-varying features.
Our method takes raw time series data as input and creates two different augmented views for two branches of the model.
To demonstrate the robustness of our model we performed extensive experiments and ablation studies on five real-world time-series datasets.
arXiv Detail & Related papers (2022-04-24T14:39:47Z) - Data-efficient Alignment of Multimodal Sequences by Aligning Gradient
Updates and Internal Feature Distributions [36.82512331179322]
Recent research suggests that network components dealing with different modalities may overfit and generalize at different speeds, creating difficulties for training.
We propose layer-wise adaptive rate scaling (LARS) to align the magnitudes of gradient updates in different layers and balance the pace of learning.
We also use sequence-wise batch normalization (SBN) to align the internal feature distributions from different modalities.
arXiv Detail & Related papers (2020-11-15T13:04:25Z) - ConFuse: Convolutional Transform Learning Fusion Framework For
Multi-Channel Data Analysis [29.58965424136611]
We propose an unsupervised fusion framework based on %the recently proposed convolutional transform learning.
We apply the framework to multi-channel financial data for stock forecasting and trading.
arXiv Detail & Related papers (2020-11-09T10:41:28Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.