Machine Learning Models in Stock Market Prediction
- URL: http://arxiv.org/abs/2202.09359v1
- Date: Sun, 6 Feb 2022 10:33:42 GMT
- Title: Machine Learning Models in Stock Market Prediction
- Authors: Gurjeet Singh
- Abstract summary: The paper focuses on predicting the Nifty 50 Index by using 8 Supervised Machine Learning Models.
Experiments are based on historical data of Nifty 50 Index of Indian Stock Market from 22nd April, 1996 to 16th April, 2021.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The paper focuses on predicting the Nifty 50 Index by using 8 Supervised
Machine Learning Models. The techniques used for empirical study are Adaptive
Boost (AdaBoost), k-Nearest Neighbors (kNN), Linear Regression (LR), Artificial
Neural Network (ANN), Random Forest (RF), Stochastic Gradient Descent (SGD),
Support Vector Machine (SVM) and Decision Trees (DT). Experiments are based on
historical data of Nifty 50 Index of Indian Stock Market from 22nd April, 1996
to 16th April, 2021, which is time series data of around 25 years. During the
period there were 6220 trading days excluding all the non trading days. The
entire trading dataset was divided into 4 subsets of different size-25% of
entire data, 50% of entire data, 75% of entire data and entire data. Each
subset was further divided into 2 parts-training data and testing data. After
applying 3 tests- Test on Training Data, Test on Testing Data and Cross
Validation Test on each subset, the prediction performance of the used models
were compared and after comparison, very interesting results were found. The
evaluation results indicate that Adaptive Boost, k- Nearest Neighbors, Random
Forest and Decision Trees under performed with increase in the size of data
set. Linear Regression and Artificial Neural Network shown almost similar
prediction results among all the models but Artificial Neural Network took more
time in training and validating the model. Thereafter Support Vector Machine
performed better among rest of the models but with increase in the size of data
set, Stochastic Gradient Descent performed better than Support Vector Machine.
Related papers
- Inferring Data Preconditions from Deep Learning Models for Trustworthy
Prediction in Deployment [25.527665632625627]
It is important to reason about the trustworthiness of the model's predictions with unseen data during deployment.
Existing methods for specifying and verifying traditional software are insufficient for this task.
We propose a novel technique that uses rules derived from neural network computations to infer data preconditions.
arXiv Detail & Related papers (2024-01-26T03:47:18Z) - Learning Defect Prediction from Unrealistic Data [57.53586547895278]
Pretrained models of code have become popular choices for code understanding and generation tasks.
Such models tend to be large and require commensurate volumes of training data.
It has become popular to train models with far larger but less realistic datasets, such as functions with artificially injected bugs.
Models trained on such data tend to only perform well on similar data, while underperforming on real world programs.
arXiv Detail & Related papers (2023-11-02T01:51:43Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Exploring the Effectiveness of Dataset Synthesis: An application of
Apple Detection in Orchards [68.95806641664713]
We explore the usability of Stable Diffusion 2.1-base for generating synthetic datasets of apple trees for object detection.
We train a YOLOv5m object detection model to predict apples in a real-world apple detection dataset.
Results demonstrate that the model trained on generated data is slightly underperforming compared to a baseline model trained on real-world images.
arXiv Detail & Related papers (2023-06-20T09:46:01Z) - Boosted Dynamic Neural Networks [53.559833501288146]
A typical EDNN has multiple prediction heads at different layers of the network backbone.
To optimize the model, these prediction heads together with the network backbone are trained on every batch of training data.
Treating training and testing inputs differently at the two phases will cause the mismatch between training and testing data distributions.
We formulate an EDNN as an additive model inspired by gradient boosting, and propose multiple training techniques to optimize the model effectively.
arXiv Detail & Related papers (2022-11-30T04:23:12Z) - Ensemble Machine Learning Model Trained on a New Synthesized Dataset
Generalizes Well for Stress Prediction Using Wearable Devices [3.006016887654771]
We investigate the generalization ability of models built on datasets containing a small number of subjects, recorded in single study protocols.
We propose and evaluate the use of ensemble techniques by combining gradient boosting with an artificial neural network to measure predictive power on new, unseen data.
arXiv Detail & Related papers (2022-09-30T00:20:57Z) - Performance and Interpretability Comparisons of Supervised Machine
Learning Algorithms: An Empirical Study [3.7881729884531805]
The paper is organized in a findings-based manner, with each section providing general conclusions.
Overall, XGB and FFNNs were competitive, with FFNNs showing better performance in smooth models.
RF did not perform well in general, confirming the findings in the literature.
arXiv Detail & Related papers (2022-04-27T12:04:33Z) - Comparing Test Sets with Item Response Theory [53.755064720563]
We evaluate 29 datasets using predictions from 18 pretrained Transformer models on individual test examples.
We find that Quoref, HellaSwag, and MC-TACO are best suited for distinguishing among state-of-the-art models.
We also observe span selection task format, which is used for QA datasets like QAMR or SQuAD2.0, is effective in differentiating between strong and weak models.
arXiv Detail & Related papers (2021-06-01T22:33:53Z) - ALT-MAS: A Data-Efficient Framework for Active Testing of Machine
Learning Algorithms [58.684954492439424]
We propose a novel framework to efficiently test a machine learning model using only a small amount of labeled test data.
The idea is to estimate the metrics of interest for a model-under-test using Bayesian neural network (BNN)
arXiv Detail & Related papers (2021-04-11T12:14:04Z) - Stock Price Prediction Using CNN and LSTM-Based Deep Learning Models [0.0]
This paper presents a suite of deep learning based models for stock price prediction.
We use the historical records of the NIFTY 50 index listed in the National Stock Exchange of India.
Our proposition includes two regression models built on convolutional neural networks and three long and short term memory network based predictive models.
arXiv Detail & Related papers (2020-10-22T03:09:07Z) - Stock Price Prediction Using Machine Learning and LSTM-Based Deep
Learning Models [1.335161061703997]
We propose an approach of hybrid modeling for stock price prediction building different machine learning and deep learning-based models.
We have used NIFTY 50 index values of the National Stock Exchange (NSE) of India during the period December 29, 2014 till July 31, 2020.
We exploit the power of LSTM regression models in forecasting the future NIFTY 50 open values using four different models that differ in their architecture and in the structure of their input data.
arXiv Detail & Related papers (2020-09-20T20:32:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.