Benchmarking Machine Learning Models to Predict Corporate Bankruptcy
- URL: http://arxiv.org/abs/2212.12051v1
- Date: Thu, 22 Dec 2022 22:01:25 GMT
- Title: Benchmarking Machine Learning Models to Predict Corporate Bankruptcy
- Authors: Emmanuel Alanis, Sudheer Chava, Agam Shah
- Abstract summary: Using a comprehensive sample of 2,585 bankruptcies from 1990 to 2019, we benchmark the performance of various machine learning models.
We find that gradient boosted trees outperform other models in one-year-ahead forecasts.
- Score: 1.933681537640272
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Using a comprehensive sample of 2,585 bankruptcies from 1990 to 2019, we
benchmark the performance of various machine learning models in predicting
financial distress of publicly traded U.S. firms. We find that gradient boosted
trees outperform other models in one-year-ahead forecasts. Variable permutation
tests show that excess stock returns, idiosyncratic risk, and relative size are
the more important variables for predictions. Textual features derived from
corporate filings do not improve performance materially. In a credit
competition model that accounts for the asymmetric cost of default
misclassification, the survival random forest is able to capture large dollar
profits.
Related papers
- Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - Diffusion Variational Autoencoder for Tackling Stochasticity in
Multi-Step Regression Stock Price Prediction [54.21695754082441]
Multi-step stock price prediction over a long-term horizon is crucial for forecasting its volatility.
Current solutions to multi-step stock price prediction are mostly designed for single-step, classification-based predictions.
We combine a deep hierarchical variational-autoencoder (VAE) and diffusion probabilistic techniques to do seq2seq stock prediction.
Our model is shown to outperform state-of-the-art solutions in terms of its prediction accuracy and variance.
arXiv Detail & Related papers (2023-08-18T16:21:15Z) - Improved Bayes Risk Can Yield Reduced Social Welfare Under Competition [99.7047087527422]
In this work, we demonstrate that competition can fundamentally alter the behavior of machine learning scaling trends.
We find many settings where improving data representation quality decreases the overall predictive accuracy across users.
At a conceptual level, our work suggests that favorable scaling trends for individual model-providers need not translate to downstream improvements in social welfare.
arXiv Detail & Related papers (2023-06-26T13:06:34Z) - Predicting Stock Price Movement after Disclosure of Corporate Annual
Reports: A Case Study of 2021 China CSI 300 Stocks [4.5885930040346565]
This work study the predicting the tendency of the stock price on the second day right after the disclosure of the companies' annual reports.
We use a variety of different models, including decision tree, logistic regression, random forest, neural network, prototypical networks.
We conclude that according to the financial indicators based on the just-released annual report of the company, the predictability of the stock price movement is weak.
arXiv Detail & Related papers (2022-06-25T01:54:53Z) - Learning to Predict Trustworthiness with Steep Slope Loss [69.40817968905495]
We study the problem of predicting trustworthiness on real-world large-scale datasets.
We observe that the trustworthiness predictors trained with prior-art loss functions are prone to view both correct predictions and incorrect predictions to be trustworthy.
We propose a novel steep slope loss to separate the features w.r.t. correct predictions from the ones w.r.t. incorrect predictions by two slide-like curves that oppose each other.
arXiv Detail & Related papers (2021-09-30T19:19:09Z) - Interpretability in Safety-Critical FinancialTrading Systems [15.060749321774136]
In 2020, some of the world's most sophisticated quant hedge funds suffered losses.
We implement a gradient-based approach for precisely stress-testing how a trading model's forecasts can be manipulated.
We find our approach discovers seemingly in-sample input settings that result in large negative shifts in return distributions.
arXiv Detail & Related papers (2021-09-24T17:05:58Z) - Stock Price Prediction Under Anomalous Circumstances [81.37657557441649]
This paper aims to capture the movement pattern of stock prices under anomalous circumstances.
We train ARIMA and LSTM models at the single-stock level, industry level, and general market level.
Based on 100 companies' stock prices in the period of 2016 to 2020, the models achieve an average prediction accuracy of 98%.
arXiv Detail & Related papers (2021-09-14T18:50:38Z) - Machine Learning Classification Methods and Portfolio Allocation: An
Examination of Market Efficiency [3.3343612552681945]
We design a novel framework to examine market efficiency through out-of-sample (OOS) predictability.
We frame the asset pricing problem as a machine learning classification problem and construct classification models to predict return states.
The prediction-based portfolios beat the market with significant OOS economic gains.
arXiv Detail & Related papers (2021-08-04T20:48:27Z) - Economic Recession Prediction Using Deep Neural Network [26.504845007567972]
We identify the deep learning methodology of Bi-LSTM with Autoencoder as the most accurate model to forecast the beginning and end of economic recessions in the U.S.
We adopt commonly-available macro and market-condition features to compare the ability of different machine learning models to generate good predictions both in-sample and out-of-sample.
arXiv Detail & Related papers (2021-07-21T22:55:14Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - Models, Markets, and the Forecasting of Elections [3.8138805042090325]
We find systematic differences in accuracy over time, with markets performing better several months before the election, and the model performing better as the election approached.
We propose a market design that incorporates model forecasts via a trading bot to generate synthetic predictions.
arXiv Detail & Related papers (2021-02-06T19:05:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.