Bankruptcy prediction using disclosure text features
- URL: http://arxiv.org/abs/2101.00719v1
- Date: Sun, 3 Jan 2021 22:23:22 GMT
- Title: Bankruptcy prediction using disclosure text features
- Authors: Sridhar Ravula
- Abstract summary: This work proposes a new distress dictionary based on the sentences used by managers in explaining financial status.
It demonstrates the significant differences in linguistic features between bankrupt and non-bankrupt firms.
Using a large sample of 500 bankrupt firms, it builds predictive models and compares the performance against two dictionaries used in financial text analysis.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A public firm's bankruptcy prediction is an important financial research
problem because of the security price downside risks. Traditional methods rely
on accounting metrics that suffer from shortcomings like window dressing and
retrospective focus. While disclosure text-based metrics overcome some of these
issues, current methods excessively focus on disclosure tone and sentiment.
There is a requirement to relate meaningful signals in the disclosure text to
financial outcomes and quantify the disclosure text data. This work proposes a
new distress dictionary based on the sentences used by managers in explaining
financial status. It demonstrates the significant differences in linguistic
features between bankrupt and non-bankrupt firms. Further, using a large sample
of 500 bankrupt firms, it builds predictive models and compares the performance
against two dictionaries used in financial text analysis. This research shows
that the proposed stress dictionary captures unique information from
disclosures and the predictive models based on its features have the highest
accuracy.
Related papers
- Detection of Temporality at Discourse Level on Financial News by Combining Natural Language Processing and Machine Learning [8.504685056067144]
Finance-related news such as Bloomberg News, CNN Business and Forbes are valuable sources of real data for market screening systems.
We propose a novel system to detect the temporality of finance-related news at discourse level.
We have tested our system on a labelled dataset of finance-related news annotated by researchers with knowledge in the field.
arXiv Detail & Related papers (2024-03-30T16:40:10Z) - On the Societal Impact of Open Foundation Models [93.67389739906561]
We focus on open foundation models, defined here as those with broadly available model weights.
We identify five distinctive properties of open foundation models that lead to both their benefits and risks.
arXiv Detail & Related papers (2024-02-27T16:49:53Z) - Interpretable Automatic Fine-grained Inconsistency Detection in Text
Summarization [56.94741578760294]
We propose the task of fine-grained inconsistency detection, the goal of which is to predict the fine-grained types of factual errors in a summary.
Motivated by how humans inspect factual inconsistency in summaries, we propose an interpretable fine-grained inconsistency detection model, FineGrainFact.
arXiv Detail & Related papers (2023-05-23T22:11:47Z) - Measuring Consistency in Text-based Financial Forecasting Models [10.339586273664725]
FinTrust is an evaluation tool that assesses logical consistency in financial text.
We show that the consistency of state-of-the-art NLP models for financial forecasting is poor.
Our analysis of the performance degradation caused by meaning-preserving alternations suggests that current text-based methods are not suitable for robustly predicting market information.
arXiv Detail & Related papers (2023-05-15T10:32:26Z) - Can ChatGPT Forecast Stock Price Movements? Return Predictability and Large Language Models [51.3422222472898]
We document the capability of large language models (LLMs) like ChatGPT to predict stock price movements using news headlines.
We develop a theoretical model incorporating information capacity constraints, underreaction, limits-to-arbitrage, and LLMs.
arXiv Detail & Related papers (2023-04-15T19:22:37Z) - A Word is Worth A Thousand Dollars: Adversarial Attack on Tweets Fools
Stock Prediction [100.9772316028191]
In this paper, we experiment with a variety of adversarial attack configurations to fool three stock prediction victim models.
Our results show that the proposed attack method can achieve consistent success rates and cause significant monetary loss in trading simulation.
arXiv Detail & Related papers (2022-05-01T05:12:22Z) - Forecasting Cryptocurrency Returns from Sentiment Signals: An Analysis
of BERT Classifiers and Weak Supervision [6.624726878647541]
We introduce weak learning, a recently proposed NLP approach to address the problem that text data is unlabeled.
We confirm that finetuning using weak labels enhances the predictive value of text-based features and raises forecast accuracy in the context of predicting cryptocurrency returns.
More fundamentally, the modeling paradigm we present, weak labeling domain-specific text and finetuning pretrained NLP models, is universally applicable in (financial) forecasting.
arXiv Detail & Related papers (2022-04-06T07:45:05Z) - Fighting Money Laundering with Statistics and Machine Learning [95.42181254494287]
There is little scientific literature on statistical and machine learning methods for anti-money laundering.
We propose a unifying terminology with two central elements: (i) client risk profiling and (ii) suspicious behavior flagging.
arXiv Detail & Related papers (2022-01-11T21:31:18Z) - Text analysis in financial disclosures [0.0]
Most of the information in a firm's financial disclosures is in unstructured text.
Researchers have started analyzing text content in disclosures recently.
This work contributes to disclosure analysis methods by highlighting the limitations of the current focus on sentiment metrics.
arXiv Detail & Related papers (2021-01-06T17:45:40Z) - Detecting Word Sense Disambiguation Biases in Machine Translation for
Model-Agnostic Adversarial Attacks [84.61578555312288]
We introduce a method for the prediction of disambiguation errors based on statistical data properties.
We develop a simple adversarial attack strategy that minimally perturbs sentences in order to elicit disambiguation errors.
Our findings indicate that disambiguation robustness varies substantially between domains and that different models trained on the same data are vulnerable to different attacks.
arXiv Detail & Related papers (2020-11-03T17:01:44Z) - Uncertainty over Uncertainty: Investigating the Assumptions,
Annotations, and Text Measurements of Economic Policy Uncertainty [12.787262921924953]
We examine an economic index that measures economic policy uncertainty from keyword occurrences in news.
We find some annotator disagreements of economic policy uncertainty can be attributed to ambiguity in language.
arXiv Detail & Related papers (2020-10-09T17:50:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.