Multimodal Generative Models for Bankruptcy Prediction Using Textual
Data
- URL: http://arxiv.org/abs/2211.08405v5
- Date: Sat, 24 Feb 2024 11:08:24 GMT
- Title: Multimodal Generative Models for Bankruptcy Prediction Using Textual
Data
- Authors: Rogelio A. Mancisidor and Kjersti Aas
- Abstract summary: This research introduces the Multimodal Discriminative Conditional (CMMD) model that learns multimodal representations that embed information from accounting, market, and textual data modalities.
With this novel methodology, it is realistic to use textual data in bankruptcy prediction models, since accounting and market data are available for all companies, unlike textual data.
- Score: 2.1756081703276
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Textual data from financial filings, e.g., the Management's Discussion &
Analysis (MDA) section in Form 10-K, has been used to improve the prediction
accuracy of bankruptcy models. In practice, however, we cannot obtain the MDA
section for all public companies, which limits the use of MDA data in
traditional bankruptcy models, as they need complete data to make predictions.
The two main reasons for the lack of MDA are: (i) not all companies are obliged
to submit the MDA and (ii) technical problems arise when crawling and scrapping
the MDA section. To solve this limitation, this research introduces the
Conditional Multimodal Discriminative (CMMD) model that learns multimodal
representations that embed information from accounting, market, and textual
data modalities. The CMMD model needs a sample with all data modalities for
model training. At test time, the CMMD model only needs access to accounting
and market modalities to generate multimodal representations, which are further
used to make bankruptcy predictions and to generate words from the missing MDA
modality. With this novel methodology, it is realistic to use textual data in
bankruptcy prediction models, since accounting and market data are available
for all companies, unlike textual data. The empirical results of this research
show that if financial regulators, or investors, were to use traditional models
using MDA data, they would only be able to make predictions for 60% of the
companies. Furthermore, the classification performance of our proposed
methodology is superior to that of a large number of traditional classifier
models, taking into account all the companies in our sample.
Related papers
- Benchmarking Multi-modal Semantic Segmentation under Sensor Failures: Missing and Noisy Modality Robustness [61.87055159919641]
Multi-modal semantic segmentation (MMSS) addresses the limitations of single-modality data by integrating complementary information across modalities.
Despite notable progress, a significant gap persists between research and real-world deployment due to variability and uncertainty in multi-modal data quality.
We introduce a robustness benchmark that evaluates MMSS models under three scenarios: Entire-Missing Modality (EMM), Random-Missing Modality (RMM), and Noisy Modality (NM)
arXiv Detail & Related papers (2025-03-24T08:46:52Z) - Triad: Empowering LMM-based Anomaly Detection with Vision Expert-guided Visual Tokenizer and Manufacturing Process [67.99194145865165]
We modify the AnyRes structure of the LLaVA model to provide the potential anomalous areas identified by existing IAD models to the LMMs.
Considering that the generation of defects is closely related to the manufacturing process, we propose a manufacturing-driven IAD paradigm.
We present Triad, a novel LMM-based method incorporating an expert-guided region-of-interest tokenizer and manufacturing process.
arXiv Detail & Related papers (2025-03-17T13:56:57Z) - AMA-LSTM: Pioneering Robust and Fair Financial Audio Analysis for Stock Volatility Prediction [25.711345527738068]
multimodal methods have faced two drawbacks.
They often fail to yield reliable models and overfit the data due to their absorption of information from the stock market.
Using multimodal models to predict stock volatility suffers from gender bias and lacks an efficient way to eliminate such bias.
Our comprehensive experiments on robustness-world financial audio datasets reveal that this method exceeds the performance of current state-of-the-art solution.
arXiv Detail & Related papers (2024-07-03T18:40:53Z) - Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning [79.53130089003986]
Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
arXiv Detail & Related papers (2023-10-05T09:09:44Z) - Incorporating Pre-trained Model Prompting in Multimodal Stock Volume
Movement Prediction [22.949484374773967]
We propose the Prompt-based MUltimodal Stock volumE prediction model (ProMUSE) to process text and time series modalities.
We use pre-trained language models for better comprehension of financial news.
We also propose a novel cross-modality contrastive alignment while reserving the unimodal heads beside the fusion head to mitigate this problem.
arXiv Detail & Related papers (2023-09-11T16:47:01Z) - MMG-Ego4D: Multi-Modal Generalization in Egocentric Action Recognition [73.80088682784587]
"Multimodal Generalization" (MMG) aims to study how systems can generalize when data from certain modalities is limited or even completely missing.
MMG consists of two novel scenarios, designed to support security, and efficiency considerations in real-world applications.
New fusion module with modality dropout training, contrastive-based alignment training, and a novel cross-modal loss for better few-shot performance.
arXiv Detail & Related papers (2023-05-12T03:05:40Z) - VERITE: A Robust Benchmark for Multimodal Misinformation Detection
Accounting for Unimodal Bias [17.107961913114778]
multimodal misinformation is a growing problem on social media platforms.
In this study, we investigate and identify the presence of unimodal bias in widely-used MMD benchmarks.
We introduce a new method -- termed Crossmodal HArd Synthetic MisAlignment (CHASMA) -- for generating realistic synthetic training data.
arXiv Detail & Related papers (2023-04-27T12:28:29Z) - Non-Invasive Fairness in Learning through the Lens of Data Drift [88.37640805363317]
We show how to improve the fairness of Machine Learning models without altering the data or the learning algorithm.
We use a simple but key insight: the divergence of trends between different populations, and, consecutively, between a learned model and minority populations, is analogous to data drift.
We explore two strategies (model-splitting and reweighing) to resolve this drift, aiming to improve the overall conformance of models to the underlying data.
arXiv Detail & Related papers (2023-03-30T17:30:42Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - A transformer-based model for default prediction in mid-cap corporate
markets [13.535770763481905]
We study mid-cap companies with less than US $10 billion in market capitalisation.
We look to predict the default probability term structure over the medium term.
We understand which data sources contribute most to the default risk.
arXiv Detail & Related papers (2021-11-18T19:01:00Z) - Test-time Collective Prediction [73.74982509510961]
Multiple parties in machine learning want to jointly make predictions on future test points.
Agents wish to benefit from the collective expertise of the full set of agents, but may not be willing to release their data or model parameters.
We explore a decentralized mechanism to make collective predictions at test time, leveraging each agent's pre-trained model.
arXiv Detail & Related papers (2021-06-22T18:29:58Z) - Families In Wild Multimedia: A Multimodal Database for Recognizing
Kinship [63.27052967981546]
We introduce the first publicly available multi-task MM kinship dataset.
To build FIW MM, we developed machinery to automatically collect, annotate, and prepare the data.
Results highlight edge cases to inspire future research with different areas of improvement.
arXiv Detail & Related papers (2020-07-28T22:36:57Z) - Industry Adoption Scenarios for Authoritative Data Stores using the ISDA
Common Domain Model [0.0]
We explore opportunities for the post-trade industry to standardize and simplify in order to significantly increase efficiency and reduce costs.
This includes transitioning to the International Swaps and Derivatives Association Common Domain Model (CDM) as a standard set of digital representations for the business events and processes throughout the life cycle of a trade.
arXiv Detail & Related papers (2020-07-13T16:59:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.