Multi-Domain Transformer-Based Counterfactual Augmentation for Earnings
Call Analysis
- URL: http://arxiv.org/abs/2112.00963v2
- Date: Fri, 3 Dec 2021 05:12:05 GMT
- Title: Multi-Domain Transformer-Based Counterfactual Augmentation for Earnings
Call Analysis
- Authors: Zixuan Yuan, Yada Zhu, Wei Zhang, Ziming Huang, Guangnan Ye, Hui Xiong
- Abstract summary: We propose a transformer-based EC encoder to attentively quantify the task-inspired significance of critical EC content for market inference.
We then develop a multi-domain counterfactual learning framework to evaluate the gradient-based variations.
Experiments on the real-world financial datasets demonstrate the effectiveness of interpretable MTCA for improving the volatility evaluation ability of the state-of-the-art by 14.2% in accuracy.
- Score: 20.087027853160627
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Earnings call (EC), as a periodic teleconference of a publicly-traded
company, has been extensively studied as an essential market indicator because
of its high analytical value in corporate fundamentals. The recent emergence of
deep learning techniques has shown great promise in creating automated
pipelines to benefit the EC-supported financial applications. However, these
methods presume all included contents to be informative without refining
valuable semantics from long-text transcript and suffer from EC scarcity issue.
Meanwhile, these black-box methods possess inherent difficulties in providing
human-understandable explanations. To this end, in this paper, we propose a
Multi-Domain Transformer-Based Counterfactual Augmentation, named MTCA, to
address the above problems. Specifically, we first propose a transformer-based
EC encoder to attentively quantify the task-inspired significance of critical
EC content for market inference. Then, a multi-domain counterfactual learning
framework is developed to evaluate the gradient-based variations after we
perturb limited EC informative texts with plentiful cross-domain documents,
enabling MTCA to perform unsupervised data augmentation. As a bonus, we
discover a way to use non-training data as instance-based explanations for
which we show the result with case studies. Extensive experiments on the
real-world financial datasets demonstrate the effectiveness of interpretable
MTCA for improving the volatility evaluation ability of the state-of-the-art by
14.2\% in accuracy.
Related papers
- Towards Faster Decentralized Stochastic Optimization with Communication Compression [27.484212303346816]
We introduce MoTEF, a novel approach that integrates communication with Momentum Tracking and Error Feedback.
Our analysis demonstrates that MoTEF integrates most of the desired properties, and significantly existing methods under data.
arXiv Detail & Related papers (2024-05-30T14:51:57Z) - Introducing User Feedback-based Counterfactual Explanations (UFCE) [49.1574468325115]
Counterfactual explanations (CEs) have emerged as a viable solution for generating comprehensible explanations in XAI.
UFCE allows for the inclusion of user constraints to determine the smallest modifications in the subset of actionable features.
UFCE outperforms two well-known CE methods in terms of textitproximity, textitsparsity, and textitfeasibility.
arXiv Detail & Related papers (2024-02-26T20:09:44Z) - Towards reducing hallucination in extracting information from financial
reports using Large Language Models [1.2289361708127877]
We show how Large Language Models (LLMs) can efficiently and rapidly extract information from earnings report transcripts.
We evaluate the outcomes of various LLMs with and without using our proposed approach based on various objective metrics for evaluating Q&A systems.
arXiv Detail & Related papers (2023-10-16T18:45:38Z) - HAlf-MAsked Model for Named Entity Sentiment analysis [0.0]
We study different transformers-based solutions NESA in RuSentNE-23 evaluation.
We present several approaches to overcome this problem, among which there is a novel technique of additional pass over given data with masked entity.
Our proposed model achieves the best result on RuSentNE-23 evaluation data and demonstrates improved consistency in entity-level sentiment analysis.
arXiv Detail & Related papers (2023-08-30T06:53:24Z) - Distributionally Robust Cross Subject EEG Decoding [15.211091130230589]
We propose a principled approach to perform dynamic evolution on the data for improvement of decoding robustness.
We derived a general data evolution framework based on Wasserstein gradient flow (WGF) and provides two different forms of evolution within the framework.
The proposed approach can be readily integrated with other data augmentation approaches for further improvements.
arXiv Detail & Related papers (2023-08-19T11:31:33Z) - DAPDAG: Domain Adaptation via Perturbed DAG Reconstruction [78.76115370275733]
We learn an auto-encoder that undertakes inference on population statistics given features and reconstruct a directed acyclic graph (DAG) as an auxiliary task.
The underlying DAG structure is assumed invariant among observed variables whose conditional distributions are allowed to vary across domains led by a latent environmental variable $E$.
We train the encoder and decoder jointly in an end-to-end manner and conduct experiments on synthetic and real datasets with mixed variables.
arXiv Detail & Related papers (2022-08-02T11:43:03Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - Towards Fair Knowledge Transfer for Imbalanced Domain Adaptation [61.317911756566126]
We propose a Towards Fair Knowledge Transfer framework to handle the fairness challenge in imbalanced cross-domain learning.
Specifically, a novel cross-domain mixup generation is exploited to augment the minority source set with target information to enhance fairness.
Our model significantly improves over 20% on two benchmarks in terms of the overall accuracy.
arXiv Detail & Related papers (2020-10-23T06:29:09Z) - Transformer Based Multi-Source Domain Adaptation [53.24606510691877]
In practical machine learning settings, the data on which a model must make predictions often come from a different distribution than the data it was trained on.
Here, we investigate the problem of unsupervised multi-source domain adaptation, where a model is trained on labelled data from multiple source domains and must make predictions on a domain for which no labelled data has been seen.
We show that the predictions of large pretrained transformer based domain experts are highly homogenous, making it challenging to learn effective functions for mixing their predictions.
arXiv Detail & Related papers (2020-09-16T16:56:23Z) - A Transductive Multi-Head Model for Cross-Domain Few-Shot Learning [72.30054522048553]
We present a new method, Transductive Multi-Head Few-Shot learning (TMHFS), to address the Cross-Domain Few-Shot Learning challenge.
The proposed methods greatly outperform the strong baseline, fine-tuning, on four different target domains.
arXiv Detail & Related papers (2020-06-08T02:39:59Z) - A BERT based Sentiment Analysis and Key Entity Detection Approach for
Online Financial Texts [4.834766555659253]
We propose a sentiment analysis and key entity detection approach based on BERT, which is applied in online financial text mining and public opinion analysis in social media.
Experimental results show that the performance of our approach is generally higher than SVM, LR, NBM, and BERT for two financial sentiment analysis and key entity detection datasets.
arXiv Detail & Related papers (2020-01-14T13:50:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.