EaSyGuide : ESG Issue Identification Framework leveraging Abilities of
Generative Large Language Models
- URL: http://arxiv.org/abs/2306.06662v2
- Date: Tue, 13 Jun 2023 13:47:15 GMT
- Title: EaSyGuide : ESG Issue Identification Framework leveraging Abilities of
Generative Large Language Models
- Authors: Hanwool Lee, Jonghyun Choi, Sohyeon Kwon, Sungbum Jung
- Abstract summary: This paper presents our participation in the FinNLP-2023 shared task on multi-lingual environmental, social, and corporate governance issue identification (ML-ESG)
The task's objective is to classify news articles based on the 35 ESG key issues defined by the MSCI ESG rating guidelines.
Our approach focuses on the English and French subtasks, employing the CerebrasGPT, OPT, and Pythia models, along with the zero-shot and GPT3Mix Augmentation techniques.
- Score: 5.388543737855513
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper presents our participation in the FinNLP-2023 shared task on
multi-lingual environmental, social, and corporate governance issue
identification (ML-ESG). The task's objective is to classify news articles
based on the 35 ESG key issues defined by the MSCI ESG rating guidelines. Our
approach focuses on the English and French subtasks, employing the CerebrasGPT,
OPT, and Pythia models, along with the zero-shot and GPT3Mix Augmentation
techniques. We utilize various encoder models, such as RoBERTa, DeBERTa, and
FinBERT, subjecting them to knowledge distillation and additional training.
Our approach yielded exceptional results, securing the first position in the
English text subtask with F1-score 0.69 and the second position in the French
text subtask with F1-score 0.78. These outcomes underscore the effectiveness of
our methodology in identifying ESG issues in news articles across different
languages. Our findings contribute to the exploration of ESG topics and
highlight the potential of leveraging advanced language models for ESG issue
identification.
Related papers
- Evaluating the performance of state-of-the-art esg domain-specific pre-trained large language models in text classification against existing models and traditional machine learning techniques [0.0]
This research investigates the classification of Environmental, Social, and Governance (ESG) information within textual disclosures.
The aim is to develop and evaluate binary classification models capable of accurately identifying and categorizing E, S and G-related content respectively.
The motivation for this research stems from the growing importance of ESG considerations in investment decisions and corporate accountability.
arXiv Detail & Related papers (2024-09-30T20:08:32Z) - ESG-FTSE: A corpus of news articles with ESG relevance labels and use cases [1.3937696730884712]
We present ESG-FTSE, the first corpus comprised of news articles with Environmental, Social and Governance (ESG) relevance annotations.
This has led to the rise of ESG scores to evaluate an investment's credentials as socially responsible.
Quantitative techniques can be applied to improve ESG scores, thus, responsible investing.
arXiv Detail & Related papers (2024-05-30T16:19:02Z) - Jetsons at FinNLP 2024: Towards Understanding the ESG Impact of a News Article using Transformer-based Models [7.010041097710465]
We describe the different approaches explored by the Jetsons team for the Multi-Lingual ESG Impact Duration Inference task.
The shared task dataset consists of 2,059 news titles and articles in English, French, Korean, and Japanese languages.
For the impact type classification task, our XLM-RoBERTa model fine-tuned using a custom fine-tuning strategy ranked first for the English language.
arXiv Detail & Related papers (2024-03-30T14:58:44Z) - SOUL: Towards Sentiment and Opinion Understanding of Language [96.74878032417054]
We propose a new task called Sentiment and Opinion Understanding of Language (SOUL)
SOUL aims to evaluate sentiment understanding through two subtasks: Review (RC) and Justification Generation (JG)
arXiv Detail & Related papers (2023-10-27T06:48:48Z) - Leveraging BERT Language Models for Multi-Lingual ESG Issue
Identification [0.30254881201174333]
Investors have increasingly recognized the significance of ESG criteria in their investment choices.
The Multi-Lingual ESG Issue Identification (ML-ESG) task encompasses the classification of news documents into 35 distinct ESG issue labels.
In this study, we explored multiple strategies harnessing BERT language models to achieve accurate classification of news documents across these labels.
arXiv Detail & Related papers (2023-09-05T12:48:21Z) - ChatGPT Beyond English: Towards a Comprehensive Evaluation of Large
Language Models in Multilingual Learning [70.57126720079971]
Large language models (LLMs) have emerged as the most important breakthroughs in natural language processing (NLP)
This paper evaluates ChatGPT on 7 different tasks, covering 37 diverse languages with high, medium, low, and extremely low resources.
Compared to the performance of previous models, our extensive experimental results demonstrate a worse performance of ChatGPT for different NLP tasks and languages.
arXiv Detail & Related papers (2023-04-12T05:08:52Z) - Bag of Tricks for Effective Language Model Pretraining and Downstream
Adaptation: A Case Study on GLUE [93.98660272309974]
This report briefly describes our submission Vega v1 on the General Language Understanding Evaluation leaderboard.
GLUE is a collection of nine natural language understanding tasks, including question answering, linguistic acceptability, sentiment analysis, text similarity, paraphrase detection, and natural language inference.
With our optimized pretraining and fine-tuning strategies, our 1.3 billion model sets new state-of-the-art on 4/9 tasks, achieving the best average score of 91.3.
arXiv Detail & Related papers (2023-02-18T09:26:35Z) - Toward Efficient Language Model Pretraining and Downstream Adaptation
via Self-Evolution: A Case Study on SuperGLUE [203.65227947509933]
This report describes our JDExplore d-team's Vega v2 submission on the SuperGLUE leaderboard.
SuperGLUE is more challenging than the widely used general language understanding evaluation (GLUE) benchmark, containing eight difficult language understanding tasks.
arXiv Detail & Related papers (2022-12-04T15:36:18Z) - Using contextual sentence analysis models to recognize ESG concepts [8.905370601886112]
This paper summarizes the joint participation of the Trading Central Labs and the L3i laboratory of the University of La Rochelle on two sub-tasks of the FinSim-4 evaluation campaign.
The first sub-task aims to enrich the 'Fortia ESG taxonomy' with new lexicon entries while the second one aims to classify sentences to either'sustainable' or 'unsustainable' with respect to ESG related factors.
arXiv Detail & Related papers (2022-07-04T13:33:21Z) - Language Models are Few-Shot Learners [61.36677350504291]
We show that scaling up language models greatly improves task-agnostic, few-shot performance.
We train GPT-3, an autoregressive language model with 175 billion parameters, and test its performance in the few-shot setting.
GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks.
arXiv Detail & Related papers (2020-05-28T17:29:03Z) - Kungfupanda at SemEval-2020 Task 12: BERT-Based Multi-Task Learning for
Offensive Language Detection [55.445023584632175]
We build an offensive language detection system, which combines multi-task learning with BERT-based models.
Our model achieves 91.51% F1 score in English Sub-task A, which is comparable to the first place.
arXiv Detail & Related papers (2020-04-28T11:27:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.