CHATREPORT: Democratizing Sustainability Disclosure Analysis through
LLM-based Tools
- URL: http://arxiv.org/abs/2307.15770v2
- Date: Wed, 11 Oct 2023 16:49:29 GMT
- Title: CHATREPORT: Democratizing Sustainability Disclosure Analysis through
LLM-based Tools
- Authors: Jingwei Ni, Julia Bingler, Chiara Colesanti-Senni, Mathias Kraus, Glen
Gostlow, Tobias Schimanski, Dominik Stammbach, Saeid Ashraf Vaghefi, Qian
Wang, Nicolas Webersinke, Tobias Wekhof, Tingyu Yu, Markus Leippold
- Abstract summary: ChatReport is a novel LLM-based system to automate the analysis of corporate sustainability reports.
We make our methodology, annotated datasets, and generated analyses of 1015 reports publicly available.
- Score: 10.653984116770234
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In the face of climate change, are companies really taking substantial steps
toward more sustainable operations? A comprehensive answer lies in the dense,
information-rich landscape of corporate sustainability reports. However, the
sheer volume and complexity of these reports make human analysis very costly.
Therefore, only a few entities worldwide have the resources to analyze these
reports at scale, which leads to a lack of transparency in sustainability
reporting. Empowering stakeholders with LLM-based automatic analysis tools can
be a promising way to democratize sustainability report analysis. However,
developing such tools is challenging due to (1) the hallucination of LLMs and
(2) the inefficiency of bringing domain experts into the AI development loop.
In this paper, we ChatReport, a novel LLM-based system to automate the analysis
of corporate sustainability reports, addressing existing challenges by (1)
making the answers traceable to reduce the harm of hallucination and (2)
actively involving domain experts in the development loop. We make our
methodology, annotated datasets, and generated analyses of 1015 reports
publicly available.
Related papers
- Explainable Natural Language Processing for Corporate Sustainability Analysis [26.267508407180465]
The concept of corporate sustainability is complex due to the diverse and intricate nature of firm operations.
Corporate sustainability assessments are plagued by subjectivity both within data that reflect corporate sustainability efforts and the analysts evaluating them.
We argue that Explainable Natural Language Processing (XNLP) can significantly enhance corporate sustainability analysis.
arXiv Detail & Related papers (2024-07-03T08:27:51Z) - Harnessing Business and Media Insights with Large Language Models [11.085865480418926]
This paper introduces Fortune Analytics Language Model (FALM)
FALM empowers users with direct access to comprehensive business analysis.
Unlike generic LLMs, FALM leverages a curated knowledge base built from professional journalism.
arXiv Detail & Related papers (2024-06-02T06:24:38Z) - Cooperate or Collapse: Emergence of Sustainable Cooperation in a Society of LLM Agents [101.17919953243107]
GovSim is a generative simulation platform designed to study strategic interactions and cooperative decision-making in large language models (LLMs)
We find that all but the most powerful LLM agents fail to achieve a sustainable equilibrium in GovSim, with the highest survival rate below 54%.
We show that agents that leverage "Universalization"-based reasoning, a theory of moral thinking, are able to achieve significantly better sustainability.
arXiv Detail & Related papers (2024-04-25T15:59:16Z) - Bangladesh Agricultural Knowledge Graph: Enabling Semantic Integration and Data-driven Analysis--Full Version [0.0]
We develop a federated knowledge graph that integrates agriculture data in Bangladesh.
BDAKG incorporates multidimensional semantics, is linked with external knowledge graphs, is compatible with OLAP, and adheres to the FAIR principles.
arXiv Detail & Related papers (2024-03-18T16:19:30Z) - AgentBoard: An Analytical Evaluation Board of Multi-turn LLM Agents [76.95062553043607]
evaluating large language models (LLMs) is essential for understanding their capabilities and facilitating their integration into practical applications.
We introduce AgentBoard, a pioneering comprehensive benchmark and accompanied open-source evaluation framework tailored to analytical evaluation of LLM agents.
arXiv Detail & Related papers (2024-01-24T01:51:00Z) - MuSR: Testing the Limits of Chain-of-thought with Multistep Soft Reasoning [63.80739044622555]
We introduce MuSR, a dataset for evaluating language models on soft reasoning tasks specified in a natural language narrative.
This dataset has two crucial features. First, it is created through a novel neurosymbolic synthetic-to-natural generation algorithm.
Second, our dataset instances are free text narratives corresponding to real-world domains of reasoning.
arXiv Detail & Related papers (2023-10-24T17:59:20Z) - Improving Open Information Extraction with Large Language Models: A
Study on Demonstration Uncertainty [52.72790059506241]
Open Information Extraction (OIE) task aims at extracting structured facts from unstructured text.
Despite the potential of large language models (LLMs) like ChatGPT as a general task solver, they lag behind state-of-the-art (supervised) methods in OIE tasks.
arXiv Detail & Related papers (2023-09-07T01:35:24Z) - Paradigm Shift in Sustainability Disclosure Analysis: Empowering
Stakeholders with CHATREPORT, a Language Model-Based Tool [10.653984116770234]
This paper introduces a novel approach to enhance Large Language Models (LLMs) with expert knowledge to automate the analysis of corporate sustainability reports.
We christen our tool CHATREPORT, and apply it in a first use case to assess corporate climate risk disclosures.
arXiv Detail & Related papers (2023-06-27T14:46:47Z) - LAMM: Language-Assisted Multi-Modal Instruction-Tuning Dataset,
Framework, and Benchmark [81.42376626294812]
We present Language-Assisted Multi-Modal instruction tuning dataset, framework, and benchmark.
Our aim is to establish LAMM as a growing ecosystem for training and evaluating MLLMs.
We present a comprehensive dataset and benchmark, which cover a wide range of vision tasks for 2D and 3D vision.
arXiv Detail & Related papers (2023-06-11T14:01:17Z) - Sentiment Analysis in the Era of Large Language Models: A Reality Check [69.97942065617664]
This paper investigates the capabilities of large language models (LLMs) in performing various sentiment analysis tasks.
We evaluate performance across 13 tasks on 26 datasets and compare the results against small language models (SLMs) trained on domain-specific datasets.
arXiv Detail & Related papers (2023-05-24T10:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.