Transformer-based Multi-task Learning for Disaster Tweet Categorisation
- URL: http://arxiv.org/abs/2110.08010v1
- Date: Fri, 15 Oct 2021 11:13:46 GMT
- Title: Transformer-based Multi-task Learning for Disaster Tweet Categorisation
- Authors: Congcong Wang, Paul Nulty, David Lillis
- Abstract summary: Social media has enabled people to circulate information in a timely fashion, thus motivating people to post messages seeking help during crisis situations.
These messages can contribute to the situational awareness of emergency responders, who have a need for them to be categorised according to information types.
We introduce a transformer-based multi-task learning (MTL) technique for classifying information types and estimating the priority of these messages.
- Score: 2.9112649816695204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social media has enabled people to circulate information in a timely fashion,
thus motivating people to post messages seeking help during crisis situations.
These messages can contribute to the situational awareness of emergency
responders, who have a need for them to be categorised according to information
types (i.e. the type of aid services the messages are requesting). We introduce
a transformer-based multi-task learning (MTL) technique for classifying
information types and estimating the priority of these messages. We evaluate
the effectiveness of our approach with a variety of metrics by submitting runs
to the TREC Incident Streams (IS) track: a research initiative specifically
designed for disaster tweet classification and prioritisation. The results
demonstrate that our approach achieves competitive performance in most metrics
as compared to other participating runs. Subsequently, we find that an ensemble
approach combining disparate transformer encoders within our approach helps to
improve the overall effectiveness to a significant extent, achieving
state-of-the-art performance in almost every metric. We make the code publicly
available so that our work can be reproduced and used as a baseline for the
community for future work in this domain.
Related papers
- Communication Learning in Multi-Agent Systems from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
We introduce a temporal gating mechanism for each agent, enabling dynamic decisions on whether to receive shared information at a given time.
arXiv Detail & Related papers (2024-11-01T05:56:51Z) - CrisisSense-LLM: Instruction Fine-Tuned Large Language Model for Multi-label Social Media Text Classification in Disaster Informatics [49.2719253711215]
This study introduces a novel approach to disaster text classification by enhancing a pre-trained Large Language Model (LLM)
Our methodology involves creating a comprehensive instruction dataset from disaster-related tweets, which is then used to fine-tune an open-source LLM.
This fine-tuned model can classify multiple aspects of disaster-related information simultaneously, such as the type of event, informativeness, and involvement of human aid.
arXiv Detail & Related papers (2024-06-16T23:01:10Z) - Learning Multi-Agent Communication from Graph Modeling Perspective [62.13508281188895]
We introduce a novel approach wherein we conceptualize the communication architecture among agents as a learnable graph.
Our proposed approach, CommFormer, efficiently optimize the communication graph and concurrently refines architectural parameters through gradient descent in an end-to-end manner.
arXiv Detail & Related papers (2024-05-14T12:40:25Z) - Context-aware Communication for Multi-agent Reinforcement Learning [6.109127175562235]
We develop a context-aware communication scheme for multi-agent reinforcement learning (MARL)
In the first stage, agents exchange coarse representations in a broadcast fashion, providing context for the second stage.
Following this, agents utilize attention mechanisms in the second stage to selectively generate messages personalized for the receivers.
To evaluate the effectiveness of CACOM, we integrate it with both actor-critic and value-based MARL algorithms.
arXiv Detail & Related papers (2023-12-25T03:33:08Z) - MASTER: Multi-task Pre-trained Bottlenecked Masked Autoencoders are
Better Dense Retrievers [140.0479479231558]
In this work, we aim to unify a variety of pre-training tasks into a multi-task pre-trained model, namely MASTER.
MASTER utilizes a shared-encoder multi-decoder architecture that can construct a representation bottleneck to compress the abundant semantic information across tasks into dense vectors.
arXiv Detail & Related papers (2022-12-15T13:57:07Z) - Enhancing Crisis-Related Tweet Classification with Entity-Masked
Language Modeling and Multi-Task Learning [0.30458514384586394]
We propose a combination of entity-masked language modeling and hierarchical multi-label classification as a multi-task learning problem.
We evaluate our method on tweets from the TREC-IS dataset and show an absolute performance gain w.r.t. F1-score of up to 10% for actionable information types.
arXiv Detail & Related papers (2022-11-21T13:54:10Z) - UCD-CS at TREC 2021 Incident Streams Track [3.4392739159262145]
The TREC Incident Streams (IS) track is a research challenge organised for this purpose.
The track asks participating systems to both classify a stream of crisis-related tweets into humanitarian aid related information types.
We report on the participation of the University College Dublin School of Computer Science (UCD-CS) in TREC-IS 2021.
arXiv Detail & Related papers (2021-12-07T14:47:27Z) - Multi-task transfer learning for finding actionable information from
crisis-related messages on social media [3.4392739159262145]
The Incident streams (IS) track is a research challenge aimed at finding important information from social media during crises for emergency response purposes.
Given a stream of crisis-related tweets, the IS challenge asks a participating system to classify what the types of users' concerns or needs are expressed in each tweet.
We describe our multi-task transfer learning approach for this challenge.
arXiv Detail & Related papers (2021-02-26T11:11:33Z) - Dif-MAML: Decentralized Multi-Agent Meta-Learning [54.39661018886268]
We propose a cooperative multi-agent meta-learning algorithm, referred to as MAML or Dif-MAML.
We show that the proposed strategy allows a collection of agents to attain agreement at a linear rate and to converge to a stationary point of the aggregate MAML.
Simulation results illustrate the theoretical findings and the superior performance relative to the traditional non-cooperative setting.
arXiv Detail & Related papers (2020-10-06T16:51:09Z) - Multimodal Categorization of Crisis Events in Social Media [81.07061295887172]
We present a new multimodal fusion method that leverages both images and texts as input.
In particular, we introduce a cross-attention module that can filter uninformative and misleading components from weak modalities.
We show that our method outperforms the unimodal approaches and strong multimodal baselines by a large margin on three crisis-related tasks.
arXiv Detail & Related papers (2020-04-10T06:31:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.