Crude Oil-related Events Extraction and Processing: A Transfer Learning
Approach
- URL: http://arxiv.org/abs/2205.00387v1
- Date: Sun, 1 May 2022 03:21:18 GMT
- Title: Crude Oil-related Events Extraction and Processing: A Transfer Learning
Approach
- Authors: Meisin Lee, Lay-Ki Soon, Eu-Gene Siew
- Abstract summary: This paper presents a complete framework for extracting and processing crude oil-related events found in CrudeOilNews corpus.
We place special emphasis on event properties (Polarity, Modality, and Intensity) classification to determine the factual certainty of each event.
- Score: 0.7476901945542385
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of the challenges in event extraction via traditional supervised learning
paradigm is the need for a sizeable annotated dataset to achieve satisfactory
model performance. It is even more challenging when it comes to event
extraction in the finance and economics domain, a domain with considerably
fewer resources. This paper presents a complete framework for extracting and
processing crude oil-related events found in CrudeOilNews corpus, addressing
the issue of annotation scarcity and class imbalance by leveraging on the
effectiveness of transfer learning. Apart from event extraction, we place
special emphasis on event properties (Polarity, Modality, and Intensity)
classification to determine the factual certainty of each event. We build
baseline models first by supervised learning and then exploit Transfer Learning
methods to boost event extraction model performance despite the limited amount
of annotated data and severe class imbalance. This is done via methods within
the transfer learning framework such as Domain Adaptive Pre-training,
Multi-task Learning and Sequential Transfer Learning. Based on experiment
results, we are able to improve all event extraction sub-task models both in F1
and MCC1-score as compared to baseline models trained via the standard
supervised learning. Accurate and holistic event extraction from crude oil news
is very useful for downstream tasks such as understanding event chains and
learning event-event relations, which can be used for other downstream tasks
such as commodity price prediction, summarisation, etc. to support a wide range
of business decision making.
Related papers
- Token-Event-Role Structure-based Multi-Channel Document-Level Event
Extraction [15.02043375212839]
This paper introduces a novel framework for document-level event extraction, incorporating a new data structure called token-event-role.
The proposed data structure enables our model to uncover the primary role of tokens in multiple events, facilitating a more comprehensive understanding of event relationships.
The results demonstrate that our approach outperforms the state-of-the-art method by 9.5 percentage points in terms of the F1 score.
arXiv Detail & Related papers (2023-06-30T15:22:57Z) - Complementary Learning Subnetworks for Parameter-Efficient
Class-Incremental Learning [40.13416912075668]
We propose a rehearsal-free CIL approach that learns continually via the synergy between two Complementary Learning Subnetworks.
Our method achieves competitive results against state-of-the-art methods, especially in accuracy gain, memory cost, training efficiency, and task-order.
arXiv Detail & Related papers (2023-06-21T01:43:25Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Rethinking Importance Weighting for Transfer Learning [71.81262398144946]
Key assumption in supervised learning is that training and test data follow the same probability distribution.
As real-world machine learning tasks are becoming increasingly complex, novel approaches are explored to cope with such challenges.
arXiv Detail & Related papers (2021-12-19T14:35:25Z) - Active Learning for Event Extraction with Memory-based Loss Prediction
Model [12.509218857483223]
Event extraction plays an important role in many industrial application scenarios.
We introduce active learning (AL) technology to reduce the cost of event annotation.
We propose a batch-based selection strategy and a Memory-Based Loss Prediction model (MBLP) to select unlabeled samples efficiently.
arXiv Detail & Related papers (2021-11-26T07:58:11Z) - Omni-Training for Data-Efficient Deep Learning [80.28715182095975]
Recent advances reveal that a properly pre-trained model endows an important property: transferability.
A tight combination of pre-training and meta-training cannot achieve both kinds of transferability.
This motivates the proposed Omni-Training framework towards data-efficient deep learning.
arXiv Detail & Related papers (2021-10-14T16:30:36Z) - Learning Constraints and Descriptive Segmentation for Subevent Detection [74.48201657623218]
We propose an approach to learning and enforcing constraints that capture dependencies between subevent detection and EventSeg prediction.
We adopt Rectifier Networks for constraint learning and then convert the learned constraints to a regularization term in the loss function of the neural model.
arXiv Detail & Related papers (2021-09-13T20:50:37Z) - Back to Prior Knowledge: Joint Event Causality Extraction via
Convolutional Semantic Infusion [5.566928318239452]
Joint event and causality extraction is a challenging yet essential task in information retrieval and data mining.
We propose convolutional knowledge infusion for frequent n-grams with different windows of length within a joint extraction framework.
Our model significantly outperforms the strong BERT+CSNN baseline.
arXiv Detail & Related papers (2021-02-19T13:31:46Z) - Active Learning for Sequence Tagging with Deep Pre-trained Models and
Bayesian Uncertainty Estimates [52.164757178369804]
Recent advances in transfer learning for natural language processing in conjunction with active learning open the possibility to significantly reduce the necessary annotation budget.
We conduct an empirical study of various Bayesian uncertainty estimation methods and Monte Carlo dropout options for deep pre-trained models in the active learning framework.
We also demonstrate that to acquire instances during active learning, a full-size Transformer can be substituted with a distilled version, which yields better computational performance.
arXiv Detail & Related papers (2021-01-20T13:59:25Z) - Minimax Lower Bounds for Transfer Learning with Linear and One-hidden
Layer Neural Networks [27.44348371795822]
We develop a statistical minimax framework to characterize the limits of transfer learning.
We derive a lower-bound for the target generalization error achievable by any algorithm as a function of the number of labeled source and target data.
arXiv Detail & Related papers (2020-06-16T22:49:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.