Domain-Guided Task Decomposition with Self-Training for Detecting
Personal Events in Social Media
- URL: http://arxiv.org/abs/2004.10201v1
- Date: Tue, 21 Apr 2020 14:50:31 GMT
- Title: Domain-Guided Task Decomposition with Self-Training for Detecting
Personal Events in Social Media
- Authors: Payam Karisani, Joyce C. Ho, and Eugene Agichtein
- Abstract summary: Mining social media for tasks such as detecting personal experiences or events, suffer from lexical sparsity, insufficient training data, and inventive lexicons.
To reduce the burden of creating extensive labeled data, we propose to perform these tasks in two steps: 1.
Decomposing the task into domain-specific sub-tasks by identifying key concepts, thus utilizing human domain understanding; 2. Combining the results of learners for each key concept using co-training to reduce the requirements for labeled training data.
- Score: 11.638298634523945
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mining social media content for tasks such as detecting personal experiences
or events, suffer from lexical sparsity, insufficient training data, and
inventive lexicons. To reduce the burden of creating extensive labeled data and
improve classification performance, we propose to perform these tasks in two
steps: 1. Decomposing the task into domain-specific sub-tasks by identifying
key concepts, thus utilizing human domain understanding; and 2. Combining the
results of learners for each key concept using co-training to reduce the
requirements for labeled training data. We empirically show the effectiveness
and generality of our approach, Co-Decomp, using three representative social
media mining tasks, namely Personal Health Mention detection, Crisis Report
detection, and Adverse Drug Reaction monitoring. The experiments show that our
model is able to outperform the state-of-the-art text classification
models--including those using the recently introduced BERT model--when small
amounts of training data are available.
Related papers
- Data-CUBE: Data Curriculum for Instruction-based Sentence Representation
Learning [85.66907881270785]
We propose a data curriculum method, namely Data-CUBE, that arranges the orders of all the multi-task data for training.
In the task level, we aim to find the optimal task order to minimize the total cross-task interference risk.
In the instance level, we measure the difficulty of all instances per task, then divide them into the easy-to-difficult mini-batches for training.
arXiv Detail & Related papers (2024-01-07T18:12:20Z) - Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Leveraging Open Data and Task Augmentation to Automated Behavioral
Coding of Psychotherapy Conversations in Low-Resource Scenarios [35.44178630251169]
In psychotherapy interactions, the quality of a session is assessed by codifying the communicative behaviors of participants during the conversation.
In this paper, we leverage a publicly available conversation-based dataset and transfer knowledge to the low-resource behavioral coding task.
We introduce a task augmentation method to produce a large number of "analogy tasks" - tasks similar to the target one - and demonstrate that the proposed framework predicts target behaviors more accurately than all the other baseline models.
arXiv Detail & Related papers (2022-10-25T18:15:25Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Boosting Supervised Learning Performance with Co-training [15.986635379046602]
We propose a new light-weight self-supervised learning framework that could boost supervised learning performance with minimum additional cost.
Our results show that both self-supervised tasks can improve the accuracy of the supervised task and, at the same time, demonstrates strong domain adaption capability.
arXiv Detail & Related papers (2021-11-18T17:01:17Z) - Reinforcement Learning with Prototypical Representations [114.35801511501639]
Proto-RL is a self-supervised framework that ties representation learning with exploration through prototypical representations.
These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations.
This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks.
arXiv Detail & Related papers (2021-02-22T18:56:34Z) - Self-supervised driven consistency training for annotation efficient
histopathology image analysis [13.005873872821066]
Training a neural network with a large labeled dataset is still a dominant paradigm in computational histopathology.
We propose a self-supervised pretext task that harnesses the underlying multi-resolution contextual cues in histology whole-slide images to learn a powerful supervisory signal for unsupervised representation learning.
We also propose a new teacher-student semi-supervised consistency paradigm that learns to effectively transfer the pretrained representations to downstream tasks based on prediction consistency with the task-specific un-labeled data.
arXiv Detail & Related papers (2021-02-07T19:46:21Z) - Parrot: Data-Driven Behavioral Priors for Reinforcement Learning [79.32403825036792]
We propose a method for pre-training behavioral priors that can capture complex input-output relationships observed in successful trials.
We show how this learned prior can be used for rapidly learning new tasks without impeding the RL agent's ability to try out novel behaviors.
arXiv Detail & Related papers (2020-11-19T18:47:40Z) - Unsupervised and Interpretable Domain Adaptation to Rapidly Filter
Tweets for Emergency Services [18.57009530004948]
We present a novel method to classify relevant tweets during an ongoing crisis using the publicly available dataset of TREC incident streams.
We use dedicated attention layers for each task to provide model interpretability; critical for real-word applications.
We show a practical implication of our work by providing a use-case for the COVID-19 pandemic.
arXiv Detail & Related papers (2020-03-04T06:40:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.