Semi-supervised New Event Type Induction and Description via Contrastive
Loss-Enforced Batch Attention
- URL: http://arxiv.org/abs/2202.05943v1
- Date: Sat, 12 Feb 2022 00:32:22 GMT
- Title: Semi-supervised New Event Type Induction and Description via Contrastive
Loss-Enforced Batch Attention
- Authors: Carl Edwards and Heng Ji
- Abstract summary: We present a novel approach to semi-supervised new event type induction using a masked contrastive loss.
We extend our approach to two new tasks: predicting the type name of the discovered clusters and linking them to FrameNet frames.
- Score: 56.46649994444616
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most event extraction methods have traditionally relied on an annotated set
of event types. However, creating event ontologies and annotating supervised
training data are expensive and time-consuming. Previous work has proposed
semi-supervised approaches which leverage seen (annotated) types to learn how
to automatically discover new event types. State-of-the-art methods, both
semi-supervised or fully unsupervised, use a form of reconstruction loss on
specific tokens in a context. In contrast, we present a novel approach to
semi-supervised new event type induction using a masked contrastive loss, which
learns similarities between event mentions by enforcing an attention mechanism
over the data minibatch. We further disentangle the discovered clusters by
approximating the underlying manifolds in the data, which allows us to increase
normalized mutual information and Fowlkes-Mallows scores by over 20% absolute.
Building on these clustering results, we extend our approach to two new tasks:
predicting the type name of the discovered clusters and linking them to
FrameNet frames.
Related papers
- Large-scale Fully-Unsupervised Re-Identification [78.47108158030213]
We propose two strategies to learn from large-scale unlabeled data.
The first strategy performs a local neighborhood sampling to reduce the dataset size in each without violating neighborhood relationships.
A second strategy leverages a novel Re-Ranking technique, which has a lower time upper bound complexity and reduces the memory complexity from O(n2) to O(kn) with k n.
arXiv Detail & Related papers (2023-07-26T16:19:19Z) - A Clustering Framework for Unsupervised and Semi-supervised New Intent
Discovery [25.900661912504397]
We propose a novel clustering framework, USNID, for unsupervised and semi-supervised new intent discovery.
First, it fully utilizes unsupervised or semi-supervised data to mine shallow semantic similarity relations.
Second, it designs a centroid-guided clustering mechanism to address the issue of cluster allocation inconsistency.
Third, it captures high-level semantics in unsupervised or semi-supervised data to discover fine-grained intent-wise clusters.
arXiv Detail & Related papers (2023-04-16T05:30:42Z) - Leveraging triplet loss for unsupervised action segmentation [0.0]
We propose a fully unsupervised framework that learns action representations suitable for the action segmentation task from the single input video itself.
Our method is a deep metric learning approach rooted in a shallow network with a triplet loss operating on similarity distributions.
Under these circumstances, we successfully recover temporal boundaries in the learned action representations with higher quality compared with existing unsupervised approaches.
arXiv Detail & Related papers (2023-04-13T11:10:16Z) - Novel Class Discovery without Forgetting [72.52222295216062]
We identify and formulate a new, pragmatic problem setting of NCDwF: Novel Class Discovery without Forgetting.
We propose a machine learning model to incrementally discover novel categories of instances from unlabeled data.
We introduce experimental protocols based on CIFAR-10, CIFAR-100 and ImageNet-1000 to measure the trade-off between knowledge retention and novel class discovery.
arXiv Detail & Related papers (2022-07-21T17:54:36Z) - An Information-theoretical Approach to Semi-supervised Learning under
Covariate-shift [24.061858945664856]
A common assumption in semi-supervised learning is that the labeled, unlabeled, and test data are drawn from the same distribution.
We propose an approach for semi-supervised learning algorithms that is capable of addressing this issue.
Our framework also recovers some popular methods, including entropy minimization and pseudo-labeling.
arXiv Detail & Related papers (2022-02-24T14:25:14Z) - PILED: An Identify-and-Localize Framework for Few-Shot Event Detection [79.66042333016478]
In our study, we employ cloze prompts to elicit event-related knowledge from pretrained language models.
We minimize the number of type-specific parameters, enabling our model to quickly adapt to event detection tasks for new types.
arXiv Detail & Related papers (2022-02-15T18:01:39Z) - Unsupervised Event Detection, Clustering, and Use Case Exposition in
Micro-PMU Measurements [0.0]
We develop an unsupervised event detection method based on the concept of Generative Adversarial Networks (GAN)
We also propose a two-step unsupervised clustering method, based on a novel linear mixed integer programming formulation.
Results show that they can outperform the prevalent methods in the literature.
arXiv Detail & Related papers (2020-07-30T05:20:29Z) - Commonality-Parsing Network across Shape and Appearance for Partially
Supervised Instance Segmentation [71.59275788106622]
We propose to learn the underlying class-agnostic commonalities that can be generalized from mask-annotated categories to novel categories.
Our model significantly outperforms the state-of-the-art methods on both partially supervised setting and few-shot setting for instance segmentation on COCO dataset.
arXiv Detail & Related papers (2020-07-24T07:23:44Z) - Extensively Matching for Few-shot Learning Event Detection [66.31312496170139]
Event detection models under super-vised learning settings fail to transfer to new event types.
Few-shot learning has not beenexplored in event detection.
We propose two novelloss factors that matching examples in the sup-port set to provide more training signals to themodel.
arXiv Detail & Related papers (2020-06-17T18:30:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.