Pseudo Siamese Network for Few-shot Intent Generation
- URL: http://arxiv.org/abs/2105.00896v1
- Date: Mon, 3 May 2021 14:30:47 GMT
- Title: Pseudo Siamese Network for Few-shot Intent Generation
- Authors: Congying Xia, Caiming Xiong, Philip Yu
- Abstract summary: We propose a Pseudo Siamese Network (PSN) to generate labeled data for few-shot intents.
PSN consists of two identicalworks with the same structure but different weights: an action network and an object network.
Experiments on two real-world datasets show that PSN achieves state-of-the-art performance for the generalized few shot intent detection task.
- Score: 54.10596778418007
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot intent detection is a challenging task due to the scare annotation
problem. In this paper, we propose a Pseudo Siamese Network (PSN) to generate
labeled data for few-shot intents and alleviate this problem. PSN consists of
two identical subnetworks with the same structure but different weights: an
action network and an object network. Each subnetwork is a transformer-based
variational autoencoder that tries to model the latent distribution of
different components in the sentence. The action network is learned to
understand action tokens and the object network focuses on object-related
expressions. It provides an interpretable framework for generating an utterance
with an action and an object existing in a given intent. Experiments on two
real-world datasets show that PSN achieves state-of-the-art performance for the
generalized few shot intent detection task.
Related papers
- Hierarchical Graph Interaction Transformer with Dynamic Token Clustering for Camouflaged Object Detection [57.883265488038134]
We propose a hierarchical graph interaction network termed HGINet for camouflaged object detection.
The network is capable of discovering imperceptible objects via effective graph interaction among the hierarchical tokenized features.
Our experiments demonstrate the superior performance of HGINet compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2024-08-27T12:53:25Z) - Neural Descriptor Fields: SE(3)-Equivariant Object Representations for
Manipulation [75.83319382105894]
We present Neural Descriptor Fields (NDFs), an object representation that encodes both points and relative poses between an object and a target.
NDFs are trained in a self-supervised fashion via a 3D auto-encoding task that does not rely on expert-labeled keypoints.
Our performance generalizes across both object instances and 6-DoF object poses, and significantly outperforms a recent baseline that relies on 2D descriptors.
arXiv Detail & Related papers (2021-12-09T18:57:15Z) - Generating Band-Limited Adversarial Surfaces Using Neural Networks [0.9208007322096533]
adversarial examples is the art of creating a noise that is added to an input signal of a classifying neural network.
In this technical report we suggest a neural network that generates the attacks.
arXiv Detail & Related papers (2021-11-14T19:16:05Z) - Anabranch Network for Camouflaged Object Segmentation [23.956327305907585]
This paper explores the camouflaged object segmentation problem, namely, segmenting the camouflaged object(s) for a given image.
To address this problem, we provide a new image dataset of camouflaged objects for benchmarking purposes.
In addition, we propose a general end-to-end network, called the Anabranch Network, that leverages both classification and segmentation tasks.
arXiv Detail & Related papers (2021-05-20T01:52:44Z) - Joint Intent Detection and Slot Filling with Wheel-Graph Attention
Networks [6.939768185086755]
We propose a new joint model with a wheel-graph attention network (Wheel-GAT) which is able to model interrelated connections directly for intent detection and slot filling.
To construct a graph structure for utterances, we create intent nodes, slot nodes, and directed edges.
arXiv Detail & Related papers (2021-02-09T02:37:56Z) - Where2Act: From Pixels to Actions for Articulated 3D Objects [54.19638599501286]
We extract highly localized actionable information related to elementary actions such as pushing or pulling for articulated objects with movable parts.
We propose a learning-from-interaction framework with an online data sampling strategy that allows us to train the network in simulation.
Our learned models even transfer to real-world data.
arXiv Detail & Related papers (2021-01-07T18:56:38Z) - Understanding the Role of Individual Units in a Deep Neural Network [85.23117441162772]
We present an analytic framework to systematically identify hidden units within image classification and image generation networks.
First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts.
Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes.
arXiv Detail & Related papers (2020-09-10T17:59:10Z) - Entropy Guided Adversarial Model for Weakly Supervised Object
Localization [11.77745060973134]
We propose to apply the shannon entropy on the CAMs generated by the network to guide it during training.
Our method does not erase any part of the image neither does it change the network architecure.
Our Entropy Guided Adversarial model (EGA model) improved performance on state of the arts benchmarks for both localization and classification accuracy.
arXiv Detail & Related papers (2020-08-04T19:39:12Z) - AGIF: An Adaptive Graph-Interactive Framework for Joint Multiple Intent
Detection and Slot Filling [69.59096090788125]
In this paper, we propose an Adaptive Graph-Interactive Framework (AGIF) for joint multiple intent detection and slot filling.
We introduce an intent-slot graph interaction layer to model the strong correlation between the slot and intents.
Such an interaction layer is applied to each token adaptively, which has the advantage to automatically extract the relevant intents information.
arXiv Detail & Related papers (2020-04-21T15:07:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.