Discriminative Adversarial Search for Abstractive Summarization
- URL: http://arxiv.org/abs/2002.10375v2
- Date: Sun, 30 Aug 2020 07:21:53 GMT
- Title: Discriminative Adversarial Search for Abstractive Summarization
- Authors: Thomas Scialom, Paul-Alexis Dray, Sylvain Lamprier, Benjamin
Piwowarski, Jacopo Staiano
- Abstract summary: We introduce a novel approach for sequence decoding, Discriminative Adversarial Search (DAS)
DAS has the desirable properties of alleviating the effects of exposure bias without requiring external metrics.
We investigate the effectiveness of the proposed approach on the task of Abstractive Summarization.
- Score: 29.943949944682196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce a novel approach for sequence decoding, Discriminative
Adversarial Search (DAS), which has the desirable properties of alleviating the
effects of exposure bias without requiring external metrics. Inspired by
Generative Adversarial Networks (GANs), wherein a discriminator is used to
improve the generator, our method differs from GANs in that the generator
parameters are not updated at training time and the discriminator is only used
to drive sequence generation at inference time.
We investigate the effectiveness of the proposed approach on the task of
Abstractive Summarization: the results obtained show that a naive application
of DAS improves over the state-of-the-art methods, with further gains obtained
via discriminator retraining. Moreover, we show how DAS can be effective for
cross-domain adaptation. Finally, all results reported are obtained without
additional rule-based filtering strategies, commonly used by the best
performing systems available: this indicates that DAS can effectively be
deployed without relying on post-hoc modifications of the generated outputs.
Related papers
- GADT: Enhancing Transferable Adversarial Attacks through Gradient-guided Adversarial Data Transformation [17.391427373945803]
We propose a novel DA-based attack algorithm, GADT, which identifies suitable DA parameters through iterative antagonism and uses posterior estimates to update AN based on these parameters.
GADT can be utilized in other black-box attack scenarios, e.g., query-based attacks, offering a new avenue to enhance attacks on real-world AI applications.
arXiv Detail & Related papers (2024-10-24T11:21:49Z) - DiffPoGAN: Diffusion Policies with Generative Adversarial Networks for Offline Reinforcement Learning [22.323173093804897]
offline reinforcement learning can learn optimal policies from pre-collected offline datasets without interacting with the environment.
Recent works address this issue by employing generative adversarial networks (GANs)
Inspired by the diffusion, we propose a new offline RL method named Diffusion Policies with Generative Adversarial Networks (DiffPoGAN)
arXiv Detail & Related papers (2024-06-13T13:15:40Z) - GE-AdvGAN: Improving the transferability of adversarial samples by
gradient editing-based adversarial generative model [69.71629949747884]
Adversarial generative models, such as Generative Adversarial Networks (GANs), are widely applied for generating various types of data.
In this work, we propose a novel algorithm named GE-AdvGAN to enhance the transferability of adversarial samples.
arXiv Detail & Related papers (2024-01-11T16:43:16Z) - Anti-Exploration by Random Network Distillation [63.04360288089277]
We show that a naive choice of conditioning for the Random Network Distillation (RND) is not discriminative enough to be used as an uncertainty estimator.
We show that this limitation can be avoided with conditioning based on Feature-wise Linear Modulation (FiLM)
We evaluate it on the D4RL benchmark, showing that it is capable of achieving performance comparable to ensemble-based methods and outperforming ensemble-free approaches by a wide margin.
arXiv Detail & Related papers (2023-01-31T13:18:33Z) - Mitigating Algorithmic Bias with Limited Annotations [65.060639928772]
When sensitive attributes are not disclosed or available, it is needed to manually annotate a small part of the training data to mitigate bias.
We propose Active Penalization Of Discrimination (APOD), an interactive framework to guide the limited annotations towards maximally eliminating the effect of algorithmic bias.
APOD shows comparable performance to fully annotated bias mitigation, which demonstrates that APOD could benefit real-world applications when sensitive information is limited.
arXiv Detail & Related papers (2022-07-20T16:31:19Z) - Augmentation-Aware Self-Supervision for Data-Efficient GAN Training [68.81471633374393]
Training generative adversarial networks (GANs) with limited data is challenging because the discriminator is prone to overfitting.
We propose a novel augmentation-aware self-supervised discriminator that predicts the augmentation parameter of the augmented data.
We compare our method with state-of-the-art (SOTA) methods using the class-conditional BigGAN and unconditional StyleGAN2 architectures.
arXiv Detail & Related papers (2022-05-31T10:35:55Z) - Reusing the Task-specific Classifier as a Discriminator:
Discriminator-free Adversarial Domain Adaptation [55.27563366506407]
We introduce a discriminator-free adversarial learning network (DALN) for unsupervised domain adaptation (UDA)
DALN achieves explicit domain alignment and category distinguishment through a unified objective.
DALN compares favorably against the existing state-of-the-art (SOTA) methods on a variety of public datasets.
arXiv Detail & Related papers (2022-04-08T04:40:18Z) - Sampling-Decomposable Generative Adversarial Recommender [84.05894139540048]
We propose a Sampling-Decomposable Generative Adversarial Recommender (SD-GAR)
In the framework, the divergence between some generator and the optimum is compensated by self-normalized importance sampling.
We extensively evaluate the proposed algorithm with five real-world recommendation datasets.
arXiv Detail & Related papers (2020-11-02T13:19:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.