Interpretable and Effective Reinforcement Learning for Attacking against
Graph-based Rumor Detection
- URL: http://arxiv.org/abs/2201.05819v1
- Date: Sat, 15 Jan 2022 10:06:29 GMT
- Title: Interpretable and Effective Reinforcement Learning for Attacking against
Graph-based Rumor Detection
- Authors: Yuefei Lyu, Xiaoyu Yang, Jiaxin Liu, Sihong Xie, Xi Zhang
- Abstract summary: Social networks are polluted by rumors, which can be detected by machine learning models.
Certain vulnerabilities are due to dependencies on the graphs and suspiciousness ranking.
With a black-box detector, we design features capturing the dependencies to allow a reinforcement learning to learn an effective and interpretable attack policy.
- Score: 12.726403718158082
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social networks are polluted by rumors, which can be detected by machine
learning models. However, the models are fragile and understanding the
vulnerabilities is critical to rumor detection. Certain vulnerabilities are due
to dependencies on the graphs and suspiciousness ranking and are difficult for
end-to-end methods to learn from limited noisy data. With a black-box detector,
we design features capturing the dependencies to allow a reinforcement learning
to learn an effective and interpretable attack policy based on the detector
output. To speed up learning, we devise: (i) a credit assignment method that
decomposes delayed rewards to individual attacking steps proportional to their
effects; (ii) a time-dependent control variate to reduce variance due to large
graphs and many attacking steps. On two social rumor datasets, we demonstrate:
(i) the effectiveness of the attacks compared to rule-based attacks and
end-to-end approaches; (ii) the usefulness of the proposed credit assignment
strategy and control variate; (iii) interpretability of the policy when
generating strong attacks.
Related papers
- FreqFed: A Frequency Analysis-Based Approach for Mitigating Poisoning
Attacks in Federated Learning [98.43475653490219]
Federated learning (FL) is susceptible to poisoning attacks.
FreqFed is a novel aggregation mechanism that transforms the model updates into the frequency domain.
We demonstrate that FreqFed can mitigate poisoning attacks effectively with a negligible impact on the utility of the aggregated model.
arXiv Detail & Related papers (2023-12-07T16:56:24Z) - PACOL: Poisoning Attacks Against Continual Learners [1.569413950416037]
In this work, we demonstrate that continual learning systems can be manipulated by malicious misinformation.
We present a new category of data poisoning attacks specific for continual learners, which we refer to as em Poisoning Attacks Against Continual learners (PACOL)
A comprehensive set of experiments shows the vulnerability of commonly used generative replay and regularization-based continual learning approaches against attack methods.
arXiv Detail & Related papers (2023-11-18T00:20:57Z) - Attacks on Online Learners: a Teacher-Student Analysis [8.567831574941252]
We study the case of adversarial attacks on machine learning models in an online learning setting.
We prove that a discontinuous transition in the learner's accuracy occurs when the attack strength exceeds a critical threshold.
Our findings show that greedy attacks can be extremely efficient, especially when data stream in small batches.
arXiv Detail & Related papers (2023-05-18T17:26:03Z) - Universal Distributional Decision-based Black-box Adversarial Attack
with Reinforcement Learning [5.240772699480865]
We propose a pixel-wise decision-based attack algorithm that finds a distribution of adversarial perturbation through a reinforcement learning algorithm.
Experiments show that the proposed approach outperforms state-of-the-art decision-based attacks with a higher attack success rate and greater transferability.
arXiv Detail & Related papers (2022-11-15T18:30:18Z) - Improving robustness of jet tagging algorithms with adversarial training [56.79800815519762]
We investigate the vulnerability of flavor tagging algorithms via application of adversarial attacks.
We present an adversarial training strategy that mitigates the impact of such simulated attacks.
arXiv Detail & Related papers (2022-03-25T19:57:19Z) - Adversarial Robustness of Deep Reinforcement Learning based Dynamic
Recommender Systems [50.758281304737444]
We propose to explore adversarial examples and attack detection on reinforcement learning-based interactive recommendation systems.
We first craft different types of adversarial examples by adding perturbations to the input and intervening on the casual factors.
Then, we augment recommendation systems by detecting potential attacks with a deep learning-based classifier based on the crafted data.
arXiv Detail & Related papers (2021-12-02T04:12:24Z) - Model-Agnostic Meta-Attack: Towards Reliable Evaluation of Adversarial
Robustness [53.094682754683255]
We propose a Model-Agnostic Meta-Attack (MAMA) approach to discover stronger attack algorithms automatically.
Our method learns the in adversarial attacks parameterized by a recurrent neural network.
We develop a model-agnostic training algorithm to improve the ability of the learned when attacking unseen defenses.
arXiv Detail & Related papers (2021-10-13T13:54:24Z) - Learning and Certification under Instance-targeted Poisoning [49.55596073963654]
We study PAC learnability and certification under instance-targeted poisoning attacks.
We show that when the budget of the adversary scales sublinearly with the sample complexity, PAC learnability and certification are achievable.
We empirically study the robustness of K nearest neighbour, logistic regression, multi-layer perceptron, and convolutional neural network on real data sets.
arXiv Detail & Related papers (2021-05-18T17:48:15Z) - Curse or Redemption? How Data Heterogeneity Affects the Robustness of
Federated Learning [51.15273664903583]
Data heterogeneity has been identified as one of the key features in federated learning but often overlooked in the lens of robustness to adversarial attacks.
This paper focuses on characterizing and understanding its impact on backdooring attacks in federated learning through comprehensive experiments using synthetic and the LEAF benchmarks.
arXiv Detail & Related papers (2021-02-01T06:06:21Z) - Subpopulation Data Poisoning Attacks [18.830579299974072]
Poisoning attacks against machine learning induce adversarial modification of data used by a machine learning algorithm to selectively change its output when it is deployed.
We introduce a novel data poisoning attack called a emphsubpopulation attack, which is particularly relevant when datasets are large and diverse.
We design a modular framework for subpopulation attacks, instantiate it with different building blocks, and show that the attacks are effective for a variety of datasets and machine learning models.
arXiv Detail & Related papers (2020-06-24T20:20:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.