Bribe & Fork: Cheap Bribing Attacks via Forking Threat
- URL: http://arxiv.org/abs/2402.01363v2
- Date: Wed, 17 Jul 2024 06:30:17 GMT
- Title: Bribe & Fork: Cheap Bribing Attacks via Forking Threat
- Authors: Zeta Avarikioti, Paweł Kędzior, Tomasz Lizurej, Tomasz Michalak,
- Abstract summary: Bribe & Fork is a modified bribing attack that leverages the threat of a so-called feather fork.
We empirically analyze historical data of some real-world blockchain implementations to evaluate the scale of this cost reduction.
Our findings shed more light on the potential vulnerability of PCNs and highlight the need for robust solutions.
- Score: 2.9061423802698565
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we reexamine the vulnerability of Payment Channel Networks (PCNs) to bribing attacks, where an adversary incentivizes blockchain miners to deliberately ignore a specific transaction to undermine the punishment mechanism of PCNs. While previous studies have posited a prohibitive cost for such attacks, we show that this cost may be dramatically reduced (to approximately \$125), thereby increasing the likelihood of these attacks. To this end, we introduce Bribe & Fork, a modified bribing attack that leverages the threat of a so-called feather fork which we analyze with a novel formal model for the mining game with forking. We empirically analyze historical data of some real-world blockchain implementations to evaluate the scale of this cost reduction. Our findings shed more light on the potential vulnerability of PCNs and highlight the need for robust solutions.
Related papers
- BM-PAW: A Profitable Mining Attack in the PoW-based Blockchain System [9.292531856119329]
We introduce a novel mining strategy, named BM-PAW, which yields superior rewards for both the attacker and the targeted pool.
We find the BM-PAW attacker can circumvent the "miner's dilemma" through equilibrium analysis in a two-pool BM-PAW game scenario.
arXiv Detail & Related papers (2024-11-09T13:59:55Z) - Examining Attacks on Consensus and Incentive Systems in Proof-of-Work Blockchains: A Systematic Literature Review [0.0]
Bitcoin's security relies on a decentralized ledger consisting of a consensus and an incentive mechanism.
As Bitcoin's acceptance grows, it faces increasing threats from attacks targeting these mechanisms.
This paper begins by examining individual attacks executed in isolation and their profitability.
It then explores how combining these attacks with each other or with other malicious and non-malicious strategies can enhance their overall effectiveness and profitability.
arXiv Detail & Related papers (2024-11-01T04:18:42Z) - Fully Automated Selfish Mining Analysis in Efficient Proof Systems Blockchains [5.864854777864723]
We study selfish mining attacks in longest-chain blockchains like Bitcoin, but where the proof of work is replaced with efficient proof systems.
We propose a novel selfish mining attack that aims to maximize expected relative revenue of the adversary.
We present a formal analysis procedure which computes an $epsilon$-tight lower bound on the optimal expected relative revenue in the MDP.
arXiv Detail & Related papers (2024-05-07T15:44:39Z) - Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery
Detection [62.595450266262645]
This paper introduces a novel and previously unrecognized threat in face forgery detection scenarios caused by backdoor attack.
By embedding backdoors into models, attackers can deceive detectors into producing erroneous predictions for forged faces.
We propose emphPoisoned Forgery Face framework, which enables clean-label backdoor attacks on face forgery detectors.
arXiv Detail & Related papers (2024-02-18T06:31:05Z) - Cost Aware Untargeted Poisoning Attack against Graph Neural Networks, [5.660584039688214]
We propose a novel attack loss framework called the Cost Aware Poisoning Attack (CA-attack) to improve the allocation of the attack budget.
Our experiments demonstrate that the proposed CA-attack significantly enhances existing attack strategies.
arXiv Detail & Related papers (2023-12-12T10:54:02Z) - Attention-Enhancing Backdoor Attacks Against BERT-based Models [54.070555070629105]
Investigating the strategies of backdoor attacks will help to understand the model's vulnerability.
We propose a novel Trojan Attention Loss (TAL) which enhances the Trojan behavior by directly manipulating the attention patterns.
arXiv Detail & Related papers (2023-10-23T01:24:56Z) - ADESS: A Proof-of-Work Protocol to Deter Double-Spend Attacks [0.0]
A principal vulnerability of a proof-of-work ("PoW") blockchain is that an attacker can re-write the history of transactions.
We propose a modification to PoW protocols, called ADESS, that contains two novel features.
arXiv Detail & Related papers (2023-09-25T21:50:23Z) - Online Adversarial Attacks [57.448101834579624]
We formalize the online adversarial attack problem, emphasizing two key elements found in real-world use-cases.
We first rigorously analyze a deterministic variant of the online threat model.
We then propose algoname, a simple yet practical algorithm yielding a provably better competitive ratio for $k=2$ over the current best single threshold algorithm.
arXiv Detail & Related papers (2021-03-02T20:36:04Z) - Quantum Multi-Solution Bernoulli Search with Applications to Bitcoin's
Post-Quantum Security [67.06003361150228]
A proof of work (PoW) is an important cryptographic construct enabling a party to convince others that they invested some effort in solving a computational task.
In this work, we examine the hardness of finding such chain of PoWs against quantum strategies.
We prove that the chain of PoWs problem reduces to a problem we call multi-solution Bernoulli search, for which we establish its quantum query complexity.
arXiv Detail & Related papers (2020-12-30T18:03:56Z) - Action-Manipulation Attacks Against Stochastic Bandits: Attacks and
Defense [45.408568528354216]
We introduce a new class of attack named action-manipulation attack.
In this attack, an adversary can change the action signal selected by the user.
To defend against this class of attacks, we introduce a novel algorithm that is robust to action-manipulation attacks.
arXiv Detail & Related papers (2020-02-19T04:09:15Z) - Adversarial Attacks on Linear Contextual Bandits [87.08004581867537]
Malicious agents may have incentives to attack the bandit algorithm to induce it to perform a desired behavior.
We show that a malicious agent can force a linear contextual bandit algorithm to pull any desired arm $T - o(T)$ times over a horizon of $T$ steps.
We also investigate the case when a malicious agent is interested in affecting the behavior of the bandit algorithm in a single context.
arXiv Detail & Related papers (2020-02-10T15:04:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.