SoK: Comprehensive Analysis of Rug Pull Causes, Datasets, and Detection Tools in DeFi
- URL: http://arxiv.org/abs/2403.16082v1
- Date: Sun, 24 Mar 2024 10:24:17 GMT
- Title: SoK: Comprehensive Analysis of Rug Pull Causes, Datasets, and Detection Tools in DeFi
- Authors: Dianxiang Sun, Wei Ma, Liming Nie, Yang Liu,
- Abstract summary: Rug pulls pose a grave threat to the cryptocurrency ecosystem, leading to substantial financial loss and undermining trust in decentralized finance (DeFi) projects.
With the emergence of new rug pull patterns, research on rug pull is out of state.
We present a taxonomy inclusive of 34 root causes, introducing six new categories inspired by industry sources: burn, hidden owner, ownership transfer, unverified contract, external call, and fake LP lock.
- Score: 14.172486637733797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rug pulls pose a grave threat to the cryptocurrency ecosystem, leading to substantial financial loss and undermining trust in decentralized finance (DeFi) projects. With the emergence of new rug pull patterns, research on rug pull is out of state. To fill this gap, we first conducted an extensive analysis of the literature review, encompassing both scholarly and industry sources. By examining existing academic articles and industrial discussions on rug pull projects, we present a taxonomy inclusive of 34 root causes, introducing six new categories inspired by industry sources: burn, hidden owner, ownership transfer, unverified contract, external call, and fake LP lock. Based on the developed taxonomy, we evaluated current rug pull datasets and explored the effectiveness and limitations of existing detection mechanisms. Our evaluation indicates that the existing datasets, which document 2,448 instances, address only 7 of the 34 root causes, amounting to a mere 20% coverage. It indicates that existing open-source datasets need to be improved to study rug pulls. In response, we have constructed a more comprehensive dataset containing 2,360 instances, expanding the coverage to 54% with the best effort. In addition, the examination of 14 detection tools showed that they can identify 25 of the 34 root causes, achieving a coverage of 73.5%. There are nine root causes (Fake LP Lock, Hidden Fee, and Destroy Token, Fake Money Transfer, Ownership Transfer, Liquidity Pool Block, Freeze Account, Wash-Trading, Hedge) that the existing tools cannot cover. Our work indicates that there is a significant gap between current research and detection tools, and the actual situation of rug pulls.
Related papers
- Retrieval-Augmented Generation with Conflicting Evidence [57.66282463340297]
Large language model (LLM) agents are increasingly employing retrieval-augmented generation (RAG) to improve the factuality of their responses.
In practice, these systems often need to handle ambiguous user queries and potentially conflicting information from multiple sources.
We propose RAMDocs (Retrieval with Ambiguity and Misinformation in Documents), a new dataset that simulates complex and realistic scenarios for conflicting evidence for a user query.
arXiv Detail & Related papers (2025-04-17T16:46:11Z) - SolRPDS: A Dataset for Analyzing Rug Pulls in Solana Decentralized Finance [0.6367946001576646]
Rug pulls in Solana have caused significant damage to users interacting with Decentralized Finance (DeFi)
A rug pull occurs when developers exploit users' trust and drain liquidity from token pools on Decentralized Exchanges (DEXs)
We introduce SolRPDS, the first public rug pull dataset derived from Solana's transactions.
arXiv Detail & Related papers (2025-04-06T11:36:48Z) - Fixing Outside the Box: Uncovering Tactics for Open-Source Security Issue Management [9.990683064304207]
We conduct a comprehensive study on the taxonomy of vulnerability remediation tactics (RT) in OSS projects.
We developed a hierarchical taxonomy of 44 distinct RT and evaluated their effectiveness and costs.
Our findings highlight a significant reliance on community-driven strategies, like using alternative libraries and bypassing vulnerabilities.
arXiv Detail & Related papers (2025-03-30T08:24:58Z) - Benchmarking Reasoning Robustness in Large Language Models [76.79744000300363]
We find significant performance degradation on novel or incomplete data.
These findings highlight the reliance on recall over rigorous logical inference.
This paper introduces a novel benchmark, termed as Math-RoB, that exploits hallucinations triggered by missing information to expose reasoning gaps.
arXiv Detail & Related papers (2025-03-06T15:36:06Z) - SoK: Detection and Repair of Accessibility Issues [10.134645262631983]
We develop a comprehensive taxonomy that categorizes 55 types of accessibility issues across four pivotal dimensions: Perceivability, Operability, Understandability, and Robustness.
We conduct an in-depth analysis of existing detection and repair tools, as well as the status of corresponding datasets.
arXiv Detail & Related papers (2024-11-29T14:19:19Z) - Contrastive Learning to Improve Retrieval for Real-world Fact Checking [84.57583869042791]
We present Contrastive Fact-Checking Reranker (CFR), an improved retriever for fact-checking complex claims.
We leverage the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents.
We find a 6% improvement in veracity classification accuracy on the dataset.
arXiv Detail & Related papers (2024-10-07T00:09:50Z) - Thinking Racial Bias in Fair Forgery Detection: Models, Datasets and Evaluations [63.52709761339949]
We first contribute a dedicated dataset called the Fair Forgery Detection (FairFD) dataset, where we prove the racial bias of public state-of-the-art (SOTA) methods.
We design novel metrics including Approach Averaged Metric and Utility Regularized Metric, which can avoid deceptive results.
We also present an effective and robust post-processing technique, Bias Pruning with Fair Activations (BPFA), which improves fairness without requiring retraining or weight updates.
arXiv Detail & Related papers (2024-07-19T14:53:18Z) - Examining Ownership Models in Software Teams: A Systematic Literature Review and a Replication Study [2.0891120283967264]
We identify 79 relevant papers published between 2005 and 2022.
We develop a taxonomy of ownership artifacts based on type, owners, and degree of ownership.
arXiv Detail & Related papers (2024-05-24T16:03:22Z) - Fact Checking Beyond Training Set [64.88575826304024]
We show that the retriever-reader suffers from performance deterioration when it is trained on labeled data from one domain and used in another domain.
We propose an adversarial algorithm to make the retriever component robust against distribution shift.
We then construct eight fact checking scenarios from these datasets, and compare our model to a set of strong baseline models.
arXiv Detail & Related papers (2024-03-27T15:15:14Z) - Exposing the Deception: Uncovering More Forgery Clues for Deepfake
Detection [36.92399832886853]
Current deepfake detection approaches may easily fall into the trap of overfitting, focusing only on forgery clues within one or a few local regions.
We present a novel framework to capture broader forgery clues by extracting multiple non-overlapping local representations and fusing them into a global semantic-rich feature.
Our method achieves state-of-the-art performance on five benchmark datasets.
arXiv Detail & Related papers (2024-03-04T07:28:23Z) - SEPSIS: I Can Catch Your Lies -- A New Paradigm for Deception Detection [9.20397189600732]
This research explores the problem of deception through the lens of psychology.
We propose a novel framework for deception detection leveraging NLP techniques.
We present a novel multi-task learning pipeline that leverages the dataless merging of fine-tuned language models.
arXiv Detail & Related papers (2023-12-01T02:13:25Z) - Inferring Resource-Oriented Intentions using LLMs for Static Resource Leak Detection [14.783216988363804]
Resource leaks, caused by resources not being released after acquisition, often lead to performance issues and system crashes.
Existing static detection techniques rely on mechanical matching of predefined resource acquisition/release APIs and null-checking conditions to find unreleased resources.
We propose InferROI, a novel approach that directly infers resource-oriented intentions (acquisition, release, and reachability validation) in code.
arXiv Detail & Related papers (2023-11-08T04:19:28Z) - Untargeted Backdoor Watermark: Towards Harmless and Stealthy Dataset
Copyright Protection [69.59980270078067]
We explore the untargeted backdoor watermarking scheme, where the abnormal model behaviors are not deterministic.
We also discuss how to use the proposed untargeted backdoor watermark for dataset ownership verification.
arXiv Detail & Related papers (2022-09-27T12:56:56Z) - Black-box Dataset Ownership Verification via Backdoor Watermarking [67.69308278379957]
We formulate the protection of released datasets as verifying whether they are adopted for training a (suspicious) third-party model.
We propose to embed external patterns via backdoor watermarking for the ownership verification to protect them.
Specifically, we exploit poison-only backdoor attacks ($e.g.$, BadNets) for dataset watermarking and design a hypothesis-test-guided method for dataset verification.
arXiv Detail & Related papers (2022-08-04T05:32:20Z) - HoVer: A Dataset for Many-Hop Fact Extraction And Claim Verification [74.66819506353086]
HoVer is a dataset for many-hop evidence extraction and fact verification.
It challenges models to extract facts from several Wikipedia articles that are relevant to a claim.
Most of the 3/4-hop claims are written in multiple sentences, which adds to the complexity of understanding long-range dependency relations.
arXiv Detail & Related papers (2020-11-05T20:33:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.