Pioplat: A Scalable, Low-Cost Framework for Latency Reduction in Ethereum Blockchain
- URL: http://arxiv.org/abs/2412.08367v1
- Date: Wed, 11 Dec 2024 13:15:16 GMT
- Title: Pioplat: A Scalable, Low-Cost Framework for Latency Reduction in Ethereum Blockchain
- Authors: Ke Wang, Qiao Wang, Yue Li, Zhi Guan, Zhong Chen,
- Abstract summary: Pioplat is a feasible, customizable, and low-cost latency reduction framework.<n>We show that Pioplat can significantly reduce the latency of receiving blocks/transactions and sending transactions.
- Score: 10.807408760944632
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As decentralized applications on permissionless blockchains are prevalent, more and more latency-sensitive usage scenarios emerged, where the lower the latency of sending and receiving messages, the better the chance of earning revenue. To reduce latency, we present Pioplat, a feasible, customizable, and low-cost latency reduction framework consisting of multiple relay nodes on different continents and at least one instrumented variant of a full node. The node selection strategy of Pioplat and the low-latency communication protocol offer an elastic way to reduce latency effectively. We demonstrate Pioplat's feasibility with an implementation running on five continents and show that Pioplat can significantly reduce the latency of receiving blocks/transactions and sending transactions, thus fulfilling the requirements of most latency-sensitive use cases. Furthermore, we provide the complete implementation of Pioplat to promote further research and allow people to apply the framework to more blockchain systems.
Related papers
- HALO: Semantic-Aware Distributed LLM Inference in Lossy Edge Network [50.33808558714122]
Large language models' (LLMs) inference at the edge can facilitate prompt service responsiveness while protecting user privacy.<n>We propose HALO, a novel framework that can boost the distributed LLM inference in lossy edge network.<n> Experimental results from a Raspberry Pi cluster demonstrate that HALO achieves a 3.41x end-to-end speedup for LLaMA-series LLMs under unreliable network conditions.
arXiv Detail & Related papers (2026-01-16T07:37:23Z) - A Range-Based Sharding (RBS) Protocol for Scalable Enterprise Blockchain [0.4941723903728161]
This paper proposes the Range-Based Sharding Protocol, a novel sharding mechanism tailored for enterprise blockchains.<n>RBS achieves significantly higher throughput and lower latency compared to existing enterprise sharding frameworks.
arXiv Detail & Related papers (2025-09-13T23:09:38Z) - Dynamic Speculative Agent Planning [57.630218933994534]
Large language-model-based agents face critical deployment challenges due to prohibitive latency and inference costs.<n>We introduce Dynamic Speculative Planning (DSP), an online reinforcement learning framework that provides lossless acceleration with substantially reduced costs.<n>Experiments on two standard agent benchmarks demonstrate that DSP achieves comparable efficiency to the fastest acceleration method while reducing total cost by 30% and unnecessary cost up to 60%.
arXiv Detail & Related papers (2025-09-02T03:34:36Z) - CSGO: Generalized Optimization for Cold Start in Wireless Collaborative Edge LLM Systems [62.24576366776727]
We propose a latency-aware scheduling framework to minimize total inference latency.<n>We show that the proposed method significantly reduces cold-start latency compared to baseline strategies.
arXiv Detail & Related papers (2025-08-15T07:49:22Z) - Task-Oriented Feature Compression for Multimodal Understanding via Device-Edge Co-Inference [49.77734021302196]
We propose a task-oriented feature compression (TOFC) method for multimodal understanding in a device-edge co-inference framework.
To enhance compression efficiency, multiple entropy models are adaptively selected based on the characteristics of the visual features.
Results show that TOFC achieves up to 60% reduction in data transmission overhead and 50% reduction in system latency.
arXiv Detail & Related papers (2025-03-17T08:37:22Z) - Zaptos: Towards Optimal Blockchain Latency [52.30047458198369]
We introduce Zaptos, a parallel pipelined architecture designed to minimize end-to-end latency.
Zaptos achieves a throughput of 20,000 transactions per second with sub-second latency.
arXiv Detail & Related papers (2025-01-18T00:22:22Z) - ConServe: Fine-Grained GPU Harvesting for LLM Online and Offline Co-Serving [61.35068981176018]
ConServe is a large language model (LLM) serving system that achieves high throughput and strong online latency guarantees.<n>We show that ConServe delivers an average of 2.2$times$ higher throughput and reduces online serving tail latency by 2.9$times$ on average compared to state-of-the-art systems.
arXiv Detail & Related papers (2024-10-02T04:12:13Z) - The Latency Price of Threshold Cryptosystem in Blockchains [52.359230560289745]
We study the interplay between threshold cryptography and a class of blockchains that use Byzantine-fault tolerant (BFT) consensus protocols.
Existing approaches for threshold cryptosystems introduce a latency overhead of at least one message delay for running the threshold cryptographic protocol.
We propose a mechanism to eliminate this overhead for blockchain-native threshold cryptosystems with tight thresholds.
arXiv Detail & Related papers (2024-07-16T20:53:04Z) - Kronos: A Secure and Generic Sharding Blockchain Consensus with Optimized Overhead [17.387814385763622]
Sharding enhances blockchain scalability by dividing the network into shards, each managing specific unspent transaction outputs or accounts.
Cross-shard transactions pose a critical challenge to the security and efficiency of sharding blockchains.
We present Kronos, a secure sharding blockchain consensus achieving optimized overhead.
arXiv Detail & Related papers (2024-03-06T12:26:04Z) - Client Orchestration and Cost-Efficient Joint Optimization for
NOMA-Enabled Hierarchical Federated Learning [55.49099125128281]
We propose a non-orthogonal multiple access (NOMA) enabled HFL system under semi-synchronous cloud model aggregation.
We show that the proposed scheme outperforms the considered benchmarks regarding HFL performance improvement and total cost reduction.
arXiv Detail & Related papers (2023-11-03T13:34:44Z) - Mysticeti: Reaching the Limits of Latency with Uncertified DAGs [5.328717371685882]
We introduce Mysticeti-C, the first DAG-based Byzantine consensus protocol to achieve the lower bounds of latency of 3 message rounds.
We extend Mysticeti-C to Mysticeti-FPC, which incorporates a fast commit path that achieves even lower latency for transferring assets.
arXiv Detail & Related papers (2023-10-23T11:40:50Z) - Relay Mining: Incentivizing Full Non-Validating Nodes Servicing All RPC Types [0.0]
Relay Mining estimates and proves the volume of Remote Procedure Calls (RPCs) made from a client to a server.
We leverage digital signatures, commit-and-reveal schemes, and Sparse Merkle Sum Tries (SMSTs) to prove the amount of work done.
A native cryptocurrency on a distributed ledger is used to rate limit applications and disincentivize over-usage.
arXiv Detail & Related papers (2023-05-18T03:23:41Z) - Efficient Parallel Split Learning over Resource-constrained Wireless
Edge Networks [44.37047471448793]
In this paper, we advocate the integration of edge computing paradigm and parallel split learning (PSL)
We propose an innovative PSL framework, namely, efficient parallel split learning (EPSL) to accelerate model training.
We show that the proposed EPSL framework significantly decreases the training latency needed to achieve a target accuracy.
arXiv Detail & Related papers (2023-03-26T16:09:48Z) - An Intelligent Deterministic Scheduling Method for Ultra-Low Latency
Communication in Edge Enabled Industrial Internet of Things [19.277349546331557]
Time Sensitive Network (TSN) is recently researched to realize low latency communication via deterministic scheduling.
Non-collision theory based deterministic scheduling (NDS) method is proposed to achieve ultra-low latency communication for the time-sensitive flows.
Experiment results demonstrate that NDS/DQS can well support deterministic ultra-low latency services and guarantee efficient bandwidth utilization.
arXiv Detail & Related papers (2022-07-17T16:52:51Z) - Jolteon and Ditto: Network-Adaptive Efficient Consensus with Asynchronous Fallback [46.30924494799245]
We develop Ditto, a Byzantine SMR protocol that enjoys the best of both worlds: optimal communication on and off the happy path and progress guarantee under asynchrony and DDoS attacks.
Specifically, we start from HotStuff, a state-of-the-art linear protocol, and gradually build Ditto. As a separate contribution and an intermediate step, we design a 2-chain version of HotStuff, Jolteon.
We implement and experimentally evaluate all our systems. Notably, Jolteon's commit latency outperforms HotStuff by 200-300ms with varying system size.
arXiv Detail & Related papers (2021-06-18T21:34:17Z) - Stochastic bandits with arm-dependent delays [102.63128271054741]
We propose a simple but efficient UCB-based algorithm called the PatientBandits.
We provide both problems-dependent and problems-independent bounds on the regret as well as performance lower bounds.
arXiv Detail & Related papers (2020-06-18T12:13:58Z) - Low Latency ASR for Simultaneous Speech Translation [27.213294097841853]
We have worked on several techniques for reducing the latency for both components, the automatic speech recognition and the speech translation module.
We combined run-on decoding with a technique for identifying stable partial hypotheses when stream decoding and a protocol for dynamic output update.
This combination reduces the latency at word level, where the words are final and will never be updated again in the future, from 18.1s to 1.1s without sacrificing performance.
arXiv Detail & Related papers (2020-03-22T13:37:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.