FLAC: Practical Failure-Aware Atomic Commit Protocol for Distributed
Transactions
- URL: http://arxiv.org/abs/2302.04500v1
- Date: Thu, 9 Feb 2023 08:52:11 GMT
- Title: FLAC: Practical Failure-Aware Atomic Commit Protocol for Distributed
Transactions
- Authors: Hexiang Pan, Quang-Trung Ta, Meihui Zhang, Yeow Meng Chee, Gang Chen,
Beng Chin Ooi
- Abstract summary: Failure-Aware Atomic Commit (FLAC) is designed for three different environments.
FLAC monitors if any failure occurs and switches to operate the most suitable sub-protocol.
It achieves up to 2.22x throughput improvement and 2.82x latency speedup.
- Score: 27.20381433013882
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In distributed transaction processing, atomic commit protocol (ACP) is used
to ensure database consistency. With the use of commodity compute nodes and
networks, failures such as system crashes and network partitioning are common.
It is therefore important for ACP to dynamically adapt to the operating
condition for efficiency while ensuring the consistency of the database.
Existing ACPs often assume stable operating conditions, hence, they are either
non-generalizable to different environments or slow in practice.
In this paper, we propose a novel and practical ACP, called Failure-Aware
Atomic Commit (FLAC). In essence, FLAC includes three sub-protocols, which are
specifically designed for three different environments: (i) no failure occurs,
(ii) participant nodes might crash but there is no delayed connection, or (iii)
both crashed nodes and delayed connection can occur. It models these
environments as the failure-free, crash-failure, and network-failure robustness
levels. During its operation, FLAC can monitor if any failure occurs and
dynamically switch to operate the most suitable sub-protocol, using a
robustness level state machine, whose parameters are fine-tuned by
reinforcement learning. Consequently, it improves both the response time and
throughput, and effectively handles nodes distributed across the Internet where
crash and network failures might occur. We implement FLAC in a distributed
transactional key-value storage system based on Google Percolator and evaluate
its performance with both a micro benchmark and a macro benchmark of real
workload. The results show that FLAC achieves up to 2.22x throughput
improvement and 2.82x latency speedup, compared to existing ACPs for
high-contention workloads.
Related papers
- C2A: Client-Customized Adaptation for Parameter-Efficient Federated Learning [21.914696641277054]
We propose a novel hypernetwork-based FL framework that generates client-specific adapters by conditioning the client information.
With the effectiveness of the hypernetworks in generating customized weights through learning, C2A can maximize the utility of shared model parameters.
Comprehensive evaluation results clearly support the superiority of C2A in terms of both efficiency and effectiveness in FL scenarios.
arXiv Detail & Related papers (2024-11-01T02:07:38Z) - FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - SeBS-Flow: Benchmarking Serverless Cloud Function Workflows [51.4200085836966]
We propose the first serverless workflow benchmarking suite SeBS-Flow.
SeBS-Flow includes six real-world application benchmarks and four microbenchmarks representing different computational patterns.
We conduct comprehensive evaluations on three major cloud platforms, assessing performance, cost, scalability, and runtime deviations.
arXiv Detail & Related papers (2024-10-04T14:52:18Z) - Continuous variable dense coding under realistic non-ideal scenarios [0.0]
We derive a general formalism for the dense coding capacity (DCC) of generic two-mode Gaussian states.
We investigate the pattern of DCC of the two-mode squeezed vacuum state (TMSV) by varying the strength of the noise.
arXiv Detail & Related papers (2024-07-10T12:49:17Z) - Scalable Federated Unlearning via Isolated and Coded Sharding [76.12847512410767]
Federated unlearning has emerged as a promising paradigm to erase the client-level data effect.
This paper proposes a scalable federated unlearning framework based on isolated sharding and coded computing.
arXiv Detail & Related papers (2024-01-29T08:41:45Z) - FLCC: Efficient Distributed Federated Learning on IoMT over CSMA/CA [0.0]
Federated Learning (FL) has emerged as a promising approach for privacy preservation.
This article investigates the performance of FL on an application that might be used to improve a remote healthcare system over ad hoc networks.
We present two metrics to evaluate the network performance: 1) probability of successful transmission while minimizing the interference, and 2) performance of distributed FL model in terms of accuracy and loss.
arXiv Detail & Related papers (2023-03-29T16:36:42Z) - HFedMS: Heterogeneous Federated Learning with Memorable Data Semantics
in Industrial Metaverse [49.1501082763252]
This paper presents HFEDMS for incorporating practical FL into the emerging Industrial Metaverse.
It reduces data heterogeneity through dynamic grouping and training mode conversion.
Then, it compensates for the forgotten knowledge by fusing compressed historical data semantics.
Experiments have been conducted on the streamed non-i.i.d. FEMNIST dataset using 368 simulated devices.
arXiv Detail & Related papers (2022-11-07T04:33:24Z) - Scalable and Sparsity-Aware Privacy-Preserving K-means Clustering with
Application to Fraud Detection [12.076075765740502]
We propose a new framework for efficient sparsity-aware K-means with three characteristics.
First, our framework is divided into a data-independent offline phase and a much faster online phase.
Second, we take advantage of the vectorization techniques in both online and offline phases.
Third, we adopt a sparse matrix multiplication for the data sparsity scenario to improve efficiency further.
arXiv Detail & Related papers (2022-08-12T02:58:26Z) - Communication-Efficient Federated Learning With Data and Client
Heterogeneity [22.432529149142976]
Federated Learning (FL) enables large-scale distributed training of machine learning models.
executing FL at scale comes with inherent practical challenges.
We present the first variant of the classic federated averaging (FedAvg) algorithm.
arXiv Detail & Related papers (2022-06-20T22:39:39Z) - Higher Performance Visual Tracking with Dual-Modal Localization [106.91097443275035]
Visual Object Tracking (VOT) has synchronous needs for both robustness and accuracy.
We propose a dual-modal framework for target localization, consisting of robust localization suppressingors via ONR and the accurate localization attending to the target center precisely via OFC.
arXiv Detail & Related papers (2021-03-18T08:47:56Z) - AQD: Towards Accurate Fully-Quantized Object Detection [94.06347866374927]
We propose an Accurate Quantized object Detection solution, termed AQD, to get rid of floating-point computation.
Our AQD achieves comparable or even better performance compared with the full-precision counterpart under extremely low-bit schemes.
arXiv Detail & Related papers (2020-07-14T09:07:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.