ABACuS: All-Bank Activation Counters for Scalable and Low Overhead RowHammer Mitigation
- URL: http://arxiv.org/abs/2310.09977v3
- Date: Thu, 2 May 2024 13:58:19 GMT
- Title: ABACuS: All-Bank Activation Counters for Scalable and Low Overhead RowHammer Mitigation
- Authors: Ataberk Olgun, Yahya Can Tugrul, Nisa Bostanci, Ismail Emir Yuksel, Haocong Luo, Steve Rhyner, Abdullah Giray Yaglikci, Geraldo F. Oliveira, Onur Mutlu,
- Abstract summary: We introduce ABACuS, a new low-cost hardware-counter-based RowHammer mitigation technique.
ABACuS uses a single shared row activation counter to track activations to the rows with the same row address in all DRAM banks.
Our evaluations show that ABACuS securely prevents RowHammer bitflip/energy overhead and low area cost.
- Score: 6.570851573752742
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce ABACuS, a new low-cost hardware-counter-based RowHammer mitigation technique that performance-, energy-, and area-efficiently scales with worsening RowHammer vulnerability. We observe that both benign workloads and RowHammer attacks tend to access DRAM rows with the same row address in multiple DRAM banks at around the same time. Based on this observation, ABACuS's key idea is to use a single shared row activation counter to track activations to the rows with the same row address in all DRAM banks. Unlike state-of-the-art RowHammer mitigation mechanisms that implement a separate row activation counter for each DRAM bank, ABACuS implements fewer counters (e.g., only one) to track an equal number of aggressor rows. Our evaluations show that ABACuS securely prevents RowHammer bitflips at low performance/energy overhead and low area cost. We compare ABACuS to four state-of-the-art mitigation mechanisms. At a near-future RowHammer threshold of 1000, ABACuS incurs only 0.58% (0.77%) performance and 1.66% (2.12%) DRAM energy overheads, averaged across 62 single-core (8-core) workloads, requiring only 9.47 KiB of storage per DRAM rank. At the RowHammer threshold of 1000, the best prior low-area-cost mitigation mechanism incurs 1.80% higher average performance overhead than ABACuS, while ABACuS requires 2.50X smaller chip area to implement. At a future RowHammer threshold of 125, ABACuS performs very similarly to (within 0.38% of the performance of) the best prior performance- and energy-efficient RowHammer mitigation mechanism while requiring 22.72X smaller chip area. ABACuS is freely and openly available at https://github.com/CMU-SAFARI/ABACuS.
Related papers
- Breaking the Memory Barrier: Near Infinite Batch Size Scaling for Contrastive Loss [59.835032408496545]
We propose a tile-based strategy that partitions the contrastive loss calculation into arbitrary small blocks.
We also introduce a multi-level tiling strategy to leverage the hierarchical structure of distributed systems.
Compared to SOTA memory-efficient solutions, it achieves a two-order-of-magnitude reduction in memory while maintaining comparable speed.
arXiv Detail & Related papers (2024-10-22T17:59:30Z) - MINT: Securely Mitigating Rowhammer with a Minimalist In-DRAM Tracker [0.9939073004351231]
This paper investigates secure low-cost in-DRAM trackers for mitigating Rowhammer (RH)
Existing low-cost in-DRAM trackers require impractical overheads of hundreds or thousands of entries per bank.
We propose a Minimalist In-DRAM Tracker (MINT), which provides secure mitigation with just a single entry.
arXiv Detail & Related papers (2024-07-22T20:29:56Z) - MOAT: Securely Mitigating Rowhammer with Per-Row Activation Counters [0.3580891736370874]
DDR5 specifications have been extended to support Per-Row Activation Counting (PRAC), with counters inlined with each row, and ALERT-Back-Off (ABO) to stop the memory controller if the DRAM needs more time to mitigate.
Although PRAC+ABO represents a strong advance in Rowhammer protection, they are just a framework, and the actual security is dependent on the implementation.
We propose MOAT, a provably secure design, which uses two internal thresholds: ETH, an "Eligibility Threshold" for mitigating a row, and ATH, an "ALERT Threshold" for initiating
arXiv Detail & Related papers (2024-07-13T20:28:02Z) - BreakHammer: Enhancing RowHammer Mitigations by Carefully Throttling Suspect Threads [5.767293823380473]
RowHammer is a read disturbance mechanism in DRAM where repeatedly accessing (hammering) a row of DRAM cells (DRAM row) induces bitflips in other physically nearby DRAM rows.
RowHammer solutions perform preventive actions (e.g., refresh neighbor rows of the hammered row) that mitigate such bitflips.
As shrinking technology node size over DRAM chip generations exacerbates RowHammer, the overheads of RowHammer solutions become prohibitively expensive.
In this work, we tackle the performance overheads of RowHammer solutions by tracking and throttling the generators of memory accesses that trigger RowHammer solutions.
arXiv Detail & Related papers (2024-04-20T22:09:38Z) - CoMeT: Count-Min-Sketch-based Row Tracking to Mitigate RowHammer at Low Cost [6.658509339087993]
CoMeT prevents RowHammer bitflips with low area, performance, and energy costs.
CoMeT uses low-cost and scalable hash-based counters to track DRAM row activations.
CoMeT achieves a good trade-off between performance, energy, and area overheads.
arXiv Detail & Related papers (2024-02-29T00:21:29Z) - EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs [68.41135269685576]
Existing methods for fine-tuning sparse LLMs often suffer from resource-intensive requirements and high retraining costs.
We propose an efficient and fast framework for fine-tuning sparse LLMs based on minimizing reconstruction error.
Our approach involves sampling a small dataset for calibration and utilizing backpropagation to iteratively optimize block-wise reconstruction error.
arXiv Detail & Related papers (2024-02-19T09:55:32Z) - HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM
Inference [68.59839755875252]
HiRE comprises of two novel components: (i) a compression scheme to cheaply predict top-$k$ rows/columns with high recall, followed by full computation restricted to the predicted subset, and (ii) DA-TOP-$k$: an efficient multi-device approximate top-$k$ operator.
We demonstrate that on a one billion parameter model, HiRE applied to both the softmax as well as feedforward layers, achieves almost matching pretraining and downstream accuracy, and speeds up inference latency by $1.47times$ on a single TPUv5e device.
arXiv Detail & Related papers (2024-02-14T18:04:36Z) - Randomized Line-to-Row Mapping for Low-Overhead Rowhammer Mitigations [0.716879432974126]
Recently proposed secure Rowhammer mitigations rely on performing mitigative action on the aggressor rather than the victims.
We propose Rubix, which breaks the spatial correlation in the line-to-row mapping by using an encrypted address to access the memory.
We also propose Rubix-D, which dynamically changes the line-to-row mapping.
arXiv Detail & Related papers (2023-08-28T21:22:15Z) - Contextual Bandits with Packing and Covering Constraints: A Modular Lagrangian Approach via Regression [65.8785736964253]
We consider contextual bandits with linear constraints (CBwLC), a variant of contextual bandits in which the algorithm consumes multiple resources subject to linear constraints on total consumption.
This problem generalizes contextual bandits with knapsacks (CBwK), allowing for packing and covering constraints, as well as positive and negative resource consumption.
We provide the first algorithm for CBwLC (or CBwK) that is based on regression oracles. The algorithm is simple, computationally efficient, and statistically optimal under mild assumptions.
arXiv Detail & Related papers (2022-11-14T16:08:44Z) - Improved Branch and Bound for Neural Network Verification via Lagrangian
Decomposition [161.09660864941603]
We improve the scalability of Branch and Bound (BaB) algorithms for formally proving input-output properties of neural networks.
We present a novel activation-based branching strategy and a BaB framework, named Branch and Dual Network Bound (BaDNB)
BaDNB outperforms previous complete verification systems by a large margin, cutting average verification times by factors up to 50 on adversarial properties.
arXiv Detail & Related papers (2021-04-14T09:22:42Z) - Scaling the Convex Barrier with Sparse Dual Algorithms [141.4085318878354]
We present two novel dual algorithms for tight and efficient neural network bounding.
Both methods recover the strengths of the new relaxation: tightness and a linear separation oracle.
We can obtain better bounds than off-the-shelf solvers in only a fraction of their running time.
arXiv Detail & Related papers (2021-01-14T19:45:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.