Smoothening block rewards: How much should miners pay for mining pools?
- URL: http://arxiv.org/abs/2309.02297v1
- Date: Tue, 5 Sep 2023 14:59:01 GMT
- Title: Smoothening block rewards: How much should miners pay for mining pools?
- Authors: Axel Cortes-Cubero, Juan P. Madrigal-Cianci, Kiran Karra, Zixuan Zhang,
- Abstract summary: We quantify the economic advantage for a given miner of having smooth rewards.
We use this to define a maximum percentage of rewards that a miner should be willing to pay for the mining pool services.
- Score: 11.245119287096419
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The rewards a blockchain miner earns vary with time. Most of the time is spent mining without receiving any rewards, and only occasionally the miner wins a block and earns a reward. Mining pools smoothen the stochastic flow of rewards, and in the ideal case, provide a steady flow of rewards over time. Smooth block rewards allow miners to choose an optimal mining power growth strategy that will result in a higher reward yield for a given investment. We quantify the economic advantage for a given miner of having smooth rewards, and use this to define a maximum percentage of rewards that a miner should be willing to pay for the mining pool services.
Related papers
- Walking the Values in Bayesian Inverse Reinforcement Learning [66.68997022043075]
Key challenge in Bayesian IRL is bridging the computational gap between the hypothesis space of possible rewards and the likelihood.
We propose ValueWalk - a new Markov chain Monte Carlo method based on this insight.
arXiv Detail & Related papers (2024-07-15T17:59:52Z) - Helping or Herding? Reward Model Ensembles Mitigate but do not Eliminate Reward Hacking [62.146953368613815]
Reward models play a key role in aligning language model applications towards human preferences.
A natural mitigation is to train an ensemble of reward models, aggregating over model outputs to obtain a more robust reward estimate.
We show that reward ensembles do not eliminate reward hacking because all reward models in the ensemble exhibit similar error patterns.
arXiv Detail & Related papers (2023-12-14T18:59:04Z) - DreamSmooth: Improving Model-based Reinforcement Learning via Reward
Smoothing [60.21269454707625]
DreamSmooth learns to predict a temporally-smoothed reward, instead of the exact reward at the given timestep.
We show that DreamSmooth achieves state-of-the-art performance on long-horizon sparse-reward tasks.
arXiv Detail & Related papers (2023-11-02T17:57:38Z) - STARC: A General Framework For Quantifying Differences Between Reward
Functions [55.33869271912095]
We provide a class of pseudometrics on the space of all reward functions that we call STARC metrics.
We show that STARC metrics induce both an upper and a lower bound on worst-case regret.
We also identify a number of issues with reward metrics proposed by earlier works.
arXiv Detail & Related papers (2023-09-26T20:31:19Z) - Undetectable Selfish Mining [4.625489011466493]
A strategic Bitcoin miner may profit by deviating from the intended Bitcoin protocol.
We develop a selfish mining variant that is provably *statistically undetectable*
We show that our strategy is strictly profitable for attackers with $38.2% ll 50%$ of the total hashrate.
arXiv Detail & Related papers (2023-09-13T09:51:32Z) - Go Beyond Imagination: Maximizing Episodic Reachability with World
Models [68.91647544080097]
In this paper, we introduce a new intrinsic reward design called GoBI - Go Beyond Imagination.
We apply learned world models to generate predicted future states with random actions.
Our method greatly outperforms previous state-of-the-art methods on 12 of the most challenging Minigrid navigation tasks.
arXiv Detail & Related papers (2023-08-25T20:30:20Z) - Cobalt: Optimizing Mining Rewards in Proof-of-Work Network Games [6.052883613180156]
A key factor affecting mining rewards earned is the connectivity between miners in the peer-to-peer network.
We formulate the problem of deciding whom to connect to for miners as a bandit problem.
A key contribution of our work is the use of network coordinates based model for learning the network structure within the bandit algorithm.
arXiv Detail & Related papers (2023-07-10T16:50:58Z) - BC-IRL: Learning Generalizable Reward Functions from Demonstrations [51.535870379280155]
inverse reinforcement learning method learns reward functions that generalize better when compared to maximum-entropy IRL approaches.
We show that BC-IRL learns rewards that generalize better on an illustrative simple task and two continuous robotic control tasks, achieving over twice the success rate of baselines in challenging generalization settings.
arXiv Detail & Related papers (2023-03-28T17:57:20Z) - Partial Selfish Mining for More Profits [21.636578888742477]
Mining attacks aim to gain an unfair share of extra rewards in the blockchain mining.
In this paper, we propose a new and feasible Partial Selfish Mining (PSM) attack.
We show that PSM attackers can be more profitable than selfish miners under a certain range of mining power and network conditions.
arXiv Detail & Related papers (2022-07-27T11:58:38Z) - A Collaboration Strategy in the Mining Pool for
Proof-of-Neural-Architecture Consensus [16.372941299296652]
In most popular public accessible cryptocurrency systems, the mining pool plays a key role because mining cryptocurrency with the mining pool turns the non-profitable situation into profitable for individual miners.
In many recent novel blockchain consensuses, the deep learning training procedure becomes the task for miners to prove their workload.
While the incentive of miners is to earn tokens, individual miners are motivated to join mining pools to become more competitive.
arXiv Detail & Related papers (2022-05-05T17:08:02Z) - Equilibrium of Blockchain Miners with Dynamic Asset Allocation [4.030037871304249]
We model and analyze blockchain miners who seek to maximize the compound return of their mining businesses.
The cost of mining determines the share of each miner or mining pool at equilibrium.
We conclude that neither miners nor mining pools who seek to maximize their compound return will have a financial incentive to occupy more than 50% of the hash rate.
arXiv Detail & Related papers (2020-06-14T20:52:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.