Maximizing NFT Incentives: References Make You Rich
- URL: http://arxiv.org/abs/2402.06459v1
- Date: Fri, 9 Feb 2024 15:04:16 GMT
- Title: Maximizing NFT Incentives: References Make You Rich
- Authors: Guangsheng Yu, Qin Wang, Caijun Sun, Lam Duc Nguyen, H.M.N. Dilum
Bandara, Shiping Chen
- Abstract summary: Current Non-Fungible Token (NFT) incentive mechanisms tend to overlook their potential for scalable organizational structures.
We propose, analyze, and implement a novel reference incentive model, which is inherently structured as a Directed Acyclic Graph (DAG)-based NFT network.
- Score: 3.943871561481494
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we study how to optimize existing Non-Fungible Token (NFT)
incentives. Upon exploring a large number of NFT-related standards and
real-world projects, we come across an unexpected finding. That is, the current
NFT incentive mechanisms, often organized in an isolated and one-time-use
fashion, tend to overlook their potential for scalable organizational
structures.
We propose, analyze, and implement a novel reference incentive model, which
is inherently structured as a Directed Acyclic Graph (DAG)-based NFT network.
This model aims to maximize connections (or references) between NFTs, enabling
each isolated NFT to expand its network and accumulate rewards derived from
subsequent or subscribed ones. We conduct both theoretical and practical
analyses of the model, demonstrating its optimal utility.
Related papers
- Joint Diffusion Processes as an Inductive Bias in Sheaf Neural Networks [14.224234978509026]
Sheaf Neural Networks (SNNs) naturally extend Graph Neural Networks (GNNs)
We propose two novel sheaf learning approaches that provide a more intuitive understanding of the involved structure maps.
In our evaluation, we show the limitations of the real-world benchmarks used so far on SNNs.
arXiv Detail & Related papers (2024-07-30T07:17:46Z) - NFTVis: Visual Analysis of NFT Performance [12.491701063977825]
A non-fungible token (NFT) is a data unit stored on the blockchain.
Current rarity models have flaws and are sometimes not convincing.
It is difficult to take comprehensive consideration and analyze NFT performance efficiently.
arXiv Detail & Related papers (2023-06-05T09:02:48Z) - A Survey on Deep Learning based Time Series Analysis with Frequency
Transformation [74.3919960186696]
Frequency transformation (FT) has been increasingly incorporated into deep learning models to enhance state-of-the-art accuracy and efficiency in time series analysis.
Despite the growing attention and the proliferation of research in this emerging field, there is currently a lack of a systematic review and in-depth analysis of deep learning-based time series models with FT.
We present a comprehensive review that systematically investigates and summarizes the recent research advancements in deep learning-based time series analysis with FT.
arXiv Detail & Related papers (2023-02-04T14:33:07Z) - Probably Something: A Multi-Layer Taxonomy of Non-Fungible Tokens [62.997667081978825]
Non-Fungible Tokens (NFTs) are hyped and increasingly marketed as essential building blocks of the Metaverse.
This paper aims to establish a fundamental and comprehensive understanding of NFTs by identifying and structuring common characteristics within a taxonomy.
arXiv Detail & Related papers (2022-08-29T18:00:30Z) - SymNMF-Net for The Symmetric NMF Problem [62.44067422984995]
We propose a neural network called SymNMF-Net for the Symmetric NMF problem.
We show that the inference of each block corresponds to a single iteration of the optimization.
Empirical results on real-world datasets demonstrate the superiority of our SymNMF-Net.
arXiv Detail & Related papers (2022-05-26T08:17:39Z) - An Analysis of the Features Considerable for NFT Recommendations [0.0]
This research explores the methods that NFTs can be recommended to people who interact with NFT-marketplaces.
The outcome highlights the necessity of using multiple Recommender Systems to present the user with the best possible NFTs when interacting with decentralized systems.
arXiv Detail & Related papers (2022-05-01T12:11:17Z) - Neural Structured Prediction for Inductive Node Classification [29.908759584092167]
This paper studies node classification in the inductive setting, aiming to learn a model on labeled training graphs and generalize it to infer node labels on unlabeled test graphs.
We present a new approach called the Structured Proxy Network (SPN), which combines the advantages of both worlds.
arXiv Detail & Related papers (2022-04-15T15:50:27Z) - EIGNN: Efficient Infinite-Depth Graph Neural Networks [51.97361378423152]
Graph neural networks (GNNs) are widely used for modelling graph-structured data in numerous applications.
Motivated by this limitation, we propose a GNN model with infinite depth, which we call Efficient Infinite-Depth Graph Neural Networks (EIGNN)
We show that EIGNN has a better ability to capture long-range dependencies than recent baselines, and consistently achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T08:16:58Z) - Under the Skin of Foundation NFT Auctions [0.0]
We studied one of the most prominent marketplaces dedicated to NFT auctions and trades, Foundation.
We performed social network analysis on a graph that we had created based on transferred NFTs on Foundation.
We built a neural network-based similarity model for retrieving and clustering similar NFTs.
arXiv Detail & Related papers (2021-09-25T09:01:44Z) - On Connections between Regularizations for Improving DNN Robustness [67.28077776415724]
This paper analyzes regularization terms proposed recently for improving the adversarial robustness of deep neural networks (DNNs)
We study possible connections between several effective methods, including input-gradient regularization, Jacobian regularization, curvature regularization, and a cross-Lipschitz functional.
arXiv Detail & Related papers (2020-07-04T23:43:32Z) - Flexible Transmitter Network [84.90891046882213]
Current neural networks are mostly built upon the MP model, which usually formulates the neuron as executing an activation function on the real-valued weighted aggregation of signals received from other neurons.
We propose the Flexible Transmitter (FT) model, a novel bio-plausible neuron model with flexible synaptic plasticity.
We present the Flexible Transmitter Network (FTNet), which is built on the most common fully-connected feed-forward architecture.
arXiv Detail & Related papers (2020-04-08T06:55:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.