Digital Asset Valuation: A Study on Domain Names, Email Addresses, and
NFTs
- URL: http://arxiv.org/abs/2210.10637v1
- Date: Thu, 6 Oct 2022 12:59:06 GMT
- Title: Digital Asset Valuation: A Study on Domain Names, Email Addresses, and
NFTs
- Authors: Kai Sun
- Abstract summary: DASH is the first Digital Asset Sales History dataset that features multiple digital asset classes spanning from classical to blockchain-based ones.
We build strong conventional feature-based models as the baselines for DASH.
We explore deep learning models based on fine-tuning pre-trained language models, which have not yet been explored in the previous literature.
- Score: 4.16271611433618
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing works on valuing digital assets on the Internet typically focus on a
single asset class. To promote the development of automated valuation
techniques, preferably those that are generally applicable to multiple asset
classes, we construct DASH, the first Digital Asset Sales History dataset that
features multiple digital asset classes spanning from classical to
blockchain-based ones. Consisting of 280K transactions of domain names
(DASH_DN), email addresses (DASH_EA), and non-fungible token (NFT)-based
identifiers (DASH_NFT), such as Ethereum Name Service names, DASH advances the
field in several aspects: the subsets DASH_DN, DASH_EA, and DASH_NFT are the
largest freely accessible domain name transaction dataset, the only publicly
available email address transaction dataset, and the first NFT transaction
dataset that focuses on identifiers, respectively.
We build strong conventional feature-based models as the baselines for DASH.
We next explore deep learning models based on fine-tuning pre-trained language
models, which have not yet been explored for digital asset valuation in the
previous literature. We find that the vanilla fine-tuned model already performs
reasonably well, outperforming all but the best-performing baselines. We
further propose improvements to make the model more aware of the time
sensitivity of transactions and the popularity of assets. Experimental results
show that our improved model consistently outperforms all the other models
across all asset classes on DASH.
Related papers
- Improving Few-Shot Cross-Domain Named Entity Recognition by Instruction Tuning a Word-Embedding based Retrieval Augmented Large Language Model [0.0]
Few-Shot Cross-Domain NER is a process of leveraging knowledge from data-rich source domains to perform entity recognition on data scarce target domains.
We propose IF-WRANER, a retrieval augmented large language model for Named Entity Recognition.
arXiv Detail & Related papers (2024-11-01T08:57:29Z) - iNeMo: Incremental Neural Mesh Models for Robust Class-Incremental Learning [22.14627083675405]
We propose incremental neural mesh models that can be extended with new meshes over time.
We demonstrate the effectiveness of our method through extensive experiments on the Pascal3D and ObjectNet3D datasets.
Our work also presents the first incremental learning approach for pose estimation.
arXiv Detail & Related papers (2024-07-12T13:57:49Z) - ID-centric Pre-training for Recommendation [51.72177873832969]
ID embeddings are challenging to be transferred to new domains.
behavioral information in ID embeddings is still verified to be dominating in PLM-based recommendation models.
We propose a novel ID-centric recommendation pre-training paradigm (IDP), which directly transfers informative ID embeddings learned in pre-training domains to item representations in new domains.
arXiv Detail & Related papers (2024-05-06T15:34:31Z) - Query of CC: Unearthing Large Scale Domain-Specific Knowledge from
Public Corpora [104.16648246740543]
We propose an efficient data collection method based on large language models.
The method bootstraps seed information through a large language model and retrieves related data from public corpora.
It not only collects knowledge-related data for specific domains but unearths the data with potential reasoning procedures.
arXiv Detail & Related papers (2024-01-26T03:38:23Z) - Towards a Foundation Purchasing Model: Pretrained Generative
Autoregression on Transaction Sequences [0.0]
We present a generative pretraining method that can be used to obtain contextualised embeddings of financial transactions.
We additionally perform large-scale pretraining of an embedding model using a corpus of data from 180 issuing banks containing 5.1 billion transactions.
arXiv Detail & Related papers (2024-01-03T09:32:48Z) - PILOT: A Pre-Trained Model-Based Continual Learning Toolbox [71.63186089279218]
This paper introduces a pre-trained model-based continual learning toolbox known as PILOT.
On the one hand, PILOT implements some state-of-the-art class-incremental learning algorithms based on pre-trained models, such as L2P, DualPrompt, and CODA-Prompt.
On the other hand, PILOT fits typical class-incremental learning algorithms within the context of pre-trained models to evaluate their effectiveness.
arXiv Detail & Related papers (2023-09-13T17:55:11Z) - Data-Free Adversarial Knowledge Distillation for Graph Neural Networks [62.71646916191515]
We propose the first end-to-end framework for data-free adversarial knowledge distillation on graph structured data (DFAD-GNN)
To be specific, our DFAD-GNN employs a generative adversarial network, which mainly consists of three components: a pre-trained teacher model and a student model are regarded as two discriminators, and a generator is utilized for deriving training graphs to distill knowledge from the teacher model into the student model.
Our DFAD-GNN significantly surpasses state-of-the-art data-free baselines in the graph classification task.
arXiv Detail & Related papers (2022-05-08T08:19:40Z) - Named Entity Normalization Model Using Edge Weight Updating Neural
Network: Assimilation Between Knowledge-Driven Graph and Data-Driven Graph [7.873525968415584]
We build the named entity normalization model with a novel Edge Weight Updating Neural Network.
Our proposed model when tested on four different datasets achieved state-of-the-art results.
arXiv Detail & Related papers (2021-06-14T16:14:58Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - MELINDA: A Multimodal Dataset for Biomedical Experiment Method
Classification [14.820951153262685]
We introduce a new dataset, MELINDA, for Multimodal biomEdicaL experImeNt methoD clAssification.
The dataset is collected in a fully automated distant supervision manner, where the labels are obtained from an existing curated database.
We benchmark various state-of-the-art NLP and computer vision models, including unimodal models which only take either caption texts or images as inputs.
arXiv Detail & Related papers (2020-12-16T19:11:36Z) - Zero-Resource Cross-Domain Named Entity Recognition [68.83177074227598]
Existing models for cross-domain named entity recognition rely on numerous unlabeled corpus or labeled NER training data in target domains.
We propose a cross-domain NER model that does not use any external resources.
arXiv Detail & Related papers (2020-02-14T09:04:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.