Word2Vec: Optimal Hyper-Parameters and Their Impact on NLP Downstream
Tasks
- URL: http://arxiv.org/abs/2003.11645v3
- Date: Sat, 17 Apr 2021 06:02:44 GMT
- Title: Word2Vec: Optimal Hyper-Parameters and Their Impact on NLP Downstream
Tasks
- Authors: Tosin P. Adewumi, Foteini Liwicki and Marcus Liwicki
- Abstract summary: We show optimal combination of hyper- parameters exists and evaluate various combinations.
We obtain better human-assigned WordSim scores, corresponding Spearman correlation and better downstream performances compared to the original model.
- Score: 1.6507910904669727
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Word2Vec is a prominent model for natural language processing (NLP) tasks.
Similar inspiration is found in distributed embeddings for new state-of-the-art
(SotA) deep neural networks. However, wrong combination of hyper-parameters can
produce poor quality vectors. The objective of this work is to empirically show
optimal combination of hyper-parameters exists and evaluate various
combinations. We compare them with the released, pre-trained original word2vec
model. Both intrinsic and extrinsic (downstream) evaluations, including named
entity recognition (NER) and sentiment analysis (SA) were carried out. The
downstream tasks reveal that the best model is usually task-specific, high
analogy scores don't necessarily correlate positively with F1 scores and the
same applies to focus on data alone. Increasing vector dimension size after a
point leads to poor quality or performance. If ethical considerations to save
time, energy and the environment are made, then reasonably smaller corpora may
do just as well or even better in some cases. Besides, using a small corpus, we
obtain better human-assigned WordSim scores, corresponding Spearman correlation
and better downstream performances (with significance tests) compared to the
original model, trained on 100 billion-word corpus.
Related papers
- Discriminative versus Generative Approaches to Simulation-based Inference [0.19999259391104385]
Deep learning has enabled unbinned and high-dimensional parameter estimation.
We compare two approaches for neural simulation-based inference (N SBI)
We find that both the direct likelihood and likelihood ratio estimation are able to effectively extract parameters with reasonable uncertainties.
arXiv Detail & Related papers (2025-03-11T01:38:54Z) - Revisiting the Evaluation of Image Synthesis with GANs [55.72247435112475]
This study presents an empirical investigation into the evaluation of synthesis performance, with generative adversarial networks (GANs) as a representative of generative models.
In particular, we make in-depth analyses of various factors, including how to represent a data point in the representation space, how to calculate a fair distance using selected samples, and how many instances to use from each set.
arXiv Detail & Related papers (2023-04-04T17:54:32Z) - Cross-Model Comparative Loss for Enhancing Neuronal Utility in Language
Understanding [82.46024259137823]
We propose a cross-model comparative loss for a broad range of tasks.
We demonstrate the universal effectiveness of comparative loss through extensive experiments on 14 datasets from 3 distinct NLU tasks.
arXiv Detail & Related papers (2023-01-10T03:04:27Z) - A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive
Coding Networks [65.34977803841007]
Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience.
We show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one.
arXiv Detail & Related papers (2022-11-16T00:11:04Z) - Embarrassingly Simple Performance Prediction for Abductive Natural
Language Inference [10.536415845097661]
We propose a method for predicting the performance of NLI models without fine-tuning them.
We show that the accuracy of the cosine similarity approach correlates strongly with the accuracy of the classification approach with a Pearson correlation coefficient of 0.65.
Our method can lead to significant time savings in the process of model selection.
arXiv Detail & Related papers (2022-02-21T18:10:24Z) - MoEfication: Conditional Computation of Transformer Models for Efficient
Inference [66.56994436947441]
Transformer-based pre-trained language models can achieve superior performance on most NLP tasks due to large parameter capacity, but also lead to huge computation cost.
We explore to accelerate large-model inference by conditional computation based on the sparse activation phenomenon.
We propose to transform a large model into its mixture-of-experts (MoE) version with equal model size, namely MoEfication.
arXiv Detail & Related papers (2021-10-05T02:14:38Z) - COM2SENSE: A Commonsense Reasoning Benchmark with Complementary
Sentences [21.11065466376105]
Commonsense reasoning is intuitive for humans but has been a long-term challenge for artificial intelligence (AI)
Recent advancements in pretrained language models have shown promising results on several commonsense benchmark datasets.
We introduce a new commonsense reasoning benchmark dataset comprising natural language true/false statements.
arXiv Detail & Related papers (2021-06-02T06:31:55Z) - Learning from Mistakes: Combining Ontologies via Self-Training for
Dialogue Generation [6.221019624345408]
Natural language generators (NLGs) for task-oriented dialogue typically take a meaning representation (MR) as input.
We create a new, larger combined ontology, and then train an NLG to produce utterances covering it.
For example, if one dataset has attributes for family-friendly and rating information, and the other has attributes for decor and service, our aim is an NLG for the combined ontology that can produce utterances that realize values for family-friendly, rating, decor and service.
arXiv Detail & Related papers (2020-09-30T23:54:38Z) - Adaptive Name Entity Recognition under Highly Unbalanced Data [5.575448433529451]
We present our experiments on a neural architecture composed of a Conditional Random Field (CRF) layer stacked on top of a Bi-directional LSTM (BI-LSTM) layer for solving NER tasks.
We introduce an add-on classification model to split sentences into two different sets: Weak and Strong classes and then designing a couple of Bi-LSTM-CRF models properly to optimize performance on each set.
arXiv Detail & Related papers (2020-03-10T06:56:52Z) - On the Discrepancy between Density Estimation and Sequence Generation [92.70116082182076]
log-likelihood is highly correlated with BLEU when we consider models within the same family.
We observe no correlation between rankings of models across different families.
arXiv Detail & Related papers (2020-02-17T20:13:35Z) - Parameter Space Factorization for Zero-Shot Learning across Tasks and
Languages [112.65994041398481]
We propose a Bayesian generative model for the space of neural parameters.
We infer the posteriors over such latent variables based on data from seen task-language combinations.
Our model yields comparable or better results than state-of-the-art, zero-shot cross-lingual transfer methods.
arXiv Detail & Related papers (2020-01-30T16:58:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.