Meta-Embeddings for Natural Language Inference and Semantic Similarity
tasks
- URL: http://arxiv.org/abs/2012.00633v1
- Date: Tue, 1 Dec 2020 16:58:01 GMT
- Title: Meta-Embeddings for Natural Language Inference and Semantic Similarity
tasks
- Authors: Shree Charran R, Rahul Kumar Dubey (Senior Member IEEE)
- Abstract summary: Word Representations form the core component for almost all advanced Natural Language Processing (NLP) applications.
In this paper, we propose to use Meta Embedding derived from few State-of-the-Art (SOTA) models to efficiently tackle mainstream NLP tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Word Representations form the core component for almost all advanced Natural
Language Processing (NLP) applications such as text mining, question-answering,
and text summarization, etc. Over the last two decades, immense research is
conducted to come up with one single model to solve all major NLP tasks. The
major problem currently is that there are a plethora of choices for different
NLP tasks. Thus for NLP practitioners, the task of choosing the right model to
be used itself becomes a challenge. Thus combining multiple pre-trained word
embeddings and forming meta embeddings has become a viable approach to improve
tackle NLP tasks. Meta embedding learning is a process of producing a single
word embedding from a given set of pre-trained input word embeddings. In this
paper, we propose to use Meta Embedding derived from few State-of-the-Art
(SOTA) models to efficiently tackle mainstream NLP tasks like classification,
semantic relatedness, and text similarity. We have compared both ensemble and
dynamic variants to identify an efficient approach. The results obtained show
that even the best State-of-the-Art models can be bettered. Thus showing us
that meta-embeddings can be used for several NLP tasks by harnessing the power
of several individual representations.
Related papers
- A Novel Cartography-Based Curriculum Learning Method Applied on RoNLI: The First Romanian Natural Language Inference Corpus [71.77214818319054]
Natural language inference is a proxy for natural language understanding.
There is no publicly available NLI corpus for the Romanian language.
We introduce the first Romanian NLI corpus (RoNLI) comprising 58K training sentence pairs.
arXiv Detail & Related papers (2024-05-20T08:41:15Z) - Meta-Task Prompting Elicits Embeddings from Large Language Models [54.757445048329735]
We introduce a new unsupervised text embedding method, Meta-Task Prompting with Explicit One-Word Limitation.
We generate high-quality sentence embeddings from Large Language Models without the need for model fine-tuning.
Our findings suggest a new scaling law, offering a versatile and resource-efficient approach for embedding generation across diverse scenarios.
arXiv Detail & Related papers (2024-02-28T16:35:52Z) - Number Entity Recognition [65.80137628972312]
Numbers are essential components of text, like any other word tokens, from which natural language processing (NLP) models are built and deployed.
In this work, we attempt to tap this potential of state-of-the-art NLP models and transfer their ability to boost performance in related tasks.
Our proposed classification of numbers into entities helps NLP models perform well on several tasks, including a handcrafted Fill-In-The-Blank (FITB) task and on question answering using joint embeddings.
arXiv Detail & Related papers (2022-05-07T05:22:43Z) - Meta Learning for Natural Language Processing: A Survey [88.58260839196019]
Deep learning has been the mainstream technique in natural language processing (NLP) area.
Deep learning requires many labeled data and is less generalizable across domains.
Meta-learning is an arising field in machine learning studying approaches to learn better algorithms.
arXiv Detail & Related papers (2022-05-03T13:58:38Z) - Unsupervised Attention-based Sentence-Level Meta-Embeddings from
Contextualised Language Models [15.900069711477542]
We propose a sentence-level meta-embedding learning method that takes independently trained contextualised word embedding models.
Our proposed method is unsupervised and is not tied to a particular downstream task.
Experimental results show that our proposed unsupervised sentence-level meta-embedding method outperforms previously proposed sentence-level meta-embedding methods.
arXiv Detail & Related papers (2022-04-16T08:20:24Z) - Grad2Task: Improved Few-shot Text Classification Using Gradients for
Task Representation [24.488427641442694]
We propose a novel conditional neural process-based approach for few-shot text classification.
Our key idea is to represent each task using gradient information from a base model.
Our approach outperforms traditional fine-tuning, sequential transfer learning, and state-of-the-art meta learning approaches.
arXiv Detail & Related papers (2022-01-27T15:29:30Z) - Between words and characters: A Brief History of Open-Vocabulary
Modeling and Tokenization in NLP [22.772546707304766]
We show how hybrid approaches of words and characters as well as subword-based approaches based on learned segmentation have been proposed and evaluated.
We conclude that there is and likely will never be a silver bullet singular solution for all applications.
arXiv Detail & Related papers (2021-12-20T13:04:18Z) - Paradigm Shift in Natural Language Processing [66.62609175829816]
In the era of deep learning, modeling for most NLP tasks has converged to several mainstream paradigms.
Recent years have observed a rising trend of Paradigm Shift, which is solving one NLP task by reformulating it as another one.
Some of these paradigms have shown great potential to unify a large number of NLP tasks, making it possible to build a single model to handle diverse tasks.
arXiv Detail & Related papers (2021-09-26T11:55:23Z) - Deconstructing word embedding algorithms [17.797952730495453]
We propose a retrospective on some of the most well-known word embedding algorithms.
In this work, we deconstruct Word2vec, GloVe, and others, into a common form, unveiling some of the common conditions that seem to be required for making performant word embeddings.
arXiv Detail & Related papers (2020-11-12T14:23:35Z) - Multiple Word Embeddings for Increased Diversity of Representation [15.279850826041066]
We show a technique that substantially and consistently improves performance over a strong baseline with negligible increase in run time.
We analyze aspects of pre-trained embedding similarity and vocabulary coverage and find that the representational diversity is the driving force of why this technique works.
arXiv Detail & Related papers (2020-09-30T02:33:09Z) - Coreferential Reasoning Learning for Language Representation [88.14248323659267]
We present CorefBERT, a novel language representation model that can capture the coreferential relations in context.
The experimental results show that, compared with existing baseline models, CorefBERT can achieve significant improvements consistently on various downstream NLP tasks.
arXiv Detail & Related papers (2020-04-15T03:57:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.