Comparison Study Between Token Classification and Sequence
Classification In Text Classification
- URL: http://arxiv.org/abs/2211.13899v1
- Date: Fri, 25 Nov 2022 05:14:58 GMT
- Title: Comparison Study Between Token Classification and Sequence
Classification In Text Classification
- Authors: Amir Jafari
- Abstract summary: Unsupervised Machine Learning techniques have been applied to Natural Language Processing tasks and surpasses the benchmarks such as GLUE with great success.
Building language models approach good results in one language and it can be applied to multiple NLP task such as classification, summarization, generation and etc as out of box models.
- Score: 0.45687771576879593
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unsupervised Machine Learning techniques have been applied to Natural
Language Processing tasks and surpasses the benchmarks such as GLUE with great
success. Building language models approach achieves good results in one
language and it can be applied to multiple NLP task such as classification,
summarization, generation and etc as an out of box model. Among all the of the
classical approaches used in NLP, the masked language modeling is the most
used. In general, the only requirement to build a language model is presence of
the large corpus of textual data. Text classification engines uses a variety of
models from classical and state of art transformer models to classify texts for
in order to save costs. Sequence Classifiers are mostly used in the domain of
text classification. However Token classifiers also are viable candidate models
as well. Sequence Classifiers and Token Classifier both tend to improve the
classification predictions due to the capturing the context information
differently. This work aims to compare the performance of Sequence Classifier
and Token Classifiers and evaluate each model on the same set of data. In this
work, we are using a pre-trained model as the base model and Token Classifier
and Sequence Classier heads results of these two scoring paradigms with be
compared..
Related papers
- Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - Learning Mutually Informed Representations for Characters and Subwords [26.189422354038978]
We introduce the entanglement model, aiming to combine character and subword language models.
Inspired by vision-language models, our model treats characters and subwords as separate modalities.
We evaluate our model on text classification, named entity recognition, POS-tagging, and character-level sequence labeling.
arXiv Detail & Related papers (2023-11-14T02:09:10Z) - T3L: Translate-and-Test Transfer Learning for Cross-Lingual Text
Classification [50.675552118811]
Cross-lingual text classification is typically built on large-scale, multilingual language models (LMs) pretrained on a variety of languages of interest.
We propose revisiting the classic "translate-and-test" pipeline to neatly separate the translation and classification stages.
arXiv Detail & Related papers (2023-06-08T07:33:22Z) - Anomaly Detection using Ensemble Classification and Evidence Theory [62.997667081978825]
We present a novel approach for novel detection using ensemble classification and evidence theory.
A pool selection strategy is presented to build a solid ensemble classifier.
We use uncertainty for the anomaly detection approach.
arXiv Detail & Related papers (2022-12-23T00:50:41Z) - TabLLM: Few-shot Classification of Tabular Data with Large Language
Models [66.03023402174138]
We study the application of large language models to zero-shot and few-shot classification.
We evaluate several serialization methods including templates, table-to-text models, and large language models.
This approach is also competitive with strong traditional baselines like gradient-boosted trees.
arXiv Detail & Related papers (2022-10-19T17:08:13Z) - Perturbations and Subpopulations for Testing Robustness in Token-Based
Argument Unit Recognition [6.502694770864571]
Argument Unit Recognition and Classification aims at identifying argument units from text and classifying them as pro or against.
One of the design choices that need to be made when developing systems for this task is what the unit of classification should be: segments of tokens or full sentences.
Previous research suggests that fine-tuning language models on the token-level yields more robust results for classifying sentences compared to training on sentences directly.
We reproduce the study that originally made this claim and further investigate what exactly token-based systems learned better compared to sentence-based ones.
arXiv Detail & Related papers (2022-09-29T13:44:28Z) - DIRECTOR: Generator-Classifiers For Supervised Language Modeling [27.86870968048833]
Current language models achieve low perplexity but their resulting generations still suffer from toxic responses, repetitiveness and contradictions.
We introduce a new architecture, sc Director, that consists of a unified generator-classifier with both a language modeling and a classification head for each output token.
arXiv Detail & Related papers (2022-06-15T17:44:08Z) - Learning and Evaluating Representations for Deep One-class
Classification [59.095144932794646]
We present a two-stage framework for deep one-class classification.
We first learn self-supervised representations from one-class data, and then build one-class classifiers on learned representations.
In experiments, we demonstrate state-of-the-art performance on visual domain one-class classification benchmarks.
arXiv Detail & Related papers (2020-11-04T23:33:41Z) - Interpretable Sequence Classification via Discrete Optimization [26.899228003677138]
In many applications such as healthcare monitoring or intrusion detection, early classification is crucial to prompt intervention.
In this work, we learn sequence classifiers that favour early classification from an evolving observation trace.
Our classifiers are interpretable---supporting explanation, counterfactual reasoning, and human-in-the-loop modification.
arXiv Detail & Related papers (2020-10-06T15:31:07Z) - Document Ranking with a Pretrained Sequence-to-Sequence Model [56.44269917346376]
We show how a sequence-to-sequence model can be trained to generate relevance labels as "target words"
Our approach significantly outperforms an encoder-only model in a data-poor regime.
arXiv Detail & Related papers (2020-03-14T22:29:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.