A Survey on Recent Advances in Sequence Labeling from Deep Learning
Models
- URL: http://arxiv.org/abs/2011.06727v1
- Date: Fri, 13 Nov 2020 02:29:50 GMT
- Title: A Survey on Recent Advances in Sequence Labeling from Deep Learning
Models
- Authors: Zhiyong He, Zanbo Wang, Wei Wei, Shanshan Feng, Xianling Mao, and
Sheng Jiang
- Abstract summary: Sequence labeling is a fundamental research problem encompassing a variety of tasks.
Deep learning has been employed for sequence labeling tasks due to its powerful capability in automatically learning complex features.
- Score: 19.753741555478793
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequence labeling (SL) is a fundamental research problem encompassing a
variety of tasks, e.g., part-of-speech (POS) tagging, named entity recognition
(NER), text chunking, etc. Though prevalent and effective in many downstream
applications (e.g., information retrieval, question answering, and knowledge
graph embedding), conventional sequence labeling approaches heavily rely on
hand-crafted or language-specific features. Recently, deep learning has been
employed for sequence labeling tasks due to its powerful capability in
automatically learning complex features of instances and effectively yielding
the stat-of-the-art performances. In this paper, we aim to present a
comprehensive review of existing deep learning-based sequence labeling models,
which consists of three related tasks, e.g., part-of-speech tagging, named
entity recognition, and text chunking. Then, we systematically present the
existing approaches base on a scientific taxonomy, as well as the widely-used
experimental datasets and popularly-adopted evaluation metrics in the SL
domain. Furthermore, we also present an in-depth analysis of different SL
models on the factors that may affect the performance and future directions in
the SL domain.
Related papers
- Classifier identification in Ancient Egyptian as a low-resource sequence-labelling task [0.7237827208209208]
Ancient Egyptian (AE) writing system was characterised by widespread use of graphemic classifiers (determinatives)
We implement a series of sequence-labelling neural models, which achieve promising performance despite the modest amount of training data.
We discuss tokenisation and operationalisation issues arising from tackling AE texts and contrast our approach with frequency-based baselines.
arXiv Detail & Related papers (2024-06-29T15:40:25Z) - Combining Supervised Learning and Reinforcement Learning for Multi-Label Classification Tasks with Partial Labels [27.53399899573121]
We propose an RL-based framework combining the exploration ability of reinforcement learning and the exploitation ability of supervised learning.
Experimental results across various tasks, including document-level relation extraction, demonstrate the generalization and effectiveness of our framework.
arXiv Detail & Related papers (2024-06-24T03:36:19Z) - Self-Supervised Skeleton-Based Action Representation Learning: A Benchmark and Beyond [19.074841631219233]
Self-supervised learning (SSL) has been proven effective for skeleton-based action understanding.
In this paper, we conduct a comprehensive survey on self-supervised skeleton-based action representation learning.
arXiv Detail & Related papers (2024-06-05T06:21:54Z) - Harnessing the Power of Beta Scoring in Deep Active Learning for
Multi-Label Text Classification [6.662167018900634]
Our study introduces a novel deep active learning strategy, capitalizing on the Beta family of proper scoring rules within the Expected Loss Reduction framework.
It computes the expected increase in scores using the Beta Scoring Rules, which are then transformed into sample vector representations.
Comprehensive evaluations across both synthetic and real datasets reveal our method's capability to often outperform established acquisition techniques in multi-label text classification.
arXiv Detail & Related papers (2024-01-15T00:06:24Z) - A Survey on Self-supervised Learning: Algorithms, Applications, and Future Trends [82.64268080902742]
Self-supervised learning (SSL) aims to learn discriminative features from unlabeled data without relying on human-annotated labels.
SSL has garnered significant attention recently, leading to the development of numerous related algorithms.
This paper presents a review of diverse SSL methods, encompassing algorithmic aspects, application domains, three key trends, and open research questions.
arXiv Detail & Related papers (2023-01-13T14:41:05Z) - SLUE Phase-2: A Benchmark Suite of Diverse Spoken Language Understanding
Tasks [88.4408774253634]
Spoken language understanding (SLU) tasks have been studied for many decades in the speech research community.
There are not nearly as many SLU task benchmarks, and many of the existing ones use data that is not freely available to all researchers.
Recent work has begun to introduce such benchmark for several tasks.
arXiv Detail & Related papers (2022-12-20T18:39:59Z) - Representation Learning for the Automatic Indexing of Sound Effects
Libraries [79.68916470119743]
We show that a task-specific but dataset-independent representation can successfully address data issues such as class imbalance, inconsistent class labels, and insufficient dataset size.
Detailed experimental results show the impact of metric learning approaches and different cross-dataset training methods on representational effectiveness.
arXiv Detail & Related papers (2022-08-18T23:46:13Z) - Recent Few-Shot Object Detection Algorithms: A Survey with Performance
Comparison [54.357707168883024]
Few-Shot Object Detection (FSOD) mimics the humans' ability of learning to learn.
FSOD intelligently transfers the learned generic object knowledge from the common heavy-tailed, to the novel long-tailed object classes.
We give an overview of FSOD, including the problem definition, common datasets, and evaluation protocols.
arXiv Detail & Related papers (2022-03-27T04:11:28Z) - Self-supervised on Graphs: Contrastive, Generative,or Predictive [25.679620842010422]
Self-supervised learning (SSL) is emerging as a new paradigm for extracting informative knowledge through well-designed pretext tasks.
We divide existing graph SSL methods into three categories: contrastive, generative, and predictive.
We also summarize the commonly used datasets, evaluation metrics, downstream tasks, and open-source implementations of various algorithms.
arXiv Detail & Related papers (2021-05-16T03:30:03Z) - Graph-based Semi-supervised Learning: A Comprehensive Review [51.26862262550445]
Semi-supervised learning (SSL) has tremendous value in practice due to its ability to utilize both labeled data and unlabelled data.
An important class of SSL methods is to naturally represent data as graphs, which corresponds to graph-based semi-supervised learning (GSSL) methods.
GSSL methods have demonstrated their advantages in various domains due to their uniqueness of structure, the universality of applications, and their scalability to large scale data.
arXiv Detail & Related papers (2021-02-26T05:11:09Z) - Adaptive Self-training for Few-shot Neural Sequence Labeling [55.43109437200101]
We develop techniques to address the label scarcity challenge for neural sequence labeling models.
Self-training serves as an effective mechanism to learn from large amounts of unlabeled data.
meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels.
arXiv Detail & Related papers (2020-10-07T22:29:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.