Enhancing Identification of Structure Function of Academic Articles
Using Contextual Information
- URL: http://arxiv.org/abs/2111.14110v2
- Date: Thu, 2 Dec 2021 04:19:34 GMT
- Title: Enhancing Identification of Structure Function of Academic Articles
Using Contextual Information
- Authors: Bowen Ma, Chengzhi Zhang, Yuzhuo Wang, Sanhong Deng
- Abstract summary: This paper takes articles of the ACL conference as the corpus to identify the structure function of academic articles.
We employ the traditional machine learning models and deep learning models to construct the classifiers based on various feature input.
Inspired by (2), this paper introduces contextual information into the deep learning models and achieved significant results.
- Score: 6.28532577139029
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the enrichment of literature resources, researchers are facing the
growing problem of information explosion and knowledge overload. To help
scholars retrieve literature and acquire knowledge successfully, clarifying the
semantic structure of the content in academic literature has become the
essential research question. In the research on identifying the structure
function of chapters in academic articles, only a few studies used the deep
learning model and explored the optimization for feature input. This limits the
application, optimization potential of deep learning models for the research
task. This paper took articles of the ACL conference as the corpus. We employ
the traditional machine learning models and deep learning models to construct
the classifiers based on various feature input. Experimental results show that
(1) Compared with the chapter content, the chapter title is more conducive to
identifying the structure function of academic articles. (2) Relative position
is a valuable feature for building traditional models. (3) Inspired by (2),
this paper further introduces contextual information into the deep learning
models and achieved significant results. Meanwhile, our models show good
migration ability in the open test containing 200 sampled non-training samples.
We also annotated the ACL main conference papers in recent five years based on
the best practice performing models and performed a time series analysis of the
overall corpus. This work explores and summarizes the practical features and
models for this task through multiple comparative experiments and provides a
reference for related text classification tasks. Finally, we indicate the
limitations and shortcomings of the current model and the direction of further
optimization.
Related papers
- Are Large Language Models Good Classifiers? A Study on Edit Intent Classification in Scientific Document Revisions [62.12545440385489]
Large language models (LLMs) have brought substantial advancements in text generation, but their potential for enhancing classification tasks remains underexplored.
We propose a framework for thoroughly investigating fine-tuning LLMs for classification, including both generation- and encoding-based approaches.
We instantiate this framework in edit intent classification (EIC), a challenging and underexplored classification task.
arXiv Detail & Related papers (2024-10-02T20:48:28Z) - Why do you cite? An investigation on citation intents and decision-making classification processes [1.7812428873698407]
This study emphasizes the importance of trustfully classifying citation intents.
We present a study utilizing advanced Ensemble Strategies for Citation Intent Classification (CIC)
One of our models sets as a new state-of-the-art (SOTA) with an 89.46% Macro-F1 score on the SciCite benchmark.
arXiv Detail & Related papers (2024-07-18T09:29:33Z) - Retrieval-Enhanced Machine Learning: Synthesis and Opportunities [60.34182805429511]
Retrieval-enhancement can be extended to a broader spectrum of machine learning (ML)
This work introduces a formal framework of this paradigm, Retrieval-Enhanced Machine Learning (REML), by synthesizing the literature in various domains in ML with consistent notations which is missing from the current literature.
The goal of this work is to equip researchers across various disciplines with a comprehensive, formally structured framework of retrieval-enhanced models, thereby fostering interdisciplinary future research.
arXiv Detail & Related papers (2024-07-17T20:01:21Z) - Learn From Model Beyond Fine-Tuning: A Survey [78.80920533793595]
Learn From Model (LFM) focuses on the research, modification, and design of foundation models (FM) based on the model interface.
The study of LFM techniques can be broadly categorized into five major areas: model tuning, model distillation, model reuse, meta learning and model editing.
This paper gives a comprehensive review of the current methods based on FM from the perspective of LFM.
arXiv Detail & Related papers (2023-10-12T10:20:36Z) - A Survey on Few-Shot Class-Incremental Learning [11.68962265057818]
Few-shot class-incremental learning (FSCIL) poses a significant challenge for deep neural networks to learn new tasks.
This paper provides a comprehensive survey on FSCIL.
FSCIL has achieved impressive achievements in various fields of computer vision.
arXiv Detail & Related papers (2023-04-17T10:15:08Z) - Algorithmic Ghost in the Research Shell: Large Language Models and
Academic Knowledge Creation in Management Research [0.0]
The paper looks at the role of large language models in academic knowledge creation.
This includes writing, editing, reviewing, dataset creation and curation.
arXiv Detail & Related papers (2023-03-10T14:25:29Z) - What Makes Good Contrastive Learning on Small-Scale Wearable-based
Tasks? [59.51457877578138]
We study contrastive learning on the wearable-based activity recognition task.
This paper presents an open-source PyTorch library textttCL-HAR, which can serve as a practical tool for researchers.
arXiv Detail & Related papers (2022-02-12T06:10:15Z) - Neural Ranking Models for Document Retrieval [11.886543741028127]
Several approaches to ranking are based on traditional machine learning algorithms using a set of hand-crafted features.
Deep learning models are trained end-to-end to extract features from the raw data for ranking tasks.
A variety of deep learning models have been proposed, and each model presents a set of neural network components to extract features that are used for ranking.
arXiv Detail & Related papers (2021-02-23T19:30:37Z) - Method and Dataset Entity Mining in Scientific Literature: A CNN +
Bi-LSTM Model with Self-attention [21.93889297841459]
We propose a novel entity recognition model, called MDER, which is able to effectively extract the method and dataset entities from scientific papers.
We evaluate the proposed model on datasets constructed from the published papers of four research areas in computer science, i.e., NLP, CV, Data Mining and AI.
arXiv Detail & Related papers (2020-10-26T13:38:43Z) - A Survey on Text Classification: From Shallow to Deep Learning [83.47804123133719]
The last decade has seen a surge of research in this area due to the unprecedented success of deep learning.
This paper fills the gap by reviewing the state-of-the-art approaches from 1961 to 2021.
We create a taxonomy for text classification according to the text involved and the models used for feature extraction and classification.
arXiv Detail & Related papers (2020-08-02T00:09:03Z) - Rethinking Generalization of Neural Models: A Named Entity Recognition
Case Study [81.11161697133095]
We take the NER task as a testbed to analyze the generalization behavior of existing models from different perspectives.
Experiments with in-depth analyses diagnose the bottleneck of existing neural NER models.
As a by-product of this paper, we have open-sourced a project that involves a comprehensive summary of recent NER papers.
arXiv Detail & Related papers (2020-01-12T04:33:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.