UIUC_BioNLP at SemEval-2021 Task 11: A Cascade of Neural Models for
Structuring Scholarly NLP Contributions
- URL: http://arxiv.org/abs/2105.05435v1
- Date: Wed, 12 May 2021 05:24:35 GMT
- Title: UIUC_BioNLP at SemEval-2021 Task 11: A Cascade of Neural Models for
Structuring Scholarly NLP Contributions
- Authors: Haoyang Liu, M. Janina Sarol and Halil Kilicoglu
- Abstract summary: We propose a cascade of neural models that performs sentence classification, phrase recognition, and triple extraction.
A BERT-CRF model was used to recognize and characterize relevant phrases in contribution sentences.
Our system was officially ranked second in Phase 1 evaluation and first in both parts of Phase 2 evaluation.
- Score: 1.5942130010323128
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a cascade of neural models that performs sentence classification,
phrase recognition, and triple extraction to automatically structure the
scholarly contributions of NLP publications. To identify the most important
contribution sentences in a paper, we used a BERT-based classifier with
positional features (Subtask 1). A BERT-CRF model was used to recognize and
characterize relevant phrases in contribution sentences (Subtask 2). We
categorized the triples into several types based on whether and how their
elements were expressed in text, and addressed each type using separate
BERT-based classifiers as well as rules (Subtask 3). Our system was officially
ranked second in Phase 1 evaluation and first in both parts of Phase 2
evaluation. After fixing a submission error in Pharse 1, our approach yields
the best results overall. In this paper, in addition to a system description,
we also provide further analysis of our results, highlighting its strengths and
limitations. We make our code publicly available at
https://github.com/Liu-Hy/nlp-contrib-graph.
Related papers
- Co-training for Low Resource Scientific Natural Language Inference [65.37685198688538]
We propose a novel co-training method that assigns weights based on the training dynamics of the classifiers to the distantly supervised labels.
By assigning importance weights instead of filtering out examples based on an arbitrary threshold on the predicted confidence, we maximize the usage of automatically labeled data.
The proposed method obtains an improvement of 1.5% in Macro F1 over the distant supervision baseline, and substantial improvements over several other strong SSL baselines.
arXiv Detail & Related papers (2024-06-20T18:35:47Z) - Software Mention Recognition with a Three-Stage Framework Based on BERTology Models at SOMD 2024 [0.0]
This paper describes our systems for the sub-task I in the Software Mention Detection in Scholarly Publications shared-task.
Our best performing system addresses the named entity recognition problem through a three-stage framework.
Our framework based on the XLM-R-based model achieves a weighted F1-score of 67.80%, delivering our team the 3rd rank in Sub-task I for the Software Mention Recognition task.
arXiv Detail & Related papers (2024-04-23T17:06:24Z) - Prefer to Classify: Improving Text Classifiers via Auxiliary Preference
Learning [76.43827771613127]
In this paper, we investigate task-specific preferences between pairs of input texts as a new alternative way for such auxiliary data annotation.
We propose a novel multi-task learning framework, called prefer-to-classify (P2C), which can enjoy the cooperative effect of learning both the given classification task and the auxiliary preferences.
arXiv Detail & Related papers (2023-06-08T04:04:47Z) - AWTE-BERT:Attending to Wordpiece Tokenization Explicitly on BERT for
Joint Intent Classification and SlotFilling [5.684659127683238]
BERT (Bidirectional Representations from Transformers) achieves the joint optimization of the two tasks.
We propose a novel joint model based on BERT, which explicitly models the multiple sub-tokens features after wordpiece tokenization.
Experimental results demonstrate that our proposed model achieves significant improvement on intent classification accuracy, slot filling F1, and sentence-level semantic frame accuracy.
arXiv Detail & Related papers (2022-11-27T13:49:19Z) - Binary Classification with Positive Labeling Sources [71.37692084951355]
We propose WEAPO, a simple yet competitive WS method for producing training labels without negative labeling sources.
We show WEAPO achieves the highest averaged performance on 10 benchmark datasets.
arXiv Detail & Related papers (2022-08-02T19:32:08Z) - Question-Based Salient Span Selection for More Controllable Text
Summarization [67.68208237480646]
We propose a method for incorporating question-answering (QA) signals into a summarization model.
Our method identifies salient noun phrases (NPs) in the input document by automatically generating wh-questions that are answered by the NPs.
This QA-based signal is incorporated into a two-stage summarization model which first marks salient NPs in the input document using a classification model, then conditionally generates a summary.
arXiv Detail & Related papers (2021-11-15T17:36:41Z) - Class Token and Knowledge Distillation for Multi-head Self-Attention
Speaker Verification Systems [20.55054374525828]
This paper explores three novel approaches to improve the performance of speaker verification systems based on deep neural networks (DNN)
Firstly, we propose the use of a learnable vector called Class token to replace the average global pooling mechanism to extract the embeddings.
Second, we have added a distilled representation token for training a teacher-student pair of networks using the Knowledge Distillation (KD) philosophy.
arXiv Detail & Related papers (2021-11-06T09:47:05Z) - Low-Resource Task-Oriented Semantic Parsing via Intrinsic Modeling [65.51280121472146]
We exploit what we intrinsically know about ontology labels to build efficient semantic parsing models.
Our model is highly efficient using a low-resource benchmark derived from TOPv2.
arXiv Detail & Related papers (2021-04-15T04:01:02Z) - KnowGraph@IITK at SemEval-2021 Task 11: Building KnowledgeGraph for NLP
Research [2.1012672709024294]
We develop a system for a research paper contributions-focused knowledge graph over Natural Language Processing literature.
The proposed system is agnostic to the subject domain and can be applied for building a knowledge graph for any area.
Our system achieved F1 score of 0.38, 0.63 and 0.76 in end-to-end pipeline testing, phrase extraction testing and triplet extraction testing respectively.
arXiv Detail & Related papers (2021-04-04T14:33:21Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.