MIDAS at SemEval-2020 Task 10: Emphasis Selection using Label
Distribution Learning and Contextual Embeddings
- URL: http://arxiv.org/abs/2009.02619v1
- Date: Sun, 6 Sep 2020 00:15:33 GMT
- Title: MIDAS at SemEval-2020 Task 10: Emphasis Selection using Label
Distribution Learning and Contextual Embeddings
- Authors: Sarthak Anand, Pradyumna Gupta, Hemant Yadav, Debanjan Mahata, Rakesh
Gosangi, Haimin Zhang, Rajiv Ratn Shah
- Abstract summary: This paper presents our submission to the SemEval 2020 - Task 10 on emphasis selection in written text.
We approach this emphasis selection problem as a sequence labeling task where we represent the underlying text with contextual embedding models.
Our best performing architecture is an ensemble of different models, which achieved an overall matching score of 0.783, placing us 15th out of 31 participating teams.
- Score: 46.973153861604416
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents our submission to the SemEval 2020 - Task 10 on emphasis
selection in written text. We approach this emphasis selection problem as a
sequence labeling task where we represent the underlying text with various
contextual embedding models. We also employ label distribution learning to
account for annotator disagreements. We experiment with the choice of model
architectures, trainability of layers, and different contextual embeddings. Our
best performing architecture is an ensemble of different models, which achieved
an overall matching score of 0.783, placing us 15th out of 31 participating
teams. Lastly, we analyze the results in terms of parts of speech tags,
sentence lengths, and word ordering.
Related papers
- LastResort at SemEval-2024 Task 3: Exploring Multimodal Emotion Cause Pair Extraction as Sequence Labelling Task [3.489826905722736]
SemEval 2024 introduces the task of Multimodal Emotion Cause Analysis in Conversations.
This paper proposes models that tackle this task as an utterance labeling and a sequence labeling problem.
In the official leaderboard for the task, our architecture was ranked 8th, achieving an F1-score of 0.1759 on the leaderboard.
arXiv Detail & Related papers (2024-04-02T16:32:49Z) - Natural Language Inference Prompts for Zero-shot Emotion Classification
in Text across Corpora [11.986676998327864]
We show that the choice of a particular prompt formulation needs to fit to the corpus.
We show that this challenge can be tackled with combinations of multiple prompts.
arXiv Detail & Related papers (2022-09-14T15:07:36Z) - Enhanced Knowledge Selection for Grounded Dialogues via Document
Semantic Graphs [123.50636090341236]
We propose to automatically convert background knowledge documents into document semantic graphs.
Our document semantic graphs preserve sentence-level information through the use of sentence nodes and provide concept connections between sentences.
Our experiments show that our semantic graph-based knowledge selection improves over sentence selection baselines for both the knowledge selection task and the end-to-end response generation task on HollE.
arXiv Detail & Related papers (2022-06-15T04:51:32Z) - A Structured Span Selector [100.0808682810258]
We propose a novel grammar-based structured span selection model.
We evaluate our model on two popular span prediction tasks: coreference resolution and semantic role labeling.
arXiv Detail & Related papers (2022-05-08T23:58:40Z) - Minimally-Supervised Structure-Rich Text Categorization via Learning on
Text-Rich Networks [61.23408995934415]
We propose a novel framework for minimally supervised categorization by learning from the text-rich network.
Specifically, we jointly train two modules with different inductive biases -- a text analysis module for text understanding and a network learning module for class-discriminative, scalable network learning.
Our experiments show that given only three seed documents per category, our framework can achieve an accuracy of about 92%.
arXiv Detail & Related papers (2021-02-23T04:14:34Z) - LTIatCMU at SemEval-2020 Task 11: Incorporating Multi-Level Features for
Multi-Granular Propaganda Span Identification [70.1903083747775]
This paper describes our submission for the task of Propaganda Span Identification in news articles.
We introduce a BERT-BiLSTM based span-level propaganda classification model that identifies which token spans within the sentence are indicative of propaganda.
arXiv Detail & Related papers (2020-08-11T16:14:47Z) - SemEval-2020 Task 10: Emphasis Selection for Written Text in Visual
Media [50.29389719723529]
We present the main findings and compare the results of SemEval-2020 Task 10, Emphasis Selection for Written Text in Visual Media.
The goal of this shared task is to design automatic methods for emphasis selection.
The analysis of systems submitted to the task indicates that BERT and RoBERTa were the most common choice of pre-trained models used.
arXiv Detail & Related papers (2020-08-07T17:24:53Z) - IITK at SemEval-2020 Task 10: Transformers for Emphasis Selection [8.352123313770552]
This paper describes the system proposed for addressing the research problem posed in Task 10 of SemEval-2020: Emphasis Selection For Written Text in Visual Media.
We propose an end-to-end model that takes as input the text and corresponding to each word gives the probability of the word to be emphasized.
arXiv Detail & Related papers (2020-07-21T14:05:56Z) - UiO-UvA at SemEval-2020 Task 1: Contextualised Embeddings for Lexical
Semantic Change Detection [5.099262949886174]
This paper focuses on Subtask 2, ranking words by the degree of their semantic drift over time.
We find that the most effective algorithms rely on the cosine similarity between averaged token embeddings and the pairwise distances between token embeddings.
arXiv Detail & Related papers (2020-04-30T18:43:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.