SemEval-2020 Task 10: Emphasis Selection for Written Text in Visual
Media
- URL: http://arxiv.org/abs/2008.03274v1
- Date: Fri, 7 Aug 2020 17:24:53 GMT
- Title: SemEval-2020 Task 10: Emphasis Selection for Written Text in Visual
Media
- Authors: Amirreza Shirani, Franck Dernoncourt, Nedim Lipka, Paul Asente, Jose
Echevarria and Thamar Solorio
- Abstract summary: We present the main findings and compare the results of SemEval-2020 Task 10, Emphasis Selection for Written Text in Visual Media.
The goal of this shared task is to design automatic methods for emphasis selection.
The analysis of systems submitted to the task indicates that BERT and RoBERTa were the most common choice of pre-trained models used.
- Score: 50.29389719723529
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present the main findings and compare the results of
SemEval-2020 Task 10, Emphasis Selection for Written Text in Visual Media. The
goal of this shared task is to design automatic methods for emphasis selection,
i.e. choosing candidates for emphasis in textual content to enable automated
design assistance in authoring. The main focus is on short text instances for
social media, with a variety of examples, from social media posts to
inspirational quotes. Participants were asked to model emphasis using plain
text with no additional context from the user or other design considerations.
SemEval-2020 Emphasis Selection shared task attracted 197 participants in the
early phase and a total of 31 teams made submissions to this task. The
highest-ranked submission achieved 0.823 Matchm score. The analysis of systems
submitted to the task indicates that BERT and RoBERTa were the most common
choice of pre-trained models used, and part of speech tag (POS) was the most
useful feature. Full results can be found on the task's website.
Related papers
- SemEval 2024 -- Task 10: Emotion Discovery and Reasoning its Flip in
Conversation (EDiReF) [61.49972925493912]
SemEval-2024 Task 10 is a shared task centred on identifying emotions in code-mixed dialogues.
This task comprises three distinct subtasks - emotion recognition in conversation for code-mixed dialogues, emotion flip reasoning for code-mixed dialogues, and emotion flip reasoning for English dialogues.
A total of 84 participants engaged in this task, with the most adept systems attaining F1-scores of 0.70, 0.79, and 0.76 for the respective subtasks.
arXiv Detail & Related papers (2024-02-29T08:20:06Z) - Text2Topic: Multi-Label Text Classification System for Efficient Topic
Detection in User Generated Content with Zero-Shot Capabilities [2.7311827519141363]
We propose Text to Topic (Text2Topic), which achieves high multi-label classification performance.
Text2Topic supports zero-shot predictions, produces domain-specific text embeddings, and enables production-scale batch-inference.
The model is deployed on a real-world stream processing platform, and it outperforms other models with 92.9% micro mAP.
arXiv Detail & Related papers (2023-10-23T11:33:24Z) - ICDAR 2023 Competition on Structured Text Extraction from Visually-Rich
Document Images [198.35937007558078]
The competition opened on 30th December, 2022 and closed on 24th March, 2023.
There are 35 participants and 91 valid submissions received for Track 1, and 15 participants and 26 valid submissions received for Track 2.
According to the performance of the submissions, we believe there is still a large gap on the expected information extraction performance for complex and zero-shot scenarios.
arXiv Detail & Related papers (2023-06-05T22:20:52Z) - Controlled Text Reduction [15.102190738450092]
We formalize textitControlled Text Reduction as a standalone task.
A model then needs to generate a coherent text that includes all and only the target information.
arXiv Detail & Related papers (2022-10-24T17:59:03Z) - MIDAS at SemEval-2020 Task 10: Emphasis Selection using Label
Distribution Learning and Contextual Embeddings [46.973153861604416]
This paper presents our submission to the SemEval 2020 - Task 10 on emphasis selection in written text.
We approach this emphasis selection problem as a sequence labeling task where we represent the underlying text with contextual embedding models.
Our best performing architecture is an ensemble of different models, which achieved an overall matching score of 0.783, placing us 15th out of 31 participating teams.
arXiv Detail & Related papers (2020-09-06T00:15:33Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - IITK at SemEval-2020 Task 10: Transformers for Emphasis Selection [8.352123313770552]
This paper describes the system proposed for addressing the research problem posed in Task 10 of SemEval-2020: Emphasis Selection For Written Text in Visual Media.
We propose an end-to-end model that takes as input the text and corresponding to each word gives the probability of the word to be emphasized.
arXiv Detail & Related papers (2020-07-21T14:05:56Z) - Let Me Choose: From Verbal Context to Font Selection [50.293897197235296]
We aim to learn associations between visual attributes of fonts and the verbal context of the texts they are typically applied to.
We introduce a new dataset, containing examples of different topics in social media posts and ads, labeled through crowd-sourcing.
arXiv Detail & Related papers (2020-05-03T17:36:17Z) - CIRCE at SemEval-2020 Task 1: Ensembling Context-Free and
Context-Dependent Word Representations [0.0]
We present an ensemble model that makes predictions based on context-free and context-dependent word representations.
The key findings are that (1) context-free word representations are a powerful and robust baseline, (2) a sentence classification objective can be used to obtain useful context-dependent word representations, and (3) combining those representations increases performance on some datasets while decreasing performance on others.
arXiv Detail & Related papers (2020-04-30T13:18:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.