DeDisCo at the DISRPT 2025 Shared Task: A System for Discourse Relation Classification
- URL: http://arxiv.org/abs/2509.11498v4
- Date: Fri, 19 Sep 2025 21:27:23 GMT
- Title: DeDisCo at the DISRPT 2025 Shared Task: A System for Discourse Relation Classification
- Authors: Zhuoxuan Ju, Jingni Wu, Abhishek Purushothama, Amir Zeldes,
- Abstract summary: This paper presents DeDisCo, Georgetown University's entry in the DISRPT 2025 shared task on discourse relation classification.<n>We test two approaches, using an mt5-based encoder and a decoder based approach using the openly available Qwen model.<n>Our system achieves a macro-accuracy score of 71.28, and we provide some interpretation and error analysis for our results.
- Score: 6.070010259231488
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents DeDisCo, Georgetown University's entry in the DISRPT 2025 shared task on discourse relation classification. We test two approaches, using an mt5-based encoder and a decoder based approach using the openly available Qwen model. We also experiment on training with augmented dataset for low-resource languages using matched data translated automatically from English, as well as using some additional linguistic features inspired by entries in previous editions of the Shared Task. Our system achieves a macro-accuracy score of 71.28, and we provide some interpretation and error analysis for our results.
Related papers
- SLRTP2025 Sign Language Production Challenge: Methodology, Results, and Future Work [87.9341538630949]
The first Sign Language Production Challenge was held as part of the third SLRTP Workshop at CVPR 2025.<n>The competition's aims are to evaluate architectures that translate from spoken language sentences to a sequence of skeleton poses.<n>This paper presents the challenge design and the winning methodologies.
arXiv Detail & Related papers (2025-08-09T11:57:33Z) - Annif at SemEval-2025 Task 5: Traditional XMTC augmented by LLMs [0.0]
This paper presents the Annif system in the SemEval-2025 Task 5 (LLMs)<n>It focussed on subject indexing using large language models.<n>Our approach combines traditional natural language processing and machine learning techniques.
arXiv Detail & Related papers (2025-04-28T11:04:23Z) - HiTZ at VarDial 2025 NorSID: Overcoming Data Scarcity with Language Transfer and Automatic Data Annotation [5.989003825349711]
We present our submission for the NorSID Shared Task, consisting of three tasks: Intent Detection, Slot Filling and Dialect Identification.<n>For Intent Detection and Slot Filling, we have fine-tuned a multitask model in a cross-lingual setting, to leverage the xSID dataset available in 17 languages.<n>In the case of Dialect Identification, our final submission consists of a model fine-tuned on the provided development set, which has obtained the highest scores.
arXiv Detail & Related papers (2024-12-13T12:31:06Z) - Cross-lingual Contextualized Phrase Retrieval [63.80154430930898]
We propose a new task formulation of dense retrieval, cross-lingual contextualized phrase retrieval.
We train our Cross-lingual Contextualized Phrase Retriever (CCPR) using contrastive learning.
On the phrase retrieval task, CCPR surpasses baselines by a significant margin, achieving a top-1 accuracy that is at least 13 points higher.
arXiv Detail & Related papers (2024-03-25T14:46:51Z) - A ML-LLM pairing for better code comment classification [0.0]
We answer the code comment classification shared task challenge by providing a two-fold evaluation.
Our best model, which took second place in the shared task, is a Neural Network with a Macro-F1 score of 88.401% on the provided seed data.
arXiv Detail & Related papers (2023-10-13T12:43:13Z) - Unify word-level and span-level tasks: NJUNLP's Participation for the
WMT2023 Quality Estimation Shared Task [59.46906545506715]
We introduce the NJUNLP team to the WMT 2023 Quality Estimation (QE) shared task.
Our team submitted predictions for the English-German language pair on all two sub-tasks.
Our models achieved the best results in English-German for both word-level and fine-grained error span detection sub-tasks.
arXiv Detail & Related papers (2023-09-23T01:52:14Z) - Strategies for improving low resource speech to text translation relying
on pre-trained ASR models [59.90106959717875]
This paper presents techniques and findings for improving the performance of low-resource speech to text translation (ST)
We conducted experiments on both simulated and real-low resource setups, on language pairs English - Portuguese, and Tamasheq - French respectively.
arXiv Detail & Related papers (2023-05-31T21:58:07Z) - Team \'UFAL at CMCL 2022 Shared Task: Figuring out the correct recipe
for predicting Eye-Tracking features using Pretrained Language Models [9.087729124428467]
We describe our systems for the CMCL 2022 shared task on predicting eye-tracking information.
Our submissions achieved an average MAE of 5.72 and ranked 5th in the shared task.
arXiv Detail & Related papers (2022-04-11T10:43:34Z) - NEMO: Frequentist Inference Approach to Constrained Linguistic Typology
Feature Prediction in SIGTYP 2020 Shared Task [83.43738174234053]
We employ frequentist inference to represent correlations between typological features and use this representation to train simple multi-class estimators that predict individual features.
Our best configuration achieved the micro-averaged accuracy score of 0.66 on 149 test languages.
arXiv Detail & Related papers (2020-10-12T19:25:43Z) - The Paradigm Discovery Problem [121.79963594279893]
We formalize the paradigm discovery problem and develop metrics for judging systems.
We report empirical results on five diverse languages.
Our code and data are available for public use.
arXiv Detail & Related papers (2020-05-04T16:38:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.