Hierarchical Multitask Learning with Dependency Parsing for Japanese
Semantic Role Labeling Improves Performance of Argument Identification
- URL: http://arxiv.org/abs/2101.06071v2
- Date: Fri, 26 Feb 2021 11:58:31 GMT
- Title: Hierarchical Multitask Learning with Dependency Parsing for Japanese
Semantic Role Labeling Improves Performance of Argument Identification
- Authors: Tomohiro Nakamura, Tomoya Miyashita, Soh Ohara
- Abstract summary: We propose a hierarchical multitask learning method with dependency parsing (DP) and show that our model achieves state-of-the-art results in Japanese SRL.
Also, we conduct experiments with a joint model that performs both argument identification and argument classification simultaneously.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the advent of FrameNet and PropBank, many semantic role labeling (SRL)
systems have been proposed in English. Although research on Japanese predicate
argument structure analysis (PASA) has been conducted, most studies focused on
surface cases. There are only few previous works on Japanese SRL for deep
cases, and their models' accuracies are low. Therefore, we propose a
hierarchical multitask learning method with dependency parsing (DP) and show
that our model achieves state-of-the-art results in Japanese SRL. Also, we
conduct experiments with a joint model that performs both argument
identification and argument classification simultaneously. The result suggests
that multitasking with DP is mainly effective for argument identification.
Related papers
- P-MMEval: A Parallel Multilingual Multitask Benchmark for Consistent Evaluation of LLMs [84.24644520272835]
Large language models (LLMs) showcase varied multilingual capabilities across tasks like translation, code generation, and reasoning.
Previous assessments often limited their scope to fundamental natural language processing (NLP) or isolated capability-specific tasks.
We present a pipeline for selecting available and reasonable benchmarks from massive ones, addressing the oversight in previous work regarding the utility of these benchmarks.
We introduce P-MMEval, a large-scale benchmark covering effective fundamental and capability-specialized datasets.
arXiv Detail & Related papers (2024-11-14T01:29:36Z) - The Power of Question Translation Training in Multilingual Reasoning: Broadened Scope and Deepened Insights [108.40766216456413]
We propose a question alignment framework to bridge the gap between large language models' English and non-English performance.
Experiment results show it can boost multilingual performance across diverse reasoning scenarios, model families, and sizes.
We analyze representation space, generated response and data scales, and reveal how question translation training strengthens language alignment within LLMs.
arXiv Detail & Related papers (2024-05-02T14:49:50Z) - Provable Benefits of Multi-task RL under Non-Markovian Decision Making
Processes [56.714690083118406]
In multi-task reinforcement learning (RL) under Markov decision processes (MDPs), the presence of shared latent structures has been shown to yield significant benefits to the sample efficiency compared to single-task RL.
We investigate whether such a benefit can extend to more general sequential decision making problems, such as partially observable MDPs (POMDPs) and more general predictive state representations (PSRs)
We propose a provably efficient algorithm UMT-PSR for finding near-optimal policies for all PSRs, and demonstrate that the advantage of multi-task learning manifests if the joint model class of PSR
arXiv Detail & Related papers (2023-10-20T14:50:28Z) - Generative Judge for Evaluating Alignment [84.09815387884753]
We propose a generative judge with 13B parameters, Auto-J, designed to address these challenges.
Our model is trained on user queries and LLM-generated responses under massive real-world scenarios.
Experimentally, Auto-J outperforms a series of strong competitors, including both open-source and closed-source models.
arXiv Detail & Related papers (2023-10-09T07:27:15Z) - Exploration of Language Dependency for Japanese Self-Supervised Speech
Representation Models [18.22157315310462]
Self-supervised learning (SSL) has been dramatically successful not only in monolingual but also in cross-lingual settings.
In this paper, we investigate how effective a cross-lingual model is in comparison with a monolingual model.
We examine how much unlabeled data collected in Japanese is needed to achieve performance comparable to a cross-lingual model pre-trained with tens of thousands of hours of English and/or multilingual data.
arXiv Detail & Related papers (2023-05-09T06:28:10Z) - Multi-level Distillation of Semantic Knowledge for Pre-training
Multilingual Language Model [15.839724725094916]
Multi-level Multilingual Knowledge Distillation (MMKD) is a novel method for improving multilingual language models.
We employ a teacher-student framework to adopt rich semantic representation knowledge in English BERT.
We conduct experiments on cross-lingual evaluation benchmarks including XNLI, PAWS-X, and XQuAD.
arXiv Detail & Related papers (2022-11-02T15:23:13Z) - DiSparse: Disentangled Sparsification for Multitask Model Compression [92.84435347164435]
DiSparse is a simple, effective, and first-of-its-kind multitask pruning and sparse training scheme.
Our experimental results demonstrate superior performance on various configurations and settings.
arXiv Detail & Related papers (2022-06-09T17:57:46Z) - A Generative Language Model for Few-shot Aspect-Based Sentiment Analysis [90.24921443175514]
We focus on aspect-based sentiment analysis, which involves extracting aspect term, category, and predicting their corresponding polarities.
We propose to reformulate the extraction and prediction tasks into the sequence generation task, using a generative language model with unidirectional attention.
Our approach outperforms the previous state-of-the-art (based on BERT) on average performance by a large margins in few-shot and full-shot settings.
arXiv Detail & Related papers (2022-04-11T18:31:53Z) - Zero-shot Cross-lingual Conversational Semantic Role Labeling [30.28068025468505]
We present an effective approach to perform zero-shot cross-lingual CSRL.
Our model implicitly learns language-agnostic, conversational structure-aware and semantically rich representations with the hierarchical encoders.
We confirm the usefulness of CSRL to non-Chinese conversational tasks such as the question-in-context rewriting task in English and the multi-turn dialogue response generation tasks in English, German and Japanese.
arXiv Detail & Related papers (2022-04-11T07:29:39Z) - Semantic Role Labeling as Dependency Parsing: Exploring Latent Tree
Structures Inside Arguments [33.95412952615206]
Recent works of SRL mainly fall into two lines:1) BIO-based and 2) span-based.
We propose to reduce SRL to a dependency parsing task and regard the flat argument spans as latent subtrees.
In particular, we equip our formulation with a novel span-constrained TreeCRF model to make tree structures span-aware.
arXiv Detail & Related papers (2021-10-13T17:02:29Z) - Incorporating Linguistic Knowledge for Abstractive Multi-document
Summarization [20.572283625521784]
We develop a neural network based abstractive multi-document summarization (MDS) model.
We process the dependency information into the linguistic-guided attention mechanism.
With the help of linguistic signals, sentence-level relations can be correctly captured.
arXiv Detail & Related papers (2021-09-23T08:13:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.