A Novel Multi-Task Learning Approach for Context-Sensitive Compound Type
Identification in Sanskrit
- URL: http://arxiv.org/abs/2208.10310v1
- Date: Mon, 22 Aug 2022 13:41:51 GMT
- Title: A Novel Multi-Task Learning Approach for Context-Sensitive Compound Type
Identification in Sanskrit
- Authors: Jivnesh Sandhan, Ashish Gupta, Hrishikesh Terdalkar, Tushar Sandhan,
Suvendu Samanta, Laxmidhar Behera and Pawan Goyal
- Abstract summary: We propose a novel multi-task learning architecture which incorporates the contextual information and enriches the complementary syntactic information.
Experiments on the benchmark datasets for SaCTI show 6.1 points (Accuracy) and 7.7 points (F1-score) absolute gain compared to the state-of-the-art system.
- Score: 13.742271198030998
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The phenomenon of compounding is ubiquitous in Sanskrit. It serves for
achieving brevity in expressing thoughts, while simultaneously enriching the
lexical and structural formation of the language. In this work, we focus on the
Sanskrit Compound Type Identification (SaCTI) task, where we consider the
problem of identifying semantic relations between the components of a compound
word. Earlier approaches solely rely on the lexical information obtained from
the components and ignore the most crucial contextual and syntactic information
useful for SaCTI. However, the SaCTI task is challenging primarily due to the
implicitly encoded context-sensitive semantic relation between the compound
components.
Thus, we propose a novel multi-task learning architecture which incorporates
the contextual information and enriches the complementary syntactic information
using morphological tagging and dependency parsing as two auxiliary tasks.
Experiments on the benchmark datasets for SaCTI show 6.1 points (Accuracy) and
7.7 points (F1-score) absolute gain compared to the state-of-the-art system.
Further, our multi-lingual experiments demonstrate the efficacy of the proposed
architecture in English and Marathi languages.The code and datasets are
publicly available at https://github.com/ashishgupta2598/SaCTI
Related papers
- Incorporating Lexical and Syntactic Knowledge for Unsupervised Cross-Lingual Transfer [4.944761231728674]
We present a novel framework called "Lexicon-Syntax Enhanced Multilingual BERT"
We use Multilingual BERT as the base model and employ two techniques to enhance its learning capabilities.
Our experimental results demonstrate this framework can consistently outperform all baselines of zero-shot cross-lingual transfer.
arXiv Detail & Related papers (2024-04-25T14:10:52Z) - SemRel2024: A Collection of Semantic Textual Relatedness Datasets for 13 Languages [44.017657230247934]
We present textitSemRel, a new semantic relatedness dataset collection annotated by native speakers across 13 languages.
These languages originate from five distinct language families and are predominantly spoken in Africa and Asia.
Each instance in the SemRel datasets is a sentence pair associated with a score that represents the degree of semantic textual relatedness between the two sentences.
arXiv Detail & Related papers (2024-02-13T18:04:53Z) - Subspace Chronicles: How Linguistic Information Emerges, Shifts and
Interacts during Language Model Training [56.74440457571821]
We analyze tasks covering syntax, semantics and reasoning, across 2M pre-training steps and five seeds.
We identify critical learning phases across tasks and time, during which subspaces emerge, share information, and later disentangle to specialize.
Our findings have implications for model interpretability, multi-task learning, and learning from limited data.
arXiv Detail & Related papers (2023-10-25T09:09:55Z) - DepNeCTI: Dependency-based Nested Compound Type Identification for
Sanskrit [7.04795623262177]
This work introduces the novel task of nested compound type identification (NeCTI)
It aims to identify nested spans of a multi-component compound and decode the implicit semantic relations between them.
To the best of our knowledge, this is the first attempt in the field of lexical semantics to propose this task.
arXiv Detail & Related papers (2023-10-14T06:11:53Z) - mCL-NER: Cross-Lingual Named Entity Recognition via Multi-view
Contrastive Learning [54.523172171533645]
Cross-lingual named entity recognition (CrossNER) faces challenges stemming from uneven performance due to the scarcity of multilingual corpora.
We propose Multi-view Contrastive Learning for Cross-lingual Named Entity Recognition (mCL-NER)
Our experiments on the XTREME benchmark, spanning 40 languages, demonstrate the superiority of mCL-NER over prior data-driven and model-based approaches.
arXiv Detail & Related papers (2023-08-17T16:02:29Z) - Linguistically-Informed Neural Architectures for Lexical, Syntactic and
Semantic Tasks in Sanskrit [1.184066113335041]
This thesis aims to make Sanskrit manuscripts more accessible to the end-users through natural language technologies.
The morphological richness, compounding, free word orderliness, and low-resource nature of Sanskrit pose significant challenges for developing deep learning solutions.
We identify four fundamental tasks, which are crucial for developing a robust NLP technology for Sanskrit.
arXiv Detail & Related papers (2023-08-17T06:33:33Z) - $\mu$PLAN: Summarizing using a Content Plan as Cross-Lingual Bridge [72.64847925450368]
Cross-lingual summarization consists of generating a summary in one language given an input document in a different language.
This work presents $mu$PLAN, an approach to cross-lingual summarization that uses an intermediate planning step as a cross-lingual bridge.
arXiv Detail & Related papers (2023-05-23T16:25:21Z) - A Comprehensive Understanding of Code-mixed Language Semantics using
Hierarchical Transformer [28.3684494647968]
We propose a hierarchical transformer-based architecture (HIT) to learn the semantics of code-mixed languages.
We evaluate the proposed method across 6 Indian languages and 9 NLP tasks on 17 datasets.
arXiv Detail & Related papers (2022-04-27T07:50:18Z) - ERICA: Improving Entity and Relation Understanding for Pre-trained
Language Models via Contrastive Learning [97.10875695679499]
We propose a novel contrastive learning framework named ERICA in pre-training phase to obtain a deeper understanding of the entities and their relations in text.
Experimental results demonstrate that our proposed ERICA framework achieves consistent improvements on several document-level language understanding tasks.
arXiv Detail & Related papers (2020-12-30T03:35:22Z) - Multilingual Irony Detection with Dependency Syntax and Neural Models [61.32653485523036]
It focuses on the contribution from syntactic knowledge, exploiting linguistic resources where syntax is annotated according to the Universal Dependencies scheme.
The results suggest that fine-grained dependency-based syntactic information is informative for the detection of irony.
arXiv Detail & Related papers (2020-11-11T11:22:05Z) - On the Language Neutrality of Pre-trained Multilingual Representations [70.93503607755055]
We investigate the language-neutrality of multilingual contextual embeddings directly and with respect to lexical semantics.
Our results show that contextual embeddings are more language-neutral and, in general, more informative than aligned static word-type embeddings.
We show how to reach state-of-the-art accuracy on language identification and match the performance of statistical methods for word alignment of parallel sentences.
arXiv Detail & Related papers (2020-04-09T19:50:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.