Enhancing Modern Supervised Word Sense Disambiguation Models by Semantic
Lexical Resources
- URL: http://arxiv.org/abs/2402.13302v1
- Date: Tue, 20 Feb 2024 13:47:51 GMT
- Title: Enhancing Modern Supervised Word Sense Disambiguation Models by Semantic
Lexical Resources
- Authors: Stefano Melacci and Achille Globo and Leonardo Rigutini
- Abstract summary: Supervised models for Word Sense Disambiguation (WSD) currently yield to state-of-the-art results in the most popular benchmarks.
We enhance "modern" supervised WSD models exploiting two popular SLRs: WordNet and WordNet Domains.
We study the effect of different types of semantic features, investigating their interaction with local contexts encoded by means of mixtures of Word Embeddings or Recurrent Neural Networks.
- Score: 11.257738983764499
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised models for Word Sense Disambiguation (WSD) currently yield to
state-of-the-art results in the most popular benchmarks. Despite the recent
introduction of Word Embeddings and Recurrent Neural Networks to design
powerful context-related features, the interest in improving WSD models using
Semantic Lexical Resources (SLRs) is mostly restricted to knowledge-based
approaches. In this paper, we enhance "modern" supervised WSD models exploiting
two popular SLRs: WordNet and WordNet Domains. We propose an effective way to
introduce semantic features into the classifiers, and we consider using the SLR
structure to augment the training data. We study the effect of different types
of semantic features, investigating their interaction with local contexts
encoded by means of mixtures of Word Embeddings or Recurrent Neural Networks,
and we extend the proposed model into a novel multi-layer architecture for WSD.
A detailed experimental comparison in the recent Unified Evaluation Framework
(Raganato et al., 2017) shows that the proposed approach leads to supervised
models that compare favourably with the state-of-the art.
Related papers
- Align-SLM: Textless Spoken Language Models with Reinforcement Learning from AI Feedback [50.84142264245052]
This work introduces the Align-SLM framework to enhance the semantic understanding of textless Spoken Language Models (SLMs)
Our approach generates multiple speech continuations from a given prompt and uses semantic metrics to create preference data for Direct Preference Optimization (DPO)
We evaluate the framework using ZeroSpeech 2021 benchmarks for lexical and syntactic modeling, the spoken version of the StoryCloze dataset for semantic coherence, and other speech generation metrics, including the GPT4-o score and human evaluation.
arXiv Detail & Related papers (2024-11-04T06:07:53Z) - LlaMaVAE: Guiding Large Language Model Generation via Continuous Latent
Sentence Spaces [1.529963465178546]
We present LlaMaVAE, which combines expressive encoder and decoder models (sentenceT5 and LlaMA) with a VAE architecture to provide better text generation control to large language models (LLMs)
Experimental results reveal that LlaMaVAE can outperform the previous state-of-the-art VAE language model, Optimus, across various tasks.
arXiv Detail & Related papers (2023-12-20T17:25:23Z) - Improved Contextual Recognition In Automatic Speech Recognition Systems
By Semantic Lattice Rescoring [4.819085609772069]
We propose a novel approach for enhancing contextual recognition within ASR systems via semantic lattice processing.
Our solution consists of using Hidden Markov Models and Gaussian Mixture Models (HMM-GMM) along with Deep Neural Networks (DNN) models for better accuracy.
We demonstrate the effectiveness of our proposed framework on the LibriSpeech dataset with empirical analyses.
arXiv Detail & Related papers (2023-10-14T23:16:05Z) - RAVEN: In-Context Learning with Retrieval-Augmented Encoder-Decoder Language Models [57.12888828853409]
RAVEN is a model that combines retrieval-augmented masked language modeling and prefix language modeling.
Fusion-in-Context Learning enables the model to leverage more in-context examples without requiring additional training.
Our work underscores the potential of retrieval-augmented encoder-decoder language models for in-context learning.
arXiv Detail & Related papers (2023-08-15T17:59:18Z) - A Comparative Study of Transformers on Word Sense Disambiguation [0.0]
This paper presents a comparative study on the contextualization power of neural network-based embedding systems.
We evaluate their contextualization power using two sample Word Sense Disambiguation (WSD) tasks, SensEval-2 and SensEval-3.
Experimental results show that the proposed techniques also achieve superior results over the current state-of-the-art on both the WSD tasks.
arXiv Detail & Related papers (2021-11-30T14:10:22Z) - Meta-Learning with Variational Semantic Memory for Word Sense
Disambiguation [56.830395467247016]
We propose a model of semantic memory for WSD in a meta-learning setting.
Our model is based on hierarchical variational inference and incorporates an adaptive memory update rule via a hypernetwork.
We show our model advances the state of the art in few-shot WSD, supports effective learning in extremely data scarce scenarios.
arXiv Detail & Related papers (2021-06-05T20:40:01Z) - A Correspondence Variational Autoencoder for Unsupervised Acoustic Word
Embeddings [50.524054820564395]
We propose a new unsupervised model for mapping a variable-duration speech segment to a fixed-dimensional representation.
The resulting acoustic word embeddings can form the basis of search, discovery, and indexing systems for low- and zero-resource languages.
arXiv Detail & Related papers (2020-12-03T19:24:42Z) - Introducing Syntactic Structures into Target Opinion Word Extraction
with Deep Learning [89.64620296557177]
We propose to incorporate the syntactic structures of the sentences into the deep learning models for targeted opinion word extraction.
We also introduce a novel regularization technique to improve the performance of the deep learning models.
The proposed model is extensively analyzed and achieves the state-of-the-art performance on four benchmark datasets.
arXiv Detail & Related papers (2020-10-26T07:13:17Z) - A Comparative Study of Lexical Substitution Approaches based on Neural
Language Models [117.96628873753123]
We present a large-scale comparative study of popular neural language and masked language models.
We show that already competitive results achieved by SOTA LMs/MLMs can be further improved if information about the target word is injected properly.
arXiv Detail & Related papers (2020-05-29T18:43:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.