Towards Explainability in NLP: Analyzing and Calculating Word Saliency
through Word Properties
- URL: http://arxiv.org/abs/2207.08083v1
- Date: Sun, 17 Jul 2022 06:02:48 GMT
- Title: Towards Explainability in NLP: Analyzing and Calculating Word Saliency
through Word Properties
- Authors: Jialiang Dong, Zhitao Guan, Longfei Wu, Zijian Zhang
- Abstract summary: We explore the relationships between the word saliency and the word properties.
We establish a mapping model, Seq2Saliency, from the words in a text sample and their properties to the saliency values.
The experimental evaluations are conducted to analyze the saliency of words with different properties.
- Score: 4.330880304715002
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The wide use of black-box models in natural language processing brings great
challenges to the understanding of the decision basis, the trustworthiness of
the prediction results, and the improvement of the model performance. The words
in text samples have properties that reflect their semantics and contextual
information, such as the part of speech, the position, etc. These properties
may have certain relationships with the word saliency, which is of great help
for studying the explainability of the model predictions. In this paper, we
explore the relationships between the word saliency and the word properties.
According to the analysis results, we further establish a mapping model,
Seq2Saliency, from the words in a text sample and their properties to the
saliency values based on the idea of sequence tagging. In addition, we
establish a new dataset called PrSalM, which contains each word in the text
samples, the word properties, and the word saliency values. The experimental
evaluations are conducted to analyze the saliency of words with different
properties. The effectiveness of the Seq2Saliency model is verified.
Related papers
- Explaining word embeddings with perfect fidelity: Case study in research impact prediction [0.0]
Self-model Rated Entities (SMER) for logistic regression-based classification models trained on word embeddings.
We show that SMER has theoretically perfect fidelity with the explained model, as its prediction corresponds exactly to the average of predictions for individual words in the text.
arXiv Detail & Related papers (2024-09-24T09:28:24Z) - Are we describing the same sound? An analysis of word embedding spaces
of expressive piano performance [4.867952721052875]
We investigate the uncertainty for the domain of characterizations of expressive piano performance.
We test five embedding models and their similarity structure for correspondence with the ground truth.
The quality of embedding models shows great variability with respect to this task.
arXiv Detail & Related papers (2023-12-31T12:20:03Z) - How Well Do Text Embedding Models Understand Syntax? [50.440590035493074]
The ability of text embedding models to generalize across a wide range of syntactic contexts remains under-explored.
Our findings reveal that existing text embedding models have not sufficiently addressed these syntactic understanding challenges.
We propose strategies to augment the generalization ability of text embedding models in diverse syntactic scenarios.
arXiv Detail & Related papers (2023-11-14T08:51:00Z) - Visualizing the Obvious: A Concreteness-based Ensemble Model for Noun
Property Prediction [34.37730333491428]
properties of nouns are more challenging to extract compared to other types of knowledge because they are rarely explicitly stated in texts.
We propose to extract these properties from images and use them in an ensemble model, in order to complement the information that is extracted from language models.
Our results show that the proposed combination of text and images greatly improves noun property prediction compared to powerful text-based language models.
arXiv Detail & Related papers (2022-10-24T01:25:21Z) - An Empirical Investigation of Commonsense Self-Supervision with
Knowledge Graphs [67.23285413610243]
Self-supervision based on the information extracted from large knowledge graphs has been shown to improve the generalization of language models.
We study the effect of knowledge sampling strategies and sizes that can be used to generate synthetic data for adapting language models.
arXiv Detail & Related papers (2022-05-21T19:49:04Z) - A Latent-Variable Model for Intrinsic Probing [93.62808331764072]
We propose a novel latent-variable formulation for constructing intrinsic probes.
We find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.
arXiv Detail & Related papers (2022-01-20T15:01:12Z) - ALL Dolphins Are Intelligent and SOME Are Friendly: Probing BERT for
Nouns' Semantic Properties and their Prototypicality [4.915907527975786]
We probe BERT (Devlin et al.) for the construction of English nouns as expressed by adjectives that do not restrict the reference scope.
We base our study on psycholinguistics datasets that capture the association strength between nouns and their semantic features.
We show that when tested in a fine-tuning setting addressing entailment, BERT successfully leverages the information needed for reasoning about the meaning of adjective-nouns.
arXiv Detail & Related papers (2021-10-12T21:43:37Z) - Did the Cat Drink the Coffee? Challenging Transformers with Generalized
Event Knowledge [59.22170796793179]
Transformers Language Models (TLMs) were tested on a benchmark for the textitdynamic estimation of thematic fit
Our results show that TLMs can reach performances that are comparable to those achieved by SDM.
However, additional analysis consistently suggests that TLMs do not capture important aspects of event knowledge.
arXiv Detail & Related papers (2021-07-22T20:52:26Z) - Understanding Synonymous Referring Expressions via Contrastive Features [105.36814858748285]
We develop an end-to-end trainable framework to learn contrastive features on the image and object instance levels.
We conduct extensive experiments to evaluate the proposed algorithm on several benchmark datasets.
arXiv Detail & Related papers (2021-04-20T17:56:24Z) - Dynamic Contextualized Word Embeddings [20.81930455526026]
We introduce dynamic contextualized word embeddings that represent words as a function of both linguistic and extralinguistic context.
Based on a pretrained language model (PLM), dynamic contextualized word embeddings model time and social space jointly.
We highlight potential application scenarios by means of qualitative and quantitative analyses on four English datasets.
arXiv Detail & Related papers (2020-10-23T22:02:40Z) - Explaining and Improving Model Behavior with k Nearest Neighbor
Representations [107.24850861390196]
We propose using k nearest neighbor representations to identify training examples responsible for a model's predictions.
We show that kNN representations are effective at uncovering learned spurious associations.
Our results indicate that the kNN approach makes the finetuned model more robust to adversarial inputs.
arXiv Detail & Related papers (2020-10-18T16:55:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.