Bird's Eye: Probing for Linguistic Graph Structures with a Simple
Information-Theoretic Approach
- URL: http://arxiv.org/abs/2105.02629v1
- Date: Thu, 6 May 2021 13:01:57 GMT
- Title: Bird's Eye: Probing for Linguistic Graph Structures with a Simple
Information-Theoretic Approach
- Authors: Yifan Hou and Mrinmaya Sachan
- Abstract summary: We propose a new information-theoretic probe, Bird's Eye, for detecting if and how representations encode the information in linguistic graphs.
We also propose an approach to use our probe to investigate localized linguistic information in the linguistic graphs using perturbation analysis.
- Score: 23.66191446048298
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: NLP has a rich history of representing our prior understanding of language in
the form of graphs. Recent work on analyzing contextualized text
representations has focused on hand-designed probe models to understand how and
to what extent do these representations encode a particular linguistic
phenomenon. However, due to the inter-dependence of various phenomena and
randomness of training probe models, detecting how these representations encode
the rich information in these linguistic graphs remains a challenging problem.
In this paper, we propose a new information-theoretic probe, Bird's Eye, which
is a fairly simple probe method for detecting if and how these representations
encode the information in these linguistic graphs. Instead of using classifier
performance, our probe takes an information-theoretic view of probing and
estimates the mutual information between the linguistic graph embedded in a
continuous space and the contextualized word representations. Furthermore, we
also propose an approach to use our probe to investigate localized linguistic
information in the linguistic graphs using perturbation analysis. We call this
probing setup Worm's Eye. Using these probes, we analyze BERT models on their
ability to encode a syntactic and a semantic graph structure, and find that
these models encode to some degree both syntactic as well as semantic
information; albeit syntactic information to a greater extent.
Related papers
- Verbalized Graph Representation Learning: A Fully Interpretable Graph Model Based on Large Language Models Throughout the Entire Process [8.820909397907274]
We propose a verbalized graph representation learning (VGRL) method which is fully interpretable.
In contrast to traditional graph machine learning models, VGRL constrains this parameter space to be text description.
We conduct several studies to empirically evaluate the effectiveness of VGRL.
arXiv Detail & Related papers (2024-10-02T12:07:47Z) - When Graph Data Meets Multimodal: A New Paradigm for Graph Understanding
and Reasoning [54.84870836443311]
The paper presents a new paradigm for understanding and reasoning about graph data by integrating image encoding and multimodal technologies.
This approach enables the comprehension of graph data through an instruction-response format, utilizing GPT-4V's advanced capabilities.
The study evaluates this paradigm on various graph types, highlighting the model's strengths and weaknesses, particularly in Chinese OCR performance and complex reasoning tasks.
arXiv Detail & Related papers (2023-12-16T08:14:11Z) - Semantic Parsing for Question Answering over Knowledge Graphs [3.10647754288788]
We introduce a novel method with graph-to-segment mapping for question answering over knowledge graphs.
This method centers on semantic parsing, a key approach for interpreting these utterances.
Our framework employs a combination of rule-based and neural-based techniques to parse and construct semantic segment sequences.
arXiv Detail & Related papers (2023-12-01T20:45:06Z) - Probing via Prompting [71.7904179689271]
This paper introduces a novel model-free approach to probing, by formulating probing as a prompting task.
We conduct experiments on five probing tasks and show that our approach is comparable or better at extracting information than diagnostic probes.
We then examine the usefulness of a specific linguistic property for pre-training by removing the heads that are essential to that property and evaluating the resulting model's performance on language modeling.
arXiv Detail & Related papers (2022-07-04T22:14:40Z) - Syntax-informed Question Answering with Heterogeneous Graph Transformer [2.139714421848487]
We present a linguistics-informed question answering approach that extends and fine-tunes a pre-trained neural language model.
We illustrate the approach by the addition of syntactic information in the form of dependency and constituency graphic structures connecting tokens and virtual tokens.
arXiv Detail & Related papers (2022-04-01T07:48:03Z) - A Latent-Variable Model for Intrinsic Probing [93.62808331764072]
We propose a novel latent-variable formulation for constructing intrinsic probes.
We find empirical evidence that pre-trained representations develop a cross-lingually entangled notion of morphosyntax.
arXiv Detail & Related papers (2022-01-20T15:01:12Z) - Probing Contextual Language Models for Common Ground with Visual
Representations [76.05769268286038]
We design a probing model that evaluates how effective are text-only representations in distinguishing between matching and non-matching visual representations.
Our findings show that language representations alone provide a strong signal for retrieving image patches from the correct object categories.
Visually grounded language models slightly outperform text-only language models in instance retrieval, but greatly under-perform humans.
arXiv Detail & Related papers (2020-05-01T21:28:28Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z) - Temporal Embeddings and Transformer Models for Narrative Text
Understanding [72.88083067388155]
We present two approaches to narrative text understanding for character relationship modelling.
The temporal evolution of these relations is described by dynamic word embeddings, that are designed to learn semantic changes over time.
A supervised learning approach based on the state-of-the-art transformer model BERT is used instead to detect static relations between characters.
arXiv Detail & Related papers (2020-03-19T14:23:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.