One Graph to Rule them All: Using NLP and Graph Neural Networks to
analyse Tolkien's Legendarium
- URL: http://arxiv.org/abs/2210.07871v1
- Date: Fri, 14 Oct 2022 14:47:56 GMT
- Title: One Graph to Rule them All: Using NLP and Graph Neural Networks to
analyse Tolkien's Legendarium
- Authors: Vincenzo Perri, Lisi Qarkaxhija, Albin Zehe, Andreas Hotho, Ingo
Scholtes
- Abstract summary: We study character networks extracted from a text corpus of J.R.R. Tolkien's Legendarium.
We show that this perspective helps us to analyse and visualise the narrative style that characterises Tolkien's works.
- Score: 3.0448872422956432
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural Language Processing and Machine Learning have considerably advanced
Computational Literary Studies. Similarly, the construction of co-occurrence
networks of literary characters, and their analysis using methods from social
network analysis and network science, have provided insights into the micro-
and macro-level structure of literary texts. Combining these perspectives, in
this work we study character networks extracted from a text corpus of J.R.R.
Tolkien's Legendarium. We show that this perspective helps us to analyse and
visualise the narrative style that characterises Tolkien's works. Addressing
character classification, embedding and co-occurrence prediction, we further
investigate the advantages of state-of-the-art Graph Neural Networks over a
popular word embedding method. Our results highlight the large potential of
graph learning in Computational Literary Studies.
Related papers
- Hierarchical Compression of Text-Rich Graphs via Large Language Models [63.75293588479027]
Text-rich graphs are prevalent in data mining contexts like e-commerce and academic graphs.
This paper introduces Hierarchical Compression'' (HiCom), a novel method to align the capabilities of LLMs with the structure of text-rich graphs.
HiCom can outperform both GNNs and LLM backbones for node classification on e-commerce and citation graphs.
arXiv Detail & Related papers (2024-06-13T07:24:46Z) - GRENADE: Graph-Centric Language Model for Self-Supervised Representation
Learning on Text-Attributed Graphs [22.282756544376493]
We develop a novel Graph-Centric Language model, GRENADE, to solve the problem of self-supervised representation learning on text-attributed graphs.
GRENADE exploits the synergistic effect of both pre-trained language model and graph neural network by optimizing with two specialized self-supervised learning algorithms.
The proposed graph-centric self-supervised learning algorithms effectively help GRENADE to capture informative textual semantics as well as structural context information on text-attributed graphs.
arXiv Detail & Related papers (2023-10-23T17:18:35Z) - Seeing in Words: Learning to Classify through Language Bottlenecks [59.97827889540685]
Humans can explain their predictions using succinct and intuitive descriptions.
We show that a vision model whose feature representations are text can effectively classify ImageNet images.
arXiv Detail & Related papers (2023-06-29T00:24:42Z) - Connecting the Dots: What Graph-Based Text Representations Work Best for
Text Classification Using Graph Neural Networks? [25.898812694174772]
This work extensively investigates graph representation methods for text classification.
We compare different graph construction schemes using a variety of GNN architectures and setups.
Two Transformer-based large language models are also included to complement the study.
arXiv Detail & Related papers (2023-05-23T23:31:24Z) - Graph Neural Networks for Text Classification: A Survey [8.414181339242706]
Graph neural network-based models can deal with complex structured text data and exploit global information.
We bring the coverage of methods up to 2023, including corpus-level and document-level graph neural networks.
As well as the technological survey, we look at issues behind and future directions addressed in text classification using graph neural networks.
arXiv Detail & Related papers (2023-04-23T04:21:50Z) - Tragic and Comical Networks. Clustering Dramatic Genres According to
Structural Properties [0.0]
A growing tradition in the joint field of network studies and drama history produces interpretations from the character networks of the plays.
Our aim is to create a method that is able to cluster texts with similar structures on the basis of the play's well-interpretable and simple properties.
Finding these features is the most important part of our research, as well as establishing the appropriate statistical procedure to calculate the similarities between the texts.
arXiv Detail & Related papers (2023-02-16T12:36:16Z) - State of the Art and Potentialities of Graph-level Learning [54.68482109186052]
Graph-level learning has been applied to many tasks including comparison, regression, classification, and more.
Traditional approaches to learning a set of graphs rely on hand-crafted features, such as substructures.
Deep learning has helped graph-level learning adapt to the growing scale of graphs by extracting features automatically and encoding graphs into low-dimensional representations.
arXiv Detail & Related papers (2023-01-14T09:15:49Z) - TeKo: Text-Rich Graph Neural Networks with External Knowledge [75.91477450060808]
We propose a novel text-rich graph neural network with external knowledge (TeKo)
We first present a flexible heterogeneous semantic network that incorporates high-quality entities.
We then introduce two types of external knowledge, that is, structured triplets and unstructured entity description.
arXiv Detail & Related papers (2022-06-15T02:33:10Z) - GraphFormers: GNN-nested Transformers for Representation Learning on
Textual Graph [53.70520466556453]
We propose GraphFormers, where layerwise GNN components are nested alongside the transformer blocks of language models.
With the proposed architecture, the text encoding and the graph aggregation are fused into an iterative workflow.
In addition, a progressive learning strategy is introduced, where the model is successively trained on manipulated data and original data to reinforce its capability of integrating information on graph.
arXiv Detail & Related papers (2021-05-06T12:20:41Z) - Structuralist analysis for neural network system diagrams [6.233820957059352]
We argue that the heterogeneous diagrammatic notations used for neural network systems has implications for signification in this domain.
We use a corpus analysis to quantitatively cluster diagrams according to the author's representational choices.
arXiv Detail & Related papers (2021-04-30T07:50:19Z) - Promoting Graph Awareness in Linearized Graph-to-Text Generation [72.83863719868364]
We study the ability of linearized models to encode local graph structures.
Our findings motivate solutions to enrich the quality of models' implicit graph encodings.
We find that these denoising scaffolds lead to substantial improvements in downstream generation in low-resource settings.
arXiv Detail & Related papers (2020-12-31T18:17:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.