Transcending the Attention Paradigm: Representation Learning from
Geospatial Social Media Data
- URL: http://arxiv.org/abs/2310.05378v3
- Date: Sun, 14 Jan 2024 09:29:18 GMT
- Title: Transcending the Attention Paradigm: Representation Learning from
Geospatial Social Media Data
- Authors: Nick DiSanto, Anthony Corso, Benjamin Sanders, Gavin Harding
- Abstract summary: This study challenges the paradigm of performance benchmarking by investigating social media data as a source of distributed patterns.
To properly represent these abstract relationships, this research dissects empirical social media corpora into their elemental components, analyzing over two billion tweets across population-dense locations.
- Score: 1.8311821879979955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: While transformers have pioneered attention-driven architectures as a
cornerstone of language modeling, their dependence on explicitly contextual
information underscores limitations in their abilities to tacitly learn
overarching textual themes. This study challenges the heuristic paradigm of
performance benchmarking by investigating social media data as a source of
distributed patterns. In stark contrast to networks that rely on capturing
complex long-term dependencies, models of online data inherently lack structure
and are forced to detect latent structures in the aggregate. To properly
represent these abstract relationships, this research dissects empirical social
media corpora into their elemental components, analyzing over two billion
tweets across population-dense locations. We create Bag-of-Word embedding
specific to each city and compare their respective representations. This finds
that even amidst noisy data, geographic location has a considerable influence
on online communication, and that hidden insights can be uncovered without the
crutch of advanced algorithms. This evidence presents valuable geospatial
implications in social science and challenges the notion that intricate models
are prerequisites for pattern recognition in natural language. This aligns with
the evolving landscape that questions the embrace of absolute interpretability
over abstract understanding and bridges the divide between sophisticated
frameworks and intangible relationships.
Related papers
- The Ontoverse: Democratising Access to Knowledge Graph-based Data Through a Cartographic Interface [33.861478826378054]
We have developed a unique approach to data navigation that leans on geographical visualisation and hierarchically structured domain knowledge.
Our approach uses natural language processing techniques to extract named entities from the underlying data and normalise them against relevant semantic domain references and navigational structures.
This allows end-users to identify entities relevant to their needs and access extensive graph analytics.
arXiv Detail & Related papers (2024-07-22T10:29:25Z) - Bridging Local Details and Global Context in Text-Attributed Graphs [62.522550655068336]
GraphBridge is a framework that bridges local and global perspectives by leveraging contextual textual information.
Our method achieves state-of-theart performance, while our graph-aware token reduction module significantly enhances efficiency and solves scalability issues.
arXiv Detail & Related papers (2024-06-18T13:35:25Z) - How Deep Networks Learn Sparse and Hierarchical Data: the Sparse Random Hierarchy Model [4.215221129670858]
We show that by introducing sparsity to generative hierarchical models of data, the task acquires insensitivity to spatial transformations that are discrete versions of smooth transformations.
We quantify how the sample complexity of CNNs learning the SRHM depends on both the sparsity and hierarchical structure of the task.
arXiv Detail & Related papers (2024-04-16T17:01:27Z) - How Well Do Text Embedding Models Understand Syntax? [50.440590035493074]
The ability of text embedding models to generalize across a wide range of syntactic contexts remains under-explored.
Our findings reveal that existing text embedding models have not sufficiently addressed these syntactic understanding challenges.
We propose strategies to augment the generalization ability of text embedding models in diverse syntactic scenarios.
arXiv Detail & Related papers (2023-11-14T08:51:00Z) - Homological Convolutional Neural Networks [4.615338063719135]
We propose a novel deep learning architecture that exploits the data structural organization through topologically constrained network representations.
We test our model on 18 benchmark datasets against 5 classic machine learning and 3 deep learning models.
arXiv Detail & Related papers (2023-08-26T08:48:51Z) - Foundational Models Defining a New Era in Vision: A Survey and Outlook [151.49434496615427]
Vision systems to see and reason about the compositional nature of visual scenes are fundamental to understanding our world.
The models learned to bridge the gap between such modalities coupled with large-scale training data facilitate contextual reasoning, generalization, and prompt capabilities at test time.
The output of such models can be modified through human-provided prompts without retraining, e.g., segmenting a particular object by providing a bounding box, having interactive dialogues by asking questions about an image or video scene or manipulating the robot's behavior through language instructions.
arXiv Detail & Related papers (2023-07-25T17:59:18Z) - Variational Cross-Graph Reasoning and Adaptive Structured Semantics
Learning for Compositional Temporal Grounding [143.5927158318524]
Temporal grounding is the task of locating a specific segment from an untrimmed video according to a query sentence.
We introduce a new Compositional Temporal Grounding task and construct two new dataset splits.
We argue that the inherent structured semantics inside the videos and language is the crucial factor to achieve compositional generalization.
arXiv Detail & Related papers (2023-01-22T08:02:23Z) - TeKo: Text-Rich Graph Neural Networks with External Knowledge [75.91477450060808]
We propose a novel text-rich graph neural network with external knowledge (TeKo)
We first present a flexible heterogeneous semantic network that incorporates high-quality entities.
We then introduce two types of external knowledge, that is, structured triplets and unstructured entity description.
arXiv Detail & Related papers (2022-06-15T02:33:10Z) - Learning Attention-based Representations from Multiple Patterns for
Relation Prediction in Knowledge Graphs [2.4028383570062606]
AEMP is a novel model for learning contextualized representations by acquiring entities' context information.
AEMP either outperforms or competes with state-of-the-art relation prediction methods.
arXiv Detail & Related papers (2022-06-07T10:53:35Z) - Understanding Spatial Relations through Multiple Modalities [78.07328342973611]
spatial relations between objects can either be explicit -- expressed as spatial prepositions, or implicit -- expressed by spatial verbs such as moving, walking, shifting, etc.
We introduce the task of inferring implicit and explicit spatial relations between two entities in an image.
We design a model that uses both textual and visual information to predict the spatial relations, making use of both positional and size information of objects and image embeddings.
arXiv Detail & Related papers (2020-07-19T01:35:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.