An enriched category theory of language: from syntax to semantics
- URL: http://arxiv.org/abs/2106.07890v1
- Date: Tue, 15 Jun 2021 05:40:51 GMT
- Title: An enriched category theory of language: from syntax to semantics
- Authors: Tai-Danae Bradley, John Terilla, Yiannis Vlassopoulos
- Abstract summary: We model probability distributions on texts as a category enriched over the unit interval.
We then pass to the enriched category of unit interval-valued copresheaves on this syntactical category to find semantic information.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Given a piece of text, the ability to generate a coherent extension of it
implies some sophistication, including a knowledge of grammar and semantics. In
this paper, we propose a mathematical framework for passing from probability
distributions on extensions of given texts to an enriched category containing
semantic information. Roughly speaking, we model probability distributions on
texts as a category enriched over the unit interval. Objects of this category
are expressions in language and hom objects are conditional probabilities that
one expression is an extension of another. This category is syntactical: it
describes what goes with what. We then pass to the enriched category of unit
interval-valued copresheaves on this syntactical category to find semantic
information.
Related papers
- AttrSeg: Open-Vocabulary Semantic Segmentation via Attribute
Decomposition-Aggregation [33.25304533086283]
Open-vocabulary semantic segmentation is a challenging task that requires segmenting novel object categories at inference time.
Recent studies have explored vision-language pre-training to handle this task, but suffer from unrealistic assumptions in practical scenarios.
This work proposes a novel attribute decomposition-aggregation framework, AttrSeg, inspired by human cognition in understanding new concepts.
arXiv Detail & Related papers (2023-08-31T19:34:09Z) - Primitive Generation and Semantic-related Alignment for Universal
Zero-Shot Segmentation [13.001629605405954]
We study universal zero-shot segmentation in this work to achieve panoptic, instance, and semantic segmentation for novel categories without any training samples.
We introduce a generative model to synthesize features for unseen categories, which links semantic and visual spaces.
The proposed approach achieves impressively state-of-the-art performance on zero-shot panoptic segmentation, instance segmentation, and semantic segmentation.
arXiv Detail & Related papers (2023-06-19T17:59:16Z) - Advancing Incremental Few-shot Semantic Segmentation via Semantic-guided
Relation Alignment and Adaptation [98.51938442785179]
Incremental few-shot semantic segmentation aims to incrementally extend a semantic segmentation model to novel classes.
This task faces a severe semantic-aliasing issue between base and novel classes due to data imbalance.
We propose the Semantic-guided Relation Alignment and Adaptation (SRAA) method that fully considers the guidance of prior semantic information.
arXiv Detail & Related papers (2023-05-18T10:40:52Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - Generalized Funnelling: Ensemble Learning and Heterogeneous Document
Embeddings for Cross-Lingual Text Classification [78.83284164605473]
emphFunnelling (Fun) is a recently proposed method for cross-lingual text classification.
We describe emphGeneralized Funnelling (gFun) as a generalization of Fun.
We show that gFun substantially improves over Fun and over state-of-the-art baselines.
arXiv Detail & Related papers (2021-09-17T23:33:04Z) - Evaluating Word Embeddings with Categorical Modularity [0.0]
We introduce categorical modularity, a novel low-resource metric to evaluate word embedding quality.
We use a core set of 500 words belonging to 59 neurobiologically motivated semantic categories in 29 languages.
We find moderate to strong positive correlations between categorical modularity and performance on the monolingual tasks of sentiment analysis and word similarity calculation.
arXiv Detail & Related papers (2021-06-02T01:29:11Z) - Semantic Analysis for Automated Evaluation of the Potential Impact of
Research Articles [62.997667081978825]
This paper presents a novel method for vector representation of text meaning based on information theory.
We show how this informational semantics is used for text classification on the basis of the Leicester Scientific Corpus.
We show that an informational approach to representing the meaning of a text has offered a way to effectively predict the scientific impact of research papers.
arXiv Detail & Related papers (2021-04-26T20:37:13Z) - Language Modeling with Reduced Densities [0.0]
We show that sequences of symbols from a finite alphabet, such as those found in a corpus of text, form a category enriched over probabilities.
We then address a second fundamental question: How can this information be stored and modeled in a way that preserves the categorical structure?
arXiv Detail & Related papers (2020-07-08T00:41:53Z) - Categorical Vector Space Semantics for Lambek Calculus with a Relevant
Modality [3.345437353879255]
We develop a categorical distributional semantics for Lambek Calculus with a Relevantity!L*.
We instantiate this category to finite dimensional vector spaces and linear maps via "quantisation" functors.
We apply the model to construct categorical and concrete semantic interpretations for the motivating example of!L*: the derivation of a phrase with a parasitic gap.
arXiv Detail & Related papers (2020-05-06T18:58:21Z) - Latent Embedding Feedback and Discriminative Features for Zero-Shot
Classification [139.44681304276]
zero-shot learning aims to classify unseen categories for which no data is available during training.
Generative Adrial Networks synthesize unseen class features by leveraging class-specific semantic embeddings.
We propose to enforce semantic consistency at all stages of zero-shot learning: training, feature synthesis and classification.
arXiv Detail & Related papers (2020-03-17T17:34:16Z) - Don't Judge an Object by Its Context: Learning to Overcome Contextual
Bias [113.44471186752018]
Existing models often leverage co-occurrences between objects and their context to improve recognition accuracy.
This work focuses on addressing such contextual biases to improve the robustness of the learnt feature representations.
arXiv Detail & Related papers (2020-01-09T18:31:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.