A CCG-Based Version of the DisCoCat Framework
- URL: http://arxiv.org/abs/2105.07720v2
- Date: Wed, 19 May 2021 11:05:47 GMT
- Title: A CCG-Based Version of the DisCoCat Framework
- Authors: Richie Yeung, Dimitri Kartsaklis
- Abstract summary: The DisCoCat model is used to study compositional aspects of language at the level of semantics.
In this paper we reformulating DisCoCat as a passage from Combinatory Categorial Grammar (CCG)
We show that standard categorial grammars can be expressed as a biclosed category, where all rules emerge as currying/uncurrying the identity.
We then proceed to model permutation-inducing rules by exploiting the symmetry of the compact closed category encoding the word meaning.
- Score: 1.7219938668142956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While the DisCoCat model (Coecke et al., 2010) has been proved a valuable
tool for studying compositional aspects of language at the level of semantics,
its strong dependency on pregroup grammars poses important restrictions: first,
it prevents large-scale experimentation due to the absence of a pregroup
parser; and second, it limits the expressibility of the model to context-free
grammars. In this paper we solve these problems by reformulating DisCoCat as a
passage from Combinatory Categorial Grammar (CCG) to a category of semantics.
We start by showing that standard categorial grammars can be expressed as a
biclosed category, where all rules emerge as currying/uncurrying the identity;
we then proceed to model permutation-inducing rules by exploiting the symmetry
of the compact closed category encoding the word meaning. We provide a proof of
concept for our method, converting "Alice in Wonderland" into DisCoCat form, a
corpus that we make available to the community.
Related papers
- Open-Vocabulary Segmentation with Semantic-Assisted Calibration [68.41025728960176]
We study open-vocabulary segmentation (OVS) through calibrating in-vocabulary and domain-biased embedding space with contextual prior of CLIP.
We present a Semantic-assisted CAlibration Network (SCAN) to achieve state-of-the-art performance on open-vocabulary segmentation benchmarks.
arXiv Detail & Related papers (2023-12-07T07:00:09Z) - Advancing Incremental Few-shot Semantic Segmentation via Semantic-guided
Relation Alignment and Adaptation [98.51938442785179]
Incremental few-shot semantic segmentation aims to incrementally extend a semantic segmentation model to novel classes.
This task faces a severe semantic-aliasing issue between base and novel classes due to data imbalance.
We propose the Semantic-guided Relation Alignment and Adaptation (SRAA) method that fully considers the guidance of prior semantic information.
arXiv Detail & Related papers (2023-05-18T10:40:52Z) - Harmonizing Base and Novel Classes: A Class-Contrastive Approach for
Generalized Few-Shot Segmentation [78.74340676536441]
We propose a class contrastive loss and a class relationship loss to regulate prototype updates and encourage a large distance between prototypes.
Our proposed approach achieves new state-of-the-art performance for the generalized few-shot segmentation task on PASCAL VOC and MS COCO datasets.
arXiv Detail & Related papers (2023-03-24T00:30:25Z) - Variational Quantum Classifiers for Natural-Language Text [0.8722210937404288]
We discuss three potential approaches to variational quantum text classifiers (VQTCs)
The first is a weighted bag-of-sentences approach which treats text as a group of independent sentences with task-specific sentence weighting.
The second is a coreference resolution approach which treats text as a consolidation of its member sentences with coreferences among them resolved.
The third approach, on the other hand, is based on the DisCoCirc model which considers both ordering of sentences and interaction of words in composing text meaning from word and sentence meanings.
arXiv Detail & Related papers (2023-03-04T18:00:05Z) - CCPrefix: Counterfactual Contrastive Prefix-Tuning for Many-Class
Classification [57.62886091828512]
We propose a brand-new prefix-tuning method, Counterfactual Contrastive Prefix-tuning (CCPrefix) for many-class classification.
Basically, an instance-dependent soft prefix, derived from fact-counterfactual pairs in the label space, is leveraged to complement the language verbalizers in many-class classification.
arXiv Detail & Related papers (2022-11-11T03:45:59Z) - Geometry-Aware Supertagging with Heterogeneous Dynamic Convolutions [0.7868449549351486]
We revisit constructive supertagging from a graph-theoretic perspective.
We propose a framework based on heterogeneous dynamic graph convolutions.
We test our approach on a number of categorial grammar datasets spanning different languages.
arXiv Detail & Related papers (2022-03-23T07:07:11Z) - Generalized Funnelling: Ensemble Learning and Heterogeneous Document
Embeddings for Cross-Lingual Text Classification [78.83284164605473]
emphFunnelling (Fun) is a recently proposed method for cross-lingual text classification.
We describe emphGeneralized Funnelling (gFun) as a generalization of Fun.
We show that gFun substantially improves over Fun and over state-of-the-art baselines.
arXiv Detail & Related papers (2021-09-17T23:33:04Z) - Exploiting a Joint Embedding Space for Generalized Zero-Shot Semantic
Segmentation [25.070027668717422]
Generalized zero-shot semantic segmentation (GZS3) predicts pixel-wise semantic labels for seen and unseen classes.
Most GZS3 methods adopt a generative approach that synthesizes visual features of unseen classes from corresponding semantic ones.
We propose a discriminative approach to address limitations in a unified framework.
arXiv Detail & Related papers (2021-08-14T13:33:58Z) - Towards Robustness Against Natural Language Word Substitutions [87.56898475512703]
Robustness against word substitutions has a well-defined and widely acceptable form, using semantically similar words as substitutions.
Previous defense methods capture word substitutions in vector space by using either $l$-ball or hyper-rectangle.
arXiv Detail & Related papers (2021-07-28T17:55:08Z) - Montague Grammar Induction [4.321645312120979]
This framework provides the analyst fine-grained control over the assumptions that the induced grammar should conform to.
We focus on the relationship between s(emantic)-selection and c(ategory)-selection, using as input a lexicon-scale acceptability judgment dataset.
arXiv Detail & Related papers (2020-10-15T23:25:01Z) - Traduction des Grammaires Cat\'egorielles de Lambek dans les Grammaires
Cat\'egorielles Abstraites [0.0]
This internship report is to demonstrate that every Lambek Grammar can be, not entirely but efficiently, expressed in Abstract Categorial Grammars (ACG)
The main idea is to transform the type rewriting system of LGs into that of Context-Free Grammars (CFG) by erasing introduction and elimination rules and generating enough axioms so that the cut rule suffices.
Although the underlying algorithm was not fully implemented, this proof provides another argument in favour of the relevance of ACGs in Natural Language Processing.
arXiv Detail & Related papers (2020-01-23T18:23:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.