A Pipeline For Discourse Circuits From CCG
- URL: http://arxiv.org/abs/2311.17892v1
- Date: Wed, 29 Nov 2023 18:46:29 GMT
- Title: A Pipeline For Discourse Circuits From CCG
- Authors: Jonathon Liu, Razin A. Shaikh, Benjamin Rodatz, Richie Yeung and Bob
Coecke
- Abstract summary: DisCoCirc represents natural language text as a circuit' that captures the core semantic information of the text.
DisCoCirc fulfils another major aim of providing an NLP model that can be implemented on near-term quantum computers.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There is a significant disconnect between linguistic theory and modern NLP
practice, which relies heavily on inscrutable black-box architectures.
DisCoCirc is a newly proposed model for meaning that aims to bridge this
divide, by providing neuro-symbolic models that incorporate linguistic
structure. DisCoCirc represents natural language text as a `circuit' that
captures the core semantic information of the text. These circuits can then be
interpreted as modular machine learning models. Additionally, DisCoCirc fulfils
another major aim of providing an NLP model that can be implemented on
near-term quantum computers.
In this paper we describe a software pipeline that converts English text to
its DisCoCirc representation. The pipeline achieves coverage over a large
fragment of the English language. It relies on Combinatory Categorial Grammar
(CCG) parses of the input text as well as coreference resolution information.
This semantic and syntactic information is used in several steps to convert the
text into a simply-typed $\lambda$-calculus term, and then into a circuit
diagram. This pipeline will enable the application of the DisCoCirc framework
to NLP tasks, using both classical and quantum approaches.
Related papers
- CoSTA: Code-Switched Speech Translation using Aligned Speech-Text Interleaving [61.73180469072787]
We focus on the problem of spoken translation (ST) of code-switched speech in Indian languages to English text.
We present a new end-to-end model architecture COSTA that scaffolds on pretrained automatic speech recognition (ASR) and machine translation (MT) modules.
COSTA significantly outperforms many competitive cascaded and end-to-end multimodal baselines by up to 3.5 BLEU points.
arXiv Detail & Related papers (2024-06-16T16:10:51Z) - Language-Oriented Communication with Semantic Coding and Knowledge
Distillation for Text-to-Image Generation [53.97155730116369]
We put forward a novel framework of language-oriented semantic communication (LSC)
In LSC, machines communicate using human language messages that can be interpreted and manipulated via natural language processing (NLP) techniques for SC efficiency.
We introduce three innovative algorithms: 1) semantic source coding (SSC), which compresses a text prompt into its key head words capturing the prompt's syntactic essence; 2) semantic channel coding ( SCC), that improves robustness against errors by substituting head words with their lenghthier synonyms; and 3) semantic knowledge distillation (SKD), that produces listener-customized prompts via in-context learning the listener's
arXiv Detail & Related papers (2023-09-20T08:19:05Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - CTC-based Non-autoregressive Speech Translation [51.37920141751813]
We investigate the potential of connectionist temporal classification for non-autoregressive speech translation.
We develop a model consisting of two encoders that are guided by CTC to predict the source and target texts.
Experiments on the MuST-C benchmarks show that our NAST model achieves an average BLEU score of 29.5 with a speed-up of 5.67$times$.
arXiv Detail & Related papers (2023-05-27T03:54:09Z) - Variational Quantum Classifiers for Natural-Language Text [0.8722210937404288]
We discuss three potential approaches to variational quantum text classifiers (VQTCs)
The first is a weighted bag-of-sentences approach which treats text as a group of independent sentences with task-specific sentence weighting.
The second is a coreference resolution approach which treats text as a consolidation of its member sentences with coreferences among them resolved.
The third approach, on the other hand, is based on the DisCoCirc model which considers both ordering of sentences and interaction of words in composing text meaning from word and sentence meanings.
arXiv Detail & Related papers (2023-03-04T18:00:05Z) - Category Theory for Quantum Natural Language Processing [0.0]
This thesis introduces quantum natural language processing (QNLP) models based on an analogy between computational linguistics and quantum mechanics.
The grammatical structure of text and sentences connects the meaning of words in the same way that entanglement structure connects the states of quantum systems.
We turn this abstract analogy into a concrete algorithm that translates the grammatical structure onto the architecture of parameterised quantum circuits.
We then use a hybrid classical-quantum algorithm to train the model so that evaluating the circuits computes the meaning of sentences in data-driven tasks.
arXiv Detail & Related papers (2022-12-13T14:38:57Z) - A multiclass Q-NLP sentiment analysis experiment using DisCoCat [0.0]
We will tackle sentiment analysis in the Noisy Intermediate Scale Computing (NISQ) era, using the DisCoCat model of language.
We will first present the basics of quantum computing and the DisCoCat model.
This will enable us to define a general framework to perform NLP tasks on a quantum computer.
arXiv Detail & Related papers (2022-09-07T13:47:35Z) - End-to-End Speech Translation for Code Switched Speech [13.97982457879585]
Code switching (CS) refers to the phenomenon of interchangeably using words and phrases from different languages.
We focus on CS in the context of English/Spanish conversations for the task of speech translation (ST), generating and evaluating both transcript and translation.
We show that our ST architectures, and especially our bidirectional end-to-end architecture, perform well on CS speech, even when no CS training data is used.
arXiv Detail & Related papers (2022-04-11T13:25:30Z) - Knowledge Transfer from Large-scale Pretrained Language Models to
End-to-end Speech Recognizers [13.372686722688325]
Training of end-to-end speech recognizers always requires transcribed utterances.
This paper proposes a method for alleviating this issue by transferring knowledge from a language model neural network that can be pretrained with text-only data.
arXiv Detail & Related papers (2022-02-16T07:02:24Z) - Coreferential Reasoning Learning for Language Representation [88.14248323659267]
We present CorefBERT, a novel language representation model that can capture the coreferential relations in context.
The experimental results show that, compared with existing baseline models, CorefBERT can achieve significant improvements consistently on various downstream NLP tasks.
arXiv Detail & Related papers (2020-04-15T03:57:45Z) - Bi-Decoder Augmented Network for Neural Machine Translation [108.3931242633331]
We propose a novel Bi-Decoder Augmented Network (BiDAN) for the neural machine translation task.
Since each decoder transforms the representations of the input text into its corresponding language, jointly training with two target ends can make the shared encoder has the potential to produce a language-independent semantic space.
arXiv Detail & Related papers (2020-01-14T02:05:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.