Variational Quantum Classifiers for Natural-Language Text
- URL: http://arxiv.org/abs/2303.02469v1
- Date: Sat, 4 Mar 2023 18:00:05 GMT
- Title: Variational Quantum Classifiers for Natural-Language Text
- Authors: Daniel T. Chang
- Abstract summary: We discuss three potential approaches to variational quantum text classifiers (VQTCs)
The first is a weighted bag-of-sentences approach which treats text as a group of independent sentences with task-specific sentence weighting.
The second is a coreference resolution approach which treats text as a consolidation of its member sentences with coreferences among them resolved.
The third approach, on the other hand, is based on the DisCoCirc model which considers both ordering of sentences and interaction of words in composing text meaning from word and sentence meanings.
- Score: 0.8722210937404288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As part of the recent research effort on quantum natural language processing
(QNLP), variational quantum sentence classifiers (VQSCs) have been implemented
and supported in lambeq / DisCoPy, based on the DisCoCat model of sentence
meaning. We discuss in some detail VQSCs, including category theory, DisCoCat
for modeling sentence as string diagram, and DisCoPy for encoding string
diagram as parameterized quantum circuit. Many NLP tasks, however, require the
handling of text consisting of multiple sentences, which is not supported in
lambeq / DisCoPy. A good example is sentiment classification of customer
feedback or product review. We discuss three potential approaches to
variational quantum text classifiers (VQTCs), in line with VQSCs. The first is
a weighted bag-of-sentences approach which treats text as a group of
independent sentences with task-specific sentence weighting. The second is a
coreference resolution approach which treats text as a consolidation of its
member sentences with coreferences among them resolved. Both approaches are
based on the DisCoCat model and should be implementable in lambeq / DisCoCat.
The third approach, on the other hand, is based on the DisCoCirc model which
considers both ordering of sentences and interaction of words in composing text
meaning from word and sentence meanings. DisCoCirc makes fundamental
modification of DisCoCat since a sentence in DisCoCirc updates meanings of
words, whereas all meanings are static in DisCoCat. It is not clear if
DisCoCirc can be implemented in lambeq / DisCoCat without breaking DisCoCat.
Related papers
- A Pipeline For Discourse Circuits From CCG [0.0]
DisCoCirc represents natural language text as a circuit' that captures the core semantic information of the text.
DisCoCirc fulfils another major aim of providing an NLP model that can be implemented on near-term quantum computers.
arXiv Detail & Related papers (2023-11-29T18:46:29Z) - CTC-based Non-autoregressive Speech Translation [51.37920141751813]
We investigate the potential of connectionist temporal classification for non-autoregressive speech translation.
We develop a model consisting of two encoders that are guided by CTC to predict the source and target texts.
Experiments on the MuST-C benchmarks show that our NAST model achieves an average BLEU score of 29.5 with a speed-up of 5.67$times$.
arXiv Detail & Related papers (2023-05-27T03:54:09Z) - Hierarchical Phrase-based Sequence-to-Sequence Learning [94.10257313923478]
We describe a neural transducer that maintains the flexibility of standard sequence-to-sequence (seq2seq) models while incorporating hierarchical phrases as a source of inductive bias during training and as explicit constraints during inference.
Our approach trains two models: a discriminative derivation based on a bracketing grammar whose tree hierarchically aligns source and target phrases, and a neural seq2seq model that learns to translate the aligned phrases one-by-one.
arXiv Detail & Related papers (2022-11-15T05:22:40Z) - CTC Alignments Improve Autoregressive Translation [145.90587287444976]
We argue that CTC does in fact make sense for translation if applied in a joint CTC/attention framework.
Our proposed joint CTC/attention models outperform pure-attention baselines across six benchmark translation tasks.
arXiv Detail & Related papers (2022-10-11T07:13:50Z) - A multiclass Q-NLP sentiment analysis experiment using DisCoCat [0.0]
We will tackle sentiment analysis in the Noisy Intermediate Scale Computing (NISQ) era, using the DisCoCat model of language.
We will first present the basics of quantum computing and the DisCoCat model.
This will enable us to define a general framework to perform NLP tasks on a quantum computer.
arXiv Detail & Related papers (2022-09-07T13:47:35Z) - Design and Implementation of a Quantum Kernel for Natural Language
Processing [0.8702432681310401]
This thesis leverage the DisCoCat model to design a quantum-based kernel function that can be used by a support vector machine (SVM) for NLP tasks.
Two similarity measures were studied: (i) the transition amplitude approach and (ii) the SWAP test.
The explicit model from previous work was used to train word embeddings and achieved a testing accuracy of $93.09 pm 0.01$%.
arXiv Detail & Related papers (2022-05-13T00:45:46Z) - A CCG-Based Version of the DisCoCat Framework [1.7219938668142956]
The DisCoCat model is used to study compositional aspects of language at the level of semantics.
In this paper we reformulating DisCoCat as a passage from Combinatory Categorial Grammar (CCG)
We show that standard categorial grammars can be expressed as a biclosed category, where all rules emerge as currying/uncurrying the identity.
We then proceed to model permutation-inducing rules by exploiting the symmetry of the compact closed category encoding the word meaning.
arXiv Detail & Related papers (2021-05-17T10:32:18Z) - Intermediate Loss Regularization for CTC-based Speech Recognition [58.33721897180646]
We present a simple and efficient auxiliary loss function for automatic speech recognition (ASR) based on the connectionist temporal classification ( CTC) objective.
We evaluate the proposed method on various corpora, reaching word error rate (WER) 9.9% on the WSJ corpus and character error rate (CER) 5.2% on the AISHELL-1 corpus respectively.
arXiv Detail & Related papers (2021-02-05T15:01:03Z) - Counterfactual Variable Control for Robust and Interpretable Question
Answering [57.25261576239862]
Deep neural network based question answering (QA) models are neither robust nor explainable in many cases.
In this paper, we inspect such spurious "capability" of QA models using causal inference.
We propose a novel approach called Counterfactual Variable Control (CVC) that explicitly mitigates any shortcut correlation.
arXiv Detail & Related papers (2020-10-12T10:09:05Z) - Extending C++ for Heterogeneous Quantum-Classical Computing [56.782064931823015]
qcor is a language extension to C++ and compiler implementation that enables heterogeneous quantum-classical programming, compilation, and execution in a single-source context.
Our work provides a first-of-its-kind C++ compiler enabling high-level quantum kernel (function) expression in a quantum-language manner.
arXiv Detail & Related papers (2020-10-08T12:49:07Z) - Meaning updating of density matrices [0.0]
The DisCoCat model of natural language meaning assigns meaning to a sentence given: (i) the meanings of its words, and (ii) its grammatical structure.
The recently introduced DisCoCirc model extends this to text consisting of multiple sentences.
While in DisCoCat all meanings are fixed, in DisCoCirc each sentence updates meanings of words.
arXiv Detail & Related papers (2020-01-03T15:28:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.