Reinforcement learning of minimalist grammars
- URL: http://arxiv.org/abs/2005.00359v1
- Date: Thu, 30 Apr 2020 14:25:58 GMT
- Title: Reinforcement learning of minimalist grammars
- Authors: Peter beim Graben, Ronald R\"omer, Werner Meyer, Markus Huber,
Matthias Wolff
- Abstract summary: State-of-the-art language technology scans the acoustically analyzed speech signal for relevant keywords.
Words are then inserted into semantic slots to interpret the user's intent.
A mental lexicon must be acquired by a cognitive agent during interaction with its users.
- Score: 0.5862282909017474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Speech-controlled user interfaces facilitate the operation of devices and
household functions to laymen. State-of-the-art language technology scans the
acoustically analyzed speech signal for relevant keywords that are subsequently
inserted into semantic slots to interpret the user's intent. In order to
develop proper cognitive information and communication technologies, simple
slot-filling should be replaced by utterance meaning transducers (UMT) that are
based on semantic parsers and a mental lexicon, comprising syntactic, phonetic
and semantic features of the language under consideration. This lexicon must be
acquired by a cognitive agent during interaction with its users. We outline a
reinforcement learning algorithm for the acquisition of syntax and semantics of
English utterances, based on minimalist grammar (MG), a recent computational
implementation of generative linguistics. English declarative sentences are
presented to the agent by a teacher in form of utterance meaning pairs (UMP)
where the meanings are encoded as formulas of predicate logic. Since MG
codifies universal linguistic competence through inference rules, thereby
separating innate linguistic knowledge from the contingently acquired lexicon,
our approach unifies generative grammar and reinforcement learning, hence
potentially resolving the still pending Chomsky-Skinner controversy.
Related papers
- Principles of semantic and functional efficiency in grammatical patterning [1.6267479602370545]
Grammatical features such as number and gender serve two central functions in human languages.
Number and gender encode salient semantic attributes like numerosity and animacy, but offload sentence processing cost by predictably linking words together.
Grammars exhibit consistent organizational patterns across diverse languages, invariably rooted in a semantic foundation.
arXiv Detail & Related papers (2024-10-21T10:49:54Z) - Word Sense Disambiguation in Native Spanish: A Comprehensive Lexical Evaluation Resource [2.7775559369441964]
A lexical meaning of a word in context can be determined automatically by Word Sense Disambiguation (WSD) algorithms.
This study introduces a new resource for Spanish WSD.
It includes a sense inventory and a lexical dataset sourced from the Diccionario de la Lengua Espanola.
arXiv Detail & Related papers (2024-09-30T17:22:33Z) - Pixel Sentence Representation Learning [67.4775296225521]
In this work, we conceptualize the learning of sentence-level textual semantics as a visual representation learning process.
We employ visually-grounded text perturbation methods like typos and word order shuffling, resonating with human cognitive patterns, and enabling perturbation to be perceived as continuous.
Our approach is further bolstered by large-scale unsupervised topical alignment training and natural language inference supervision.
arXiv Detail & Related papers (2024-02-13T02:46:45Z) - Leveraging Language ID to Calculate Intermediate CTC Loss for Enhanced
Code-Switching Speech Recognition [5.3545957730615905]
We introduce language identification information into the middle layer of the ASR model's encoder.
We aim to generate acoustic features that imply language distinctions in a more implicit way, reducing the model's confusion when dealing with language switching.
arXiv Detail & Related papers (2023-12-15T07:46:35Z) - Language-Oriented Communication with Semantic Coding and Knowledge
Distillation for Text-to-Image Generation [53.97155730116369]
We put forward a novel framework of language-oriented semantic communication (LSC)
In LSC, machines communicate using human language messages that can be interpreted and manipulated via natural language processing (NLP) techniques for SC efficiency.
We introduce three innovative algorithms: 1) semantic source coding (SSC), which compresses a text prompt into its key head words capturing the prompt's syntactic essence; 2) semantic channel coding ( SCC), that improves robustness against errors by substituting head words with their lenghthier synonyms; and 3) semantic knowledge distillation (SKD), that produces listener-customized prompts via in-context learning the listener's
arXiv Detail & Related papers (2023-09-20T08:19:05Z) - An Empirical Revisiting of Linguistic Knowledge Fusion in Language
Understanding Tasks [33.765874588342285]
Infusing language models with syntactic or semantic knowledge from structural linguistic priors has shown improvements on many language understanding tasks.
We conduct empirical study of replacing parsed graphs or trees with trivial ones for tasks in the GLUE benchmark.
It reveals that the gains might not be significantly attributed to explicit linguistic priors but rather to more feature interactions brought by fusion layers.
arXiv Detail & Related papers (2022-10-24T07:47:32Z) - Code-Switching without Switching: Language Agnostic End-to-End Speech
Translation [68.8204255655161]
We treat speech recognition and translation as one unified end-to-end speech translation problem.
By training LAST with both input languages, we decode speech into one target language, regardless of the input language.
arXiv Detail & Related papers (2022-10-04T10:34:25Z) - Wav-BERT: Cooperative Acoustic and Linguistic Representation Learning
for Low-Resource Speech Recognition [159.9312272042253]
Wav-BERT is a cooperative acoustic and linguistic representation learning method.
We unify a pre-trained acoustic model (wav2vec 2.0) and a language model (BERT) into an end-to-end trainable framework.
arXiv Detail & Related papers (2021-09-19T16:39:22Z) - Leveraging Acoustic and Linguistic Embeddings from Pretrained speech and
language Models for Intent Classification [81.80311855996584]
We propose a novel intent classification framework that employs acoustic features extracted from a pretrained speech recognition system and linguistic features learned from a pretrained language model.
We achieve 90.86% and 99.07% accuracy on ATIS and Fluent speech corpus, respectively.
arXiv Detail & Related papers (2021-02-15T07:20:06Z) - Semantics-Aware Inferential Network for Natural Language Understanding [79.70497178043368]
We propose a Semantics-Aware Inferential Network (SAIN) to meet such a motivation.
Taking explicit contextualized semantics as a complementary input, the inferential module of SAIN enables a series of reasoning steps over semantic clues.
Our model achieves significant improvement on 11 tasks including machine reading comprehension and natural language inference.
arXiv Detail & Related papers (2020-04-28T07:24:43Z) - Detect Language of Transliterated Texts [0.0]
Informal transliteration from other languages to English is prevalent in social media threads, instant messaging, and discussion forums.
We propose a Language Identification (LID) system, with an approach for feature extraction.
We tokenize the words into phonetic syllables and use a simple Long Short-term Memory (LSTM) network architecture to detect the language of transliterated texts.
arXiv Detail & Related papers (2020-04-26T10:28:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.