Expressivity of Emergent Language is a Trade-off between Contextual
Complexity and Unpredictability
- URL: http://arxiv.org/abs/2106.03982v1
- Date: Mon, 7 Jun 2021 21:57:11 GMT
- Title: Expressivity of Emergent Language is a Trade-off between Contextual
Complexity and Unpredictability
- Authors: Shangmin Guo, Yi Ren, Kory Mathewson, Simon Kirby, Stefano V.
Albrecht, Kenny Smith
- Abstract summary: We propose a definition of partial order between expressivity based on the generalisation performance across different language games.
We also validate the hypothesis that expressivity of emergent languages is a trade-off between the complexity and unpredictability of the context.
We show that using our contrastive loss alleviates the collapse of message types seen using standard referential loss functions.
- Score: 7.765925231148388
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Researchers are now using deep learning models to explore the emergence of
language in various language games, where simulated agents interact and develop
an emergent language to solve a task. Although it is quite intuitive that
different types of language games posing different communicative challenges
might require emergent languages which encode different levels of information,
there is no existing work exploring the expressivity of the emergent languages.
In this work, we propose a definition of partial order between expressivity
based on the generalisation performance across different language games. We
also validate the hypothesis that expressivity of emergent languages is a
trade-off between the complexity and unpredictability of the context those
languages are used in. Our second novel contribution is introducing contrastive
loss into the implementation of referential games. We show that using our
contrastive loss alleviates the collapse of message types seen using standard
referential loss functions.
Related papers
- Multilingual Multi-Figurative Language Detection [14.799109368073548]
figurative language understanding is highly understudied in a multilingual setting.
We introduce multilingual multi-figurative language modelling, and provide a benchmark for sentence-level figurative language detection.
We develop a framework for figurative language detection based on template-based prompt learning.
arXiv Detail & Related papers (2023-05-31T18:52:41Z) - Transparency Helps Reveal When Language Models Learn Meaning [71.96920839263457]
Our systematic experiments with synthetic data reveal that, with languages where all expressions have context-independent denotations, both autoregressive and masked language models learn to emulate semantic relations between expressions.
Turning to natural language, our experiments with a specific phenomenon -- referential opacity -- add to the growing body of evidence that current language models do not well-represent natural language semantics.
arXiv Detail & Related papers (2022-10-14T02:35:19Z) - Testing the Ability of Language Models to Interpret Figurative Language [69.59943454934799]
Figurative and metaphorical language are commonplace in discourse.
It remains an open question to what extent modern language models can interpret nonliteral phrases.
We introduce Fig-QA, a Winograd-style nonliteral language understanding task.
arXiv Detail & Related papers (2022-04-26T23:42:22Z) - Linking Emergent and Natural Languages via Corpus Transfer [98.98724497178247]
We propose a novel way to establish a link by corpus transfer between emergent languages and natural languages.
Our approach showcases non-trivial transfer benefits for two different tasks -- language modeling and image captioning.
We also introduce a novel metric to predict the transferability of an emergent language by translating emergent messages to natural language captions grounded on the same images.
arXiv Detail & Related papers (2022-03-24T21:24:54Z) - Emergent Communication of Generalizations [13.14792537601313]
We argue that communicating about a single object in a shared visual context is prone to overfitting and does not encourage language useful beyond concrete reference.
We propose games that require communicating generalizations over sets of objects representing abstract visual concepts.
We find that these games greatly improve systematicity and interpretability of the learned languages.
arXiv Detail & Related papers (2021-06-04T19:02:18Z) - AM2iCo: Evaluating Word Meaning in Context across Low-ResourceLanguages
with Adversarial Examples [51.048234591165155]
We present AM2iCo, Adversarial and Multilingual Meaning in Context.
It aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts.
Results reveal that current SotA pretrained encoders substantially lag behind human performance.
arXiv Detail & Related papers (2021-04-17T20:23:45Z) - Inductive Bias and Language Expressivity in Emergent Communication [6.043034177891378]
We investigate how the type of the language game affects the emergent language.
We show that languages emerged from different games have different compositionality and further different expressivity.
arXiv Detail & Related papers (2020-12-04T22:20:55Z) - VECO: Variable and Flexible Cross-lingual Pre-training for Language
Understanding and Generation [77.82373082024934]
We plug a cross-attention module into the Transformer encoder to explicitly build the interdependence between languages.
It can effectively avoid the degeneration of predicting masked words only conditioned on the context in its own language.
The proposed cross-lingual model delivers new state-of-the-art results on various cross-lingual understanding tasks of the XTREME benchmark.
arXiv Detail & Related papers (2020-10-30T03:41:38Z) - Bridging Linguistic Typology and Multilingual Machine Translation with
Multi-View Language Representations [83.27475281544868]
We use singular vector canonical correlation analysis to study what kind of information is induced from each source.
We observe that our representations embed typology and strengthen correlations with language relationships.
We then take advantage of our multi-view language vector space for multilingual machine translation, where we achieve competitive overall translation accuracy.
arXiv Detail & Related papers (2020-04-30T16:25:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.