Multi-lingual agents through multi-headed neural networks
- URL: http://arxiv.org/abs/2111.11129v1
- Date: Mon, 22 Nov 2021 11:39:42 GMT
- Title: Multi-lingual agents through multi-headed neural networks
- Authors: J. D. Thomas, R. Santos-Rodr\'iguez, R. Piechocki, M. Anca
- Abstract summary: This paper focuses on cooperative Multi-Agent Reinforcement Learning.
In this context, multiple distinct and incompatible languages can emerge.
We take inspiration from the Continual Learning literature and equip our agents with multi-headed neural networks which enable our agents to be multi-lingual.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper considers cooperative Multi-Agent Reinforcement Learning, focusing
on emergent communication in settings where multiple pairs of independent
learners interact at varying frequencies. In this context, multiple distinct
and incompatible languages can emerge. When an agent encounters a speaker of an
alternative language, there is a requirement for a period of adaptation before
they can efficiently converse. This adaptation results in the emergence of a
new language and the forgetting of the previous language. In principle, this is
an example of the Catastrophic Forgetting problem which can be mitigated by
enabling the agents to learn and maintain multiple languages. We take
inspiration from the Continual Learning literature and equip our agents with
multi-headed neural networks which enable our agents to be multi-lingual. Our
method is empirically validated within a referential MNIST based communication
game and is shown to be able to maintain multiple languages where existing
approaches cannot.
Related papers
- Lens: Rethinking Multilingual Enhancement for Large Language Models [70.85065197789639]
Lens is a novel approach to enhance multilingual capabilities of large language models (LLMs)
It operates by manipulating the hidden representations within the language-agnostic and language-specific subspaces from top layers of LLMs.
It achieves superior results with much fewer computational resources compared to existing post-training approaches.
arXiv Detail & Related papers (2024-10-06T08:51:30Z) - Bidirectional Emergent Language in Situated Environments [4.950411915351642]
We introduce two novel cooperative environments: Multi-Agent Pong and Collectors.
optimal performance requires the emergence of a communication protocol, but moderate success can be achieved without one.
We find that the emerging communication is sparse, with the agents only generating meaningful messages and acting upon incoming messages in states where they cannot succeed without coordination.
arXiv Detail & Related papers (2024-08-26T21:25:44Z) - A Roadmap for Multilingual, Multimodal Domain Independent Deception Detection [2.1506382989223782]
Deception, a prevalent aspect of human communication, has undergone a significant transformation in the digital age.
Recent studies have shown the possibility of the existence of universal linguistic cues to deception across domains within the English language.
The practical task of deception detection in low-resource languages is not a well-studied problem due to the lack of labeled data.
arXiv Detail & Related papers (2024-05-07T00:38:34Z) - Communication Drives the Emergence of Language Universals in Neural
Agents: Evidence from the Word-order/Case-marking Trade-off [3.631024220680066]
We propose a new Neural-agent Language Learning and Communication framework (NeLLCom) where pairs of speaking and listening agents first learn a miniature language.
We succeed in replicating the trade-off with the new framework without hard-coding specific biases in the agents.
arXiv Detail & Related papers (2023-01-30T17:22:33Z) - On the cross-lingual transferability of multilingual prototypical models
across NLU tasks [2.44288434255221]
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven to be effective for limited domain and language applications.
In practice, these approaches suffer from the drawbacks of domain-driven design and under-resourced languages.
This article proposes to investigate the cross-lingual transferability of using synergistically few-shot learning with prototypical neural networks and multilingual Transformers-based models.
arXiv Detail & Related papers (2022-07-19T09:55:04Z) - Learning to Ground Multi-Agent Communication with Autoencoders [43.22048280036316]
Communication requires a common language, a lingua franca, between agents.
We demonstrate a simple way to ground language in learned representations.
We find that a standard representation learning algorithm is sufficient for arriving at a grounded common language.
arXiv Detail & Related papers (2021-10-28T17:57:26Z) - Discovering Representation Sprachbund For Multilingual Pre-Training [139.05668687865688]
We generate language representation from multilingual pre-trained models and conduct linguistic analysis.
We cluster all the target languages into multiple groups and name each group as a representation sprachbund.
Experiments are conducted on cross-lingual benchmarks and significant improvements are achieved compared to strong baselines.
arXiv Detail & Related papers (2021-09-01T09:32:06Z) - Are Multilingual Models Effective in Code-Switching? [57.78477547424949]
We study the effectiveness of multilingual language models to understand their capability and adaptability to the mixed-language setting.
Our findings suggest that pre-trained multilingual models do not necessarily guarantee high-quality representations on code-switching.
arXiv Detail & Related papers (2021-03-24T16:20:02Z) - VECO: Variable and Flexible Cross-lingual Pre-training for Language
Understanding and Generation [77.82373082024934]
We plug a cross-attention module into the Transformer encoder to explicitly build the interdependence between languages.
It can effectively avoid the degeneration of predicting masked words only conditioned on the context in its own language.
The proposed cross-lingual model delivers new state-of-the-art results on various cross-lingual understanding tasks of the XTREME benchmark.
arXiv Detail & Related papers (2020-10-30T03:41:38Z) - Gender Bias in Multilingual Embeddings and Cross-Lingual Transfer [101.58431011820755]
We study gender bias in multilingual embeddings and how it affects transfer learning for NLP applications.
We create a multilingual dataset for bias analysis and propose several ways for quantifying bias in multilingual representations.
arXiv Detail & Related papers (2020-05-02T04:34:37Z) - Bridging Linguistic Typology and Multilingual Machine Translation with
Multi-View Language Representations [83.27475281544868]
We use singular vector canonical correlation analysis to study what kind of information is induced from each source.
We observe that our representations embed typology and strengthen correlations with language relationships.
We then take advantage of our multi-view language vector space for multilingual machine translation, where we achieve competitive overall translation accuracy.
arXiv Detail & Related papers (2020-04-30T16:25:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.