Cetacean Translation Initiative: a roadmap to deciphering the
communication of sperm whales
- URL: http://arxiv.org/abs/2104.08614v1
- Date: Sat, 17 Apr 2021 18:39:22 GMT
- Title: Cetacean Translation Initiative: a roadmap to deciphering the
communication of sperm whales
- Authors: Jacob Andreas, Ga\v{s}per Begu\v{s}, Michael M. Bronstein, Roee
Diamant, Denley Delaney, Shane Gero, Shafi Goldwasser, David F. Gruber, Sarah
de Haas, Peter Malkin, Roger Payne, Giovanni Petri, Daniela Rus, Pratyusha
Sharma, Dan Tchernov, Pernille T{\o}nnesen, Antonio Torralba, Daniel Vogt,
Robert J. Wood
- Abstract summary: Recent research showed the promise of machine learning tools for analyzing acoustic communication in nonhuman species.
We outline the key elements required for the collection and processing of massive bioacoustic data of sperm whales.
The technological capabilities developed are likely to yield cross-applications and advancements in broader communities investigating non-human communication and animal behavioral research.
- Score: 97.41394631426678
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The past decade has witnessed a groundbreaking rise of machine learning for
human language analysis, with current methods capable of automatically
accurately recovering various aspects of syntax and semantics - including
sentence structure and grounded word meaning - from large data collections.
Recent research showed the promise of such tools for analyzing acoustic
communication in nonhuman species. We posit that machine learning will be the
cornerstone of future collection, processing, and analysis of multimodal
streams of data in animal communication studies, including bioacoustic,
behavioral, biological, and environmental data. Cetaceans are unique non-human
model species as they possess sophisticated acoustic communications, but
utilize a very different encoding system that evolved in an aquatic rather than
terrestrial medium. Sperm whales, in particular, with their highly-developed
neuroanatomical features, cognitive abilities, social structures, and discrete
click-based encoding make for an excellent starting point for advanced machine
learning tools that can be applied to other animals in the future. This paper
details a roadmap toward this goal based on currently existing technology and
multidisciplinary scientific community effort. We outline the key elements
required for the collection and processing of massive bioacoustic data of sperm
whales, detecting their basic communication units and language-like
higher-level structures, and validating these models through interactive
playback experiments. The technological capabilities developed by such an
undertaking are likely to yield cross-applications and advancements in broader
communities investigating non-human communication and animal behavioral
research.
Related papers
- Multi Modal Information Fusion of Acoustic and Linguistic Data for Decoding Dairy Cow Vocalizations in Animal Welfare Assessment [0.0]
This study aims to decode dairy cow contact calls by employing multi-modal data fusion techniques.
We utilize the Natural Language Processing model to transcribe audio recordings of cow vocalizations into written form.
We categorized vocalizations into high frequency calls associated with distress or arousal, and low frequency calls linked to contentment or calmness.
arXiv Detail & Related papers (2024-11-01T09:48:30Z) - Multimodal Fusion with LLMs for Engagement Prediction in Natural Conversation [70.52558242336988]
We focus on predicting engagement in dyadic interactions by scrutinizing verbal and non-verbal cues, aiming to detect signs of disinterest or confusion.
In this work, we collect a dataset featuring 34 participants engaged in casual dyadic conversations, each providing self-reported engagement ratings at the end of each conversation.
We introduce a novel fusion strategy using Large Language Models (LLMs) to integrate multiple behavior modalities into a multimodal transcript''
arXiv Detail & Related papers (2024-09-13T18:28:12Z) - Feature Representations for Automatic Meerkat Vocalization Classification [15.642602544201308]
This paper investigates feature representations for automatic meerkat vocalization analysis.
Call type classification studies conducted on two data sets reveal that feature extraction methods developed for human speech processing can be effectively employed for automatic meerkat call analysis.
arXiv Detail & Related papers (2024-08-27T10:51:51Z) - Unveiling the pressures underlying language learning and use in neural networks, large language models, and humans: Lessons from emergent machine-to-machine communication [5.371337604556311]
We review three cases where mismatches between the emergent linguistic behavior of neural agents and humans were resolved.
We identify key pressures at play for language learning and emergence: communicative success, production effort, learnability, and other psycho-/sociolinguistic factors.
arXiv Detail & Related papers (2024-03-21T14:33:34Z) - Language Evolution with Deep Learning [49.879239655532324]
Computational modeling plays an essential role in the study of language emergence.
It aims to simulate the conditions and learning processes that could trigger the emergence of a structured language.
This chapter explores another class of computational models that have recently revolutionized the field of machine learning: deep learning models.
arXiv Detail & Related papers (2024-03-18T16:52:54Z) - OmniMotionGPT: Animal Motion Generation with Limited Data [70.35662376853163]
We introduce AnimalML3D, the first text-animal motion dataset with 1240 animation sequences spanning 36 different animal identities.
We are able to generate animal motions with high diversity and fidelity, quantitatively and qualitatively outperforming the results of training human motion generation baselines on animal data.
arXiv Detail & Related papers (2023-11-30T07:14:00Z) - Transferable Models for Bioacoustics with Human Language Supervision [0.0]
BioLingual is a new model for bioacoustics based on contrastive language-audio pretraining.
It can identify over a thousand species' calls across taxa, complete bioacoustic tasks zero-shot, and retrieve animal vocalization recordings from natural text queries.
arXiv Detail & Related papers (2023-08-09T14:22:18Z) - Co-Located Human-Human Interaction Analysis using Nonverbal Cues: A
Survey [71.43956423427397]
We aim to identify the nonverbal cues and computational methodologies resulting in effective performance.
This survey differs from its counterparts by involving the widest spectrum of social phenomena and interaction settings.
Some major observations are: the most often used nonverbal cue, computational method, interaction environment, and sensing approach are speaking activity, support vector machines, and meetings composed of 3-4 persons equipped with microphones and cameras, respectively.
arXiv Detail & Related papers (2022-07-20T13:37:57Z) - Seeing biodiversity: perspectives in machine learning for wildlife
conservation [49.15793025634011]
We argue that machine learning can meet this analytic challenge to enhance our understanding, monitoring capacity, and conservation of wildlife species.
In essence, by combining new machine learning approaches with ecological domain knowledge, animal ecologists can capitalize on the abundance of data generated by modern sensor technologies.
arXiv Detail & Related papers (2021-10-25T13:40:36Z) - Decoding EEG Brain Activity for Multi-Modal Natural Language Processing [9.35961671939495]
We present the first large-scale study of systematically analyzing the potential of EEG brain activity data for improving natural language processing tasks.
We find that filtering the EEG signals into frequency bands is more beneficial than using the broadband signal.
For a range of word embedding types, EEG data improves binary and ternary sentiment classification and outperforms multiple baselines.
arXiv Detail & Related papers (2021-02-17T09:44:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.