Objective hearing threshold identification from auditory brainstem
response measurements using supervised and self-supervised approaches
- URL: http://arxiv.org/abs/2112.08961v1
- Date: Thu, 16 Dec 2021 15:24:31 GMT
- Title: Objective hearing threshold identification from auditory brainstem
response measurements using supervised and self-supervised approaches
- Authors: Dominik Thalmeier, Gregor Miller, Elida Schneltzer, Anja Hurt, Martin
Hrab\v{e} de Angelis, Lore Becker, Christian L. M\"uller, Holger Maier
- Abstract summary: We develop and compare two methods for automated hearing threshold identification from averaged ABR raw data.
We show that both models work well, outperform human threshold detection, and are suitable for fast, reliable, and unbiased hearing threshold detection and quality control.
- Score: 1.0627340704073347
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Hearing loss is a major health problem and psychological burden in humans.
Mouse models offer a possibility to elucidate genes involved in the underlying
developmental and pathophysiological mechanisms of hearing impairment. To this
end, large-scale mouse phenotyping programs include auditory phenotyping of
single-gene knockout mouse lines. Using the auditory brainstem response (ABR)
procedure, the German Mouse Clinic and similar facilities worldwide have
produced large, uniform data sets of averaged ABR raw data of mutant and
wildtype mice. In the course of standard ABR analysis, hearing thresholds are
assessed visually by trained staff from series of signal curves of increasing
sound pressure level. This is time-consuming and prone to be biased by the
reader as well as the graphical display quality and scale. In an attempt to
reduce workload and improve quality and reproducibility, we developed and
compared two methods for automated hearing threshold identification from
averaged ABR raw data: a supervised approach involving two combined neural
networks trained on human-generated labels and a self-supervised approach,
which exploits the signal power spectrum and combines random forest sound level
estimation with a piece-wise curve fitting algorithm for threshold finding. We
show that both models work well, outperform human threshold detection, and are
suitable for fast, reliable, and unbiased hearing threshold detection and
quality control. In a high-throughput mouse phenotyping environment, both
methods perform well as part of an automated end-to-end screening pipeline to
detect candidate genes for hearing involvement. Code for both models as well as
data used for this work are freely available.
Related papers
- All Thresholds Barred: Direct Estimation of Call Density in Bioacoustic
Data [1.7916003204531015]
We propose a validation scheme for estimating call density in a body of data.
We use these distributions to predict site-level densities, which may be subject to distribution shifts.
arXiv Detail & Related papers (2024-02-23T14:52:44Z) - Attention-Based Recurrent Neural Network For Automatic Behavior Laying
Hen Recognition [0.0]
This work focuses on the recognition of the types of calls of the laying hens in order to propose a robust system of characterization of their behavior.
We first collected and annotated laying hen call signals, then designed an optimal acoustic characterization based on the combination of time and frequency domain features.
We then used these features to build the multi-label classification models based on recurrent neural network to assign a semantic class to the vocalization that characterize the laying hen behavior.
arXiv Detail & Related papers (2024-01-18T10:52:46Z) - Multi-Transfer Learning Techniques for Detecting Auditory Brainstem
Response [4.023511716339818]
Inaccurate assessment of auditory brainstem response (ABR) tests may lead to incorrect judgments regarding the integrity of the auditory nerve system.
This study proposed deep-learning models using the transfer-learning (TL) approach to extract features from ABR testing and diagnose Hearing Loss using support vector machines (SVM)
It has been decided to use six measures accuracy, precision, recall, geometric mean (GM), standard deviation (SD), and area under the ROC curve to measure the effectiveness of the proposed model.
arXiv Detail & Related papers (2023-08-29T10:40:12Z) - Leveraging Pretrained Representations with Task-related Keywords for
Alzheimer's Disease Detection [69.53626024091076]
Alzheimer's disease (AD) is particularly prominent in older adults.
Recent advances in pre-trained models motivate AD detection modeling to shift from low-level features to high-level representations.
This paper presents several efficient methods to extract better AD-related cues from high-level acoustic and linguistic features.
arXiv Detail & Related papers (2023-03-14T16:03:28Z) - Utilizing synthetic training data for the supervised classification of
rat ultrasonic vocalizations [0.0]
Murine rodents generate ultrasonic vocalizations (USVs) with frequencies that extend to around 120kHz.
These calls are important in social behaviour, and so their analysis can provide insights into the function of vocal communication, and its dysfunction.
We compare the detection and classification performance of a trained human against two convolutional neural networks (CNNs), DeepSqueak and VocalMat, on audio containing rat USVs.
arXiv Detail & Related papers (2023-03-03T03:17:45Z) - Decision Forest Based EMG Signal Classification with Low Volume Dataset
Augmented with Random Variance Gaussian Noise [51.76329821186873]
We produce a model that can classify six different hand gestures with a limited number of samples that generalizes well to a wider audience.
We appeal to a set of more elementary methods such as the use of random bounds on a signal, but desire to show the power these methods can carry in an online setting.
arXiv Detail & Related papers (2022-06-29T23:22:18Z) - Self-supervised models of audio effectively explain human cortical
responses to speech [71.57870452667369]
We capitalize on the progress of self-supervised speech representation learning to create new state-of-the-art models of the human auditory system.
We show that these results show that self-supervised models effectively capture the hierarchy of information relevant to different stages of speech processing in human cortex.
arXiv Detail & Related papers (2022-05-27T22:04:02Z) - Discriminative Singular Spectrum Classifier with Applications on
Bioacoustic Signal Recognition [67.4171845020675]
We present a bioacoustic signal classifier equipped with a discriminative mechanism to extract useful features for analysis and classification efficiently.
Unlike current bioacoustic recognition methods, which are task-oriented, the proposed model relies on transforming the input signals into vector subspaces.
The validity of the proposed method is verified using three challenging bioacoustic datasets containing anuran, bee, and mosquito species.
arXiv Detail & Related papers (2021-03-18T11:01:21Z) - Improving Medical Image Classification with Label Noise Using
Dual-uncertainty Estimation [72.0276067144762]
We discuss and define the two common types of label noise in medical images.
We propose an uncertainty estimation-based framework to handle these two label noise amid the medical image classification task.
arXiv Detail & Related papers (2021-02-28T14:56:45Z) - Decoding Methods for Neural Narrative Generation [74.37264021226308]
Narrative generation is an open-ended NLP task in which a model generates a story given a prompt.
We apply and evaluate advances in decoding methods for neural response generation to neural narrative generation.
arXiv Detail & Related papers (2020-10-14T19:32:56Z) - Exploration of Audio Quality Assessment and Anomaly Localisation Using
Attention Models [37.60722440434528]
In this paper, a novel model for audio quality assessment is proposed by jointly using bidirectional long short-term memory and an attention mechanism.
The former is to mimic a human auditory perception ability to learn information from a recording, and the latter is to further discriminate interferences from desired signals by highlighting target related features.
To evaluate our proposed approach, the TIMIT dataset is used and augmented by mixing with various natural sounds.
arXiv Detail & Related papers (2020-05-16T17:54:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.