Unsupervised classification to improve the quality of a bird song
recording dataset
- URL: http://arxiv.org/abs/2302.07560v1
- Date: Wed, 15 Feb 2023 10:01:58 GMT
- Title: Unsupervised classification to improve the quality of a bird song
recording dataset
- Authors: F\'elix Michaud (ISYEB ), J\'er\^ome Sueur (ISYEB ), Maxime Le Cesne
(ISYEB ), Sylvain Haupert (ISYEB )
- Abstract summary: We introduce a data-centric novel labelling function composed of three successive steps: time-frequency sound unit segmentation, feature computation for each sound unit, and classification of each sound unit as bird song or noise.
Our labelling function was able to significantly reduce the initial label noise present in the dataset by up to a factor of three.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open audio databases such as Xeno-Canto are widely used to build datasets to
explore bird song repertoire or to train models for automatic bird sound
classification by deep learning algorithms. However, such databases suffer from
the fact that bird sounds are weakly labelled: a species name is attributed to
each audio recording without timestamps that provide the temporal localization
of the bird song of interest. Manual annotations can solve this issue, but they
are time consuming, expert-dependent, and cannot run on large datasets. Another
solution consists in using a labelling function that automatically segments
audio recordings before assigning a label to each segmented audio sample.
Although labelling functions were introduced to expedite strong label
assignment, their classification performance remains mostly unknown. To address
this issue and reduce label noise (wrong label assignment) in large bird song
datasets, we introduce a data-centric novel labelling function composed of
three successive steps: 1) time-frequency sound unit segmentation, 2) feature
computation for each sound unit, and 3) classification of each sound unit as
bird song or noise with either an unsupervised DBSCAN algorithm or the
supervised BirdNET neural network. The labelling function was optimized,
validated, and tested on the songs of 44 West-Palearctic common bird species.
We first showed that the segmentation of bird songs alone aggregated from 10%
to 83% of label noise depending on the species. We also demonstrated that our
labelling function was able to significantly reduce the initial label noise
present in the dataset by up to a factor of three. Finally, we discuss
different opportunities to design suitable labelling functions to build
high-quality animal vocalizations with minimum expert annotation effort.
Related papers
- Unsupervised outlier detection to improve bird audio dataset labels [0.0]
Non-target bird species sounds can result in dataset labeling discrepancies referred to as label noise.
We present a cleaning process consisting of audio preprocessing followed by dimensionality reduction and unsupervised outlier detection.
arXiv Detail & Related papers (2025-04-25T19:04:40Z) - A Bird Song Detector for improving bird identification through Deep Learning: a case study from DoƱana [2.7924253850013416]
We develop a pipeline for automatic bird vocalization identification in Donana National Park (SW Spain)
We manually annotated 461 minutes of audio from three habitats across nine locations, yielding 3,749 annotations for 34 classes.
Applying the Bird Song Detector before classification improved species identification, as all classification models performed better when analyzing only the segments where birds were detected.
arXiv Detail & Related papers (2025-03-19T13:19:06Z) - AlleNoise: large-scale text classification benchmark dataset with real-world label noise [40.11095094521714]
We present AlleNoise, a new curated text classification benchmark dataset with real-world instance-dependent label noise.
The noise distribution comes from actual users of a major e-commerce marketplace, so it realistically reflects the semantics of human mistakes.
We demonstrate that a representative selection of established methods for learning with noisy labels is inadequate to handle such real-world noise.
arXiv Detail & Related papers (2024-06-24T09:29:14Z) - Learning from Time Series under Temporal Label Noise [23.39598516168891]
We first propose and formalize temporal label noise, an unstudied problem for sequential classification of time series.
We show that our methods lead to state-of-the-art performance in the presence of diverse temporal label noise functions using real and synthetic data.
arXiv Detail & Related papers (2024-02-06T20:56:31Z) - Exploring Meta Information for Audio-based Zero-shot Bird Classification [113.17261694996051]
This study investigates how meta-information can improve zero-shot audio classification.
We use bird species as an example case study due to the availability of rich and diverse meta-data.
arXiv Detail & Related papers (2023-09-15T13:50:16Z) - S3: Supervised Self-supervised Learning under Label Noise [53.02249460567745]
In this paper we address the problem of classification in the presence of label noise.
In the heart of our method is a sample selection mechanism that relies on the consistency between the annotated label of a sample and the distribution of the labels in its neighborhood in the feature space.
Our method significantly surpasses previous methods on both CIFARCIFAR100 with artificial noise and real-world noisy datasets such as WebVision and ANIMAL-10N.
arXiv Detail & Related papers (2021-11-22T15:49:20Z) - ARCA23K: An audio dataset for investigating open-set label noise [48.683197172795865]
This paper introduces ARCA23K, an automatically Retrieved and curated audio dataset comprised of over 23000 labelled Freesound clips.
We show that the majority of labelling errors in ARCA23K are due to out-of-vocabulary audio clips, and we refer to this type of label noise as open-set label noise.
arXiv Detail & Related papers (2021-09-19T21:10:25Z) - Instance-dependent Label-noise Learning under a Structural Causal Model [92.76400590283448]
Label noise will degenerate the performance of deep learning algorithms.
By leveraging a structural causal model, we propose a novel generative approach for instance-dependent label-noise learning.
arXiv Detail & Related papers (2021-09-07T10:42:54Z) - Learning with Feature-Dependent Label Noise: A Progressive Approach [19.425199841491246]
We propose a new family of feature-dependent label noise, which is much more general than commonly used i.i.d. label noise.
We provide theoretical guarantees showing that for a wide variety of (unknown) noise patterns, a classifier trained with this strategy converges to be consistent with the Bayes classifier.
arXiv Detail & Related papers (2021-03-13T17:34:22Z) - Towards Robustness to Label Noise in Text Classification via Noise
Modeling [7.863638253070439]
Large datasets in NLP suffer from noisy labels, due to erroneous automatic and human annotation procedures.
We study the problem of text classification with label noise, and aim to capture this noise through an auxiliary noise model over the classifier.
arXiv Detail & Related papers (2021-01-27T05:41:57Z) - A Second-Order Approach to Learning with Instance-Dependent Label Noise [58.555527517928596]
The presence of label noise often misleads the training of deep neural networks.
We show that the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks.
arXiv Detail & Related papers (2020-12-22T06:36:58Z) - Attention-Aware Noisy Label Learning for Image Classification [97.26664962498887]
Deep convolutional neural networks (CNNs) learned on large-scale labeled samples have achieved remarkable progress in computer vision.
The cheapest way to obtain a large body of labeled visual data is to crawl from websites with user-supplied labels, such as Flickr.
This paper proposes the attention-aware noisy label learning approach to improve the discriminative capability of the network trained on datasets with potential label noise.
arXiv Detail & Related papers (2020-09-30T15:45:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.