English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Journal Article

Bioacoustic classification of a small dataset of mammalian vocalisations using deep learning

MPS-Authors
/persons/resource/persons19791

Kotz,  Sonja A.       
Department of Neuropsychology and Psychopharmacology, Maastricht University, the Netherlands;
Department Neuropsychology, MPI for Human Cognitive and Brain Sciences, Max Planck Society;

External Resource
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)

Manriquez_2024.pdf
(Publisher version), 3MB

Supplementary Material (public)
There is no public supplementary material available
Citation

Manriquez P., R., Kotz, S. A., Ravignani, A., & de Boer, B. (2024). Bioacoustic classification of a small dataset of mammalian vocalisations using deep learning. Bioacoustics. doi:10.1080/09524622.2024.2354468.


Cite as: https://hdl.handle.net/21.11116/0000-000F-89E9-0
Abstract
In the last few years, automatic extraction and classification of animal vocalisations has been facilitated by machine learning (ML) and deep learning (DL) methods. Different frameworks allowed researchers to automatically extract features and perform classification tasks, aiding in call identification and species recognition. However, the success of these applications relies on the amount of available data to train these algorithms. The lack of sufficient data can also lead to overfitting and affect generalisation (i.e. poor performance on out-of-sample data). Further, acquiring large data sets is costly and annotating them is time consuming. Thus, how small can a dataset be to still provide useful information by means of ML or DL? Here, we show how convolutional neural network architectures can handle small datasets in a bioacoustic classification task of affective mammalian vocalisations. We explain how these techniques can be used (e.g. pre-training and data augmentation), and emphasise how to implement them in concordance with features of bioacoustic signals. We further discuss whether these networks can generalise the affective quality of vocalisations across different taxa.