Highlights: What are the main findings? A CNN was developed for multiclass classification of dolphin vocalizations. The model achieved a 95.2% mean accuracy with a mean F1-score of 87.8%. What is the implication of the main finding? This approach enhances the passive acoustic monitoring of dolphin vocalizations. It provides a scalable AI-based solution for marine bioacoustics research. Monitoring dolphins in the open sea is essential for understanding their behavior and the impact of human activities on the marine ecosystems. Passive Acoustic Monitoring (PAM) is a non-invasive technique for tracking dolphins, providing continuous data. This study presents a novel approach for classifying dolphin vocalizations from a PAM acoustic recording using a convolutional neural network (CNN). Four types of common bottlenose dolphin (Tursiops truncatus) vocalizations were identified from underwater recordings: whistles, echolocation clicks, burst pulse sounds, and feeding buzzes. To enhance classification performances, edge-detection filters were applied to spectrograms, with the aim of removing unwanted noise components. A dataset of nearly 10,000 spectrograms was used to train and test the CNN through a 10-fold cross-validation procedure. The results showed that the CNN achieved an average accuracy of 95.2% and an F1-score of 87.8%. The class-specific results showed a high accuracy for whistles (97.9%), followed by echolocation clicks (94.5%), feeding buzzes (94.0%), and burst pulse sounds (92.3%). The highest F1-score was obtained for whistles, exceeding 95%, while the other three vocalization typologies maintained an F1-score above 80%. This method provides a promising step toward improving the passive acoustic monitoring of dolphins, contributing to both species conservation and the mitigation of conflicts with fisheries.

Multiclass CNN Approach for Automatic Classification of Dolphin Vocalizations / Di Nardo, Francesco; De Marco, Rocco; Li Veli, Daniel; Screpanti, Laura; Castagna, Benedetta; Lucchetti, Alessandro; Scaradozzi, David. - In: SENSORS. - ISSN 1424-8220. - ELETTRONICO. - 25:8(2025). [10.3390/s25082499]

Multiclass CNN Approach for Automatic Classification of Dolphin Vocalizations

Di Nardo, Francesco
Primo
;
Screpanti, Laura
Secondo
;
Castagna, Benedetta;Scaradozzi, David
Ultimo
2025-01-01

Abstract

Highlights: What are the main findings? A CNN was developed for multiclass classification of dolphin vocalizations. The model achieved a 95.2% mean accuracy with a mean F1-score of 87.8%. What is the implication of the main finding? This approach enhances the passive acoustic monitoring of dolphin vocalizations. It provides a scalable AI-based solution for marine bioacoustics research. Monitoring dolphins in the open sea is essential for understanding their behavior and the impact of human activities on the marine ecosystems. Passive Acoustic Monitoring (PAM) is a non-invasive technique for tracking dolphins, providing continuous data. This study presents a novel approach for classifying dolphin vocalizations from a PAM acoustic recording using a convolutional neural network (CNN). Four types of common bottlenose dolphin (Tursiops truncatus) vocalizations were identified from underwater recordings: whistles, echolocation clicks, burst pulse sounds, and feeding buzzes. To enhance classification performances, edge-detection filters were applied to spectrograms, with the aim of removing unwanted noise components. A dataset of nearly 10,000 spectrograms was used to train and test the CNN through a 10-fold cross-validation procedure. The results showed that the CNN achieved an average accuracy of 95.2% and an F1-score of 87.8%. The class-specific results showed a high accuracy for whistles (97.9%), followed by echolocation clicks (94.5%), feeding buzzes (94.0%), and burst pulse sounds (92.3%). The highest F1-score was obtained for whistles, exceeding 95%, while the other three vocalization typologies maintained an F1-score above 80%. This method provides a promising step toward improving the passive acoustic monitoring of dolphins, contributing to both species conservation and the mitigation of conflicts with fisheries.
2025
convolutional neural networks; deep learning; dolphins; passive acoustic monitoring
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11566/349565
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
social impact