Identifying and Classifying Whale Sounds in Underwater Soundscapes Based on Faster Regionbased Convolutional Neural Networks
Journal
2025 IEEE Underwater Technology, UT 2025
Part Of
2025 IEEE Underwater Technology, UT 2025
Start Page
1
End Page
6
Date Issued
2025-03-02
Author(s)
Abstract
Passive acoustic monitoring is a well-established tool for studying underwater soundscapes, including ship noise, the activity and ecology of marine animal species. As people pay more and more attention to marine ecology and the amount of passive acoustic data collected increases exponentially, an efficient detection model is needed to assist in the analysis of underwater acoustic data. This study proposes Faster R-CNN to identify the sounds of Chinese white dolphins in 36 hours of acoustic data collected at a monitoring point off the coast of Yunlin, Taiwan. The model had an average accuracy of 0.87 and an average area under the receiver operating characteristic curve (AUC-ROC) of ${0. 8 0 2}$. This model output was used to analyze the spatial and temporal patterns of Chinese white dolphin calls, confirming the behavioral patterns of Chinese white dolphins living near Taiwan. This study demonstrates that Faster R-CNN trained on a small data set generalizes well to highly variable signal types under various recording and noise conditions. We demonstrate the utility of transfer learning methods. These results validate the feasibility of applying deep learning models to identify highly variable signals across a wide range of spatial and temporal scales, enabling new discoveries by combining large datasets with cutting-edge tools.
Event(s)
2025 IEEE Underwater Technology, UT 2025, Taipei, 2 March 2025 through 5 March 2025. Code 208171
Subjects
Faster R-CNN
Passive acoustic monitoring
underwater soundscape
whale voiceprint
SDGs
Publisher
IEEE
Type
conference paper
