Insect population numbers and biodiversity have been rapidly declining with time, and monitoring these trends has become increasingly important for conservation measures to be effectively implemented. But monitoring methods are often invasive, time and resource intense, and prone to various biases. Many insect species produce characteristic sounds that can easily be detected and recorded without large cost or effort. Using deep learning methods, insect sounds from field recordings could be automatically detected and classified to monitor biodiversity and species distribution ranges. We implement this using recently published datasets of insect sounds (up to 66 species of Orthoptera and Cicadidae) and machine learning methods and evaluate their potential for acoustic insect monitoring. We compare the performance of the conventional spectrogram-based audio representation against LEAF, a new adaptive and waveform-based frontend. LEAF achieved better classification performance than the mel-spectrogram frontend by adapting its feature extraction parameters during training. This result is encouraging for future implementations of deep learning technology for automatic insect sound recognition, especially as larger datasets become available.

Public Library of Science (PLoS)
doi.org/10.1371/journal.pcbi.1011541
PLOS Computational Biology

Released under the CC-BY 4.0 ("Attribution 4.0 International") License

Staff publications

Faiß, M., & Stowell, D. (2023). Adaptive representations of sound for automatic insect recognition. PLOS Computational Biology, 19(10), e1011541–e1011541. doi:10.1371/journal.pcbi.1011541