DeepSqueak tool identifies marine mammal calls
Lurking beneath the ocean's surface, marine mammals use sound for navigation, prey detection, and a wide range of natural behaviors. Passive acoustic data from underwater environments can provide valuable information on these animals, such as their presence or absence within an area, their density and abundance, and their vocal response to anthropogenic noise sources.
As the size and number of acoustic datasets increase, accurately and quickly matching the bioacoustics signals to their corresponding sources becomes more challenging and important. This is especially difficult in noisy, natural acoustic environments.
Elizabeth Ferguson, from Ocean Science Analytics, will describe how DeepSqueak, a deep learning tool, can classify underwater acoustic signals at the 182nd Meeting of the Acoustical Society of America during her presentation, "Development of deep neural networks for marine mammal call detection using an open-source, user friendly tool." The session will take place May 23 at 11:25 a.m. Eastern U.S. as part of the conference at the Sheraton Denver Downtown Hotel.
Spectrograms show how acoustic signals of different frequencies vary with time. They look like heat maps, with brighter regions indicating higher sound intensity at that frequency and time. DeepSqueak uses deep neural network image recognition and classification methods to determine the important features within spectrograms, then match those features to specific sources.
"Although we used DeepSqueak to detect underwater sounds, this user-friendly, open source tool would be useful for a variety of terrestrial species," said Ferguson. "The capabilities of call detection extend to frequencies below the ultrasonic sounds it was originally intended for. Due to this and the capability of DeepSqueak to detect variable call types, development of neural networks is possible for many species of interest."
DeepSqueak was originally developed to classify ultrasound signals from rodents, but its neural network framework allows the technique to adapt to detect sounds at other frequencies. Ferguson and her team used the method and data from hydrophones on the Ocean Observatories Initiative's Coastal Endurance Array to detect humpback whales, delphinids, and fin whales, which have highly variable calls with a wide range of frequencies.