AIP Publishing LLC
AIP Publishing LLC
  • pubs.aip.org
  • AIP
  • AIP China
  • University Science Books
  • Resources
    • Researchers
    • Librarians
    • Publishing Partners
    • Topical Portfolios
    • Commercial Partners
  • Publications

    Find the Right Journal

    Explore the AIP Publishing collection by title, topic, impact, citations, and more.
    Browse Journals

    Latest Content

    Read about the newest discoveries and developments in the physical sciences.
    See What's New

    Publications

    • Journals
    • Books
    • Physics Today
    • AIP Conference Proceedings
    • Scilight
    • Find the Right Journal
    • Latest Content
  • About
    • About Us
    • News and Announcements
    • Careers
    • Events
    • Leadership
    • Contact
  • pubs.aip.org
  • AIP
  • AIP China
  • University Science Books

‘Fishial’ Recognition: Neural Network Identifies Coral Reef Sounds

  • March 11, 2025
  • The Journal of the Acoustical Society of America
  • News
Share:

Faster identification of fish sounds from acoustic recordings can improve research, conservation efforts

From the Journal: The Journal of the Acoustical Society of America

CUREE, an autonomous underwater robot, is used by the researchers to collect acoustic data for analysis. Image by Austin Greene, Woods Hole Oceanographic Institution
CUREE, an autonomous underwater robot, is used by the researchers to collect acoustic data for analysis. Image by Austin Greene, Woods Hole Oceanographic Institution

WASHINGTON, March 11, 2025 – Coral reefs are some of the world’s most diverse ecosystems. Despite making up less than 1% of the world’s oceans, one quarter of all marine species spend some portion of their life on a reef. With so much life in one spot, researchers can struggle to gain a clear understanding of which species are present and in what numbers.

In JASA, published on behalf of the Acoustical Society of America by AIP Publishing, researchers from Woods Hole Oceanographic Institution combined acoustic monitoring with a neural network to identify fish activity on coral reefs by sound.

For years, researchers have used passive acoustic monitoring to track coral reef activity. Typically, an acoustic recorder would be deployed underwater, where it would spend months recording audio from a reef. Existing signal processing tools can be used to analyze large batches of acoustic data at a time, but they cannot be used to find specific sounds — to do that, scientists usually need to go through all that data by hand.

“But for the people that are doing that, it’s awful work, to be quite honest,” said author Seth McCammon. “It’s incredibly tedious work. It’s miserable.”

Equally as important, this type of manual analysis is too slow for practical use. With many of the world’s coral reefs under threat from climate change and human activity, being able to rapidly identify and track changes in reef populations is crucial for conservation efforts.

“It takes years to analyze data to that level with humans,” said McCammon. “The analysis of the data in this way is not useful at scale.”

As an alternative, the researchers trained a neural network to sort through the deluge of acoustic data automatically, analyzing audio recordings in real time. Their algorithm can match the accuracy of human experts in deciphering acoustical trends on a reef, but it can do so more than 25 times faster, and it could change the way ocean monitoring and research is conducted.

“Now that we no longer need to have a human in the loop, what other sorts of devices — moving beyond just recorders — could we use?” said McCammon. “Some work that my co-author Aran Mooney is doing involves integrating this type of neural network onto a floating mooring that’s broadcasting real-time updates of fish call counts. We are also working on putting our neural network onto our autonomous underwater vehicle, CUREE, so that it can listen for fish and map out hot spots of biological activity.”

This technology also has the potential to solve a long-standing problem in marine acoustic studies: matching each unique sound to a fish.

“For the vast majority of species, we haven’t gotten to the point yet where we can say with certainty that a call came from a particular species of fish,” said McCammon. “That’s, at least in my mind, the holy grail we’re looking for. By being able to do fish call detection in real time, we can start to build devices that are able to automatically hear a call and then see what fish are nearby.”

Eventually, McCammon hopes that this neural network will provide researchers with the ability to monitor fish populations in real time, identify species in trouble, and respond to disasters. This technology will help conservationists gain a clearer picture of the health of coral reefs, in an era where reefs need all the help they can get.

###

Article Title

Rapid detection of fish calls within diverse coral reef soundscapes using a convolutional neural network

Authors

Seth McCammon, Nathan Formel, Sierra Jarriel, and T. Aran Mooney

Author Affiliations

Woods Hole Oceanographic Institution


The Journal of the Acoustical Society of America

Since 1929, The Journal of the Acoustical Society of America (JASA) has been the leading source of theoretical and experimental research results in the broad interdisciplinary subject of sound.

https://asa.scitation.org/journal/jas

Share:
  • Heat from the Sun Affects Seismic Activity on Earth
  • What’s Behind the ‘Pop and Slosh’ When Opening a Swing-Top Bottle of Beer?

Keep Up With AIP Publishing

Sign up for the AIP newsletter to receive the latest news and information from AIP Publishing.
Sign Up

AIP PUBLISHING

1305 Walt Whitman Road,
Suite 110
Melville, NY 11747
(516) 576-2200

Resources

  • Researchers
  • Librarians
  • Publishing Partners
  • Commercial Partners

About

  • About Us
  • Careers 
  • Leadership

Support

  • Contact Us
  • Terms Of Use
  • Privacy Policy

© 2025 AIP Publishing LLC
  • 𝕏