A keystone species, burrowing owls convey crucial ecological information through their vocalizations. This project creates a whole TinyML pipeline that listens for, recognizes, and categorizes six different burrowing-owl sounds in real time, all of which are controlled by an STM32H7 microcontroller. To train lightweight CNNs (Custom Tiny CNN, MobileNetV2, ProxylessNAS) on the BUOWSET dataset, we transform raw audio into 64-band Mel spectrograms, quantize the model to int8, and embed it as a C header for on-device inference.
Our goal is to build a TinyML-based system that runs entirely on an STM32H7 board to listen for, detect, and classify six different burrowing-owl vocalizations in real time. By converting audio to 64-band Mel spectrograms, training lightweight CNNs (MobileNetV2 and ProxylessNAS) on the BUOWSET dataset, then quantizing to int8 and embedding the model as a C header, we achieve:
We trained three compact CNN architectures optimized for on-device inference on STM32H7. Final validation accuracy and F1 score (epoch 20) are shown below:
We gratefully acknowledge the support and data provided by the Engineers For Exploration (E4E) Acoustic Species Identification Lab at UC San Diego. E4E is a research group focused on protecting the environment, uncovering mysteries related to cultural heritage, and providing experiential learning experiences for undergraduate and graduate students.
This work was conducted in collaboration with the San Diego Zoo Wildlife Alliance, whose expertise in burrowing-owl ecology and field data collection was invaluable.
We also thank Professor Ryan Kastner for his guidance and support throughout this project.
Contact & Project Lead at E4E:
Email:
lvonschoenfeldt@ucsd.edu