Sound as Photograph

Point a smartphone at the sky at dawn and press record. Within three seconds, the free BirdNET app will tell you what you're hearing — not by matching a waveform to a library of known songs, but by converting sound into a visual image and running that image through a convolutional neural network. More than 6,000 bird species across every continent, identified from a three-second clip.

BirdNET is a joint research project of the K. Lisa Yang Center for Conservation Bioacoustics at the Cornell Lab of Ornithology and the Chair of Media Informatics at Chemnitz University of Technology in Germany. Its principal architect, Stefan Kahl, developed the system as a doctoral project and continues to maintain it as an open-source tool — free, with no financial benefit from downloads or usage. Since its public launch in 2018, the app has been downloaded millions of times, generated over 40 million submissions, and engaged more participants in bird monitoring than eBird — Cornell's own flagship citizen-science platform — within its first two years of existence.

What makes BirdNET remarkable is not its popularity. It is the transparency of its pipeline. Unlike most commercial AI systems, BirdNET's architecture, training data sources, model weights, and deployment code are all openly documented. This makes it a near-perfect case study for understanding what a modern deep learning system actually does when it "recognizes" something — and where, precisely, the intelligence lives.

6,522
Species
identified
3 sec
Audio
window
48k
Sample rate
(Hz)
50 MB
Model
size

From Air Pressure to Pixels

A bird call is a pressure wave. A microphone converts that wave into an electrical signal sampled at 48,000 points per second. BirdNET slices incoming audio into three-second windows — 144,000 samples each. This duration was chosen deliberately: most bird vocalizations — songs, calls, alarm notes — fit within a three-second envelope.

The critical transformation happens next. BirdNET does not feed raw audio into its neural network. Instead, it converts each three-second clip into a mel spectrogram — a two-dimensional image where the horizontal axis represents time, the vertical axis represents frequency, and pixel intensity represents energy. The "mel" scale is a perceptual mapping: it compresses higher frequencies relative to lower ones, matching how vertebrate hearing actually works. The result is a visual fingerprint of sound.

BirdNET computes not one but two mel spectrograms per clip, each tuned to a different frequency band. The low-frequency channel covers 0–3,000 Hz with an FFT window of 2,048 samples — capturing owls, bitterns, and large raptors. The high-frequency channel covers 500–15,000 Hz with a 1,024-sample window — targeting warblers, kinglets, and most passerine calls. Both produce 96 mel bins at 96×511 pixel resolution.

This dual-channel approach avoids the detail loss that would result from representing the entire 0–15 kHz range in a single spectrogram. Once the audio is an image, the problem becomes image classification.

// The BirdNET Pipeline — Sound to Species

Audio is recorded, sliced into 3-second windows, converted to dual mel spectrograms, and classified by an EfficientNet CNN. Location metadata refines the final ranking.

EfficientNet: Seeing Sound

The backbone of BirdNET V2.4 — the current production model, released March 2025 — is an EfficientNetB0-like convolutional neural network. EfficientNet, developed by Google Brain researchers Tan and Le in 2019, introduced compound scaling — a method for uniformly scaling a network's depth, width, and input resolution using a single coefficient, rather than expanding one dimension at a time. The B0 variant is the smallest in the family, optimized for inference speed on mobile and embedded hardware.

BirdNET's implementation modifies the standard EfficientNetB0 with a custom spectrogram input layer and produces a 1,024-dimensional embedding vector before the final classification head. The model weighs 50.5 MB in FP32 precision and requires just 0.826 GFLOPs per inference — small enough to run in real time on a five-year-old smartphone.

The classification head maps the 1,024-dimensional embedding to 6,522 output classes — one per supported species. Each class receives a confidence score. But raw visual classification alone produces too many false positives: spectrograms of different species can look remarkably similar, especially among closely related taxa. This is where geography becomes critical.

Dual Spectrogram Channels

Low freq: 0–3,000 Hz · nfft 2048 · hop 278 · 96 mel bins

High freq: 500–15,000 Hz · nfft 1024 · hop 280 · 96 mel bins

Resolution: 96×511 px per channel · Embedding: 1,024 dimensions

Location as Prior

BirdNET does not classify in a vacuum. When a recording includes GPS coordinates and a date, the system applies a biogeographic prior — a probability mask derived from species range data — that suppresses classes for species not expected in that region at that time of year. A Great Tit spectrogram recorded in January in Berlin will correctly trigger Parus major; the same spectrogram submitted from São Paulo will be re-ranked, because Great Tits do not occur in South America.

This is not a simple whitelist. The system uses continuous probability distributions over geographic space, allowing it to handle edge-of-range detections, vagrants, and seasonal migrants. The interplay between visual confidence and geographic prior is what transforms BirdNET from a pattern-matching parlor trick into a genuinely useful ecological instrument.

"Our guiding design principles were that we needed an accurate algorithm and a simple user interface. Otherwise, users would not return to the app."

— Stefan Kahl, BirdNET lead developer

Training on the World's Sound Libraries

The model trains on curated recordings from two of the world's largest bioacoustic archives: Xeno-canto, a community-driven repository with over 900,000 recordings, and the Macaulay Library at Cornell, holding more than 1.5 million audio and video recordings of wildlife.

Training data is filtered for quality, segmented into three-second windows, and subjected to data augmentation — synthetic modifications including time stretching, pitch shifting, background noise injection, and random frequency masking. A dawn chorus in a tropical forest might contain twenty species vocalizing simultaneously, layered over insect noise, wind, and rain. The model must learn to identify a target species despite these overlapping signals.

// Model Version History — Species Coverage Over Time

Each model version expanded both the species list and the geographic breadth of the training set, pushing into underrepresented regions of sub-Saharan Africa, Southeast Asia, and Oceania.

The Deployment Ecosystem

BirdNET is not a single app — it is an ecosystem. The mobile apps (iOS and Android) are the consumer-facing products, free with no account required. BirdNET-Analyzer is a desktop application with GUI and command-line interfaces for batch-processing large audio datasets — the tool most commonly used by researchers. A Progressive Web App runs TensorFlow.js entirely in the browser with no server round-trip. TFLite models deploy to edge devices including Raspberry Pi for autonomous recording stations. Python packages and an R bridge (birdnetR) integrate with scientific computing pipelines.

This multi-platform strategy reflects a deliberate philosophy: the model should go where the microphones are. A researcher deploying autonomous recording units in a Malaysian rainforest canopy needs TFLite on a Raspberry Pi. A casual birder in Central Park needs an Android app. A conservation agency analyzing 10,000 hours of archival recordings needs the command-line analyzer. BirdNET serves all three from the same trained weights.

Citizen Science at Machine Scale

By mid-2022 — four years after launch — BirdNET had engaged more than 2.2 million participants and accumulated over 40 million submissions. In 2020 alone, the app recorded more than 1.1 million active participants, exceeding eBird's 317,792 in the same period. The difference is instructive: eBird requires a registered account, taxonomic knowledge, and structured data entry. BirdNET requires a microphone and a finger.

The model weights are published under Creative Commons Attribution-NonCommercial-ShareAlike 4.0. The analyzer source code is on GitHub. Kahl has explicitly declined to commercialize the system. This matters because it means BirdNET functions as scientific infrastructure, not a product. Researchers can inspect the model, critique its biases, retrain it on local data, and integrate it into larger monitoring pipelines without licensing negotiations.

// Spectrogram Visualization — Simulated Birdsong Analysis

A simulated mel spectrogram showing how frequency energy is distributed across a 3-second window. The neural network reads these images the way a human reads a photograph.

What Listening Machines Still Cannot Do

BirdNET struggles with species that produce highly variable vocalizations — some thrushes, some warblers — where the same individual generates spectrograms that look fundamentally different from one call to the next. It can be fooled by anthropogenic sounds — car alarms, phone ringtones — that occupy the same frequency bands as certain species. And its accuracy drops in high-noise environments with many overlapping vocalizers, precisely the conditions that matter most in tropical biodiversity hotspots.

The deeper issue is epistemological: a correct identification is not the same as a confirmed presence. BirdNET tells you what a sound most resembles. The gap between resemblance and truth is where human expertise remains irreplaceable.

But the question for conservation is whether the sheer volume of BirdNET data — millions of identifications per month, spanning every continent — compensates for the per-observation error rate. The answer, increasingly, is yes. What no human surveyor can match is the temporal and geographic density of a global network of microphones, running 24 hours a day, in places no ornithologist has ever set foot. The ear in the machine does not replace the ear in the field. It extends the field to every phone, every forest, every dawn chorus on Earth.

Sources: BirdNET (birdnet.cornell.edu) • BirdNET Model V2.4 (Zenodo, March 2025) • Kahl et al. (2021), "BirdNET: A deep learning solution for avian diversity monitoring," Ecological Informatics • Wood, Kahl, Rahaman & Klinck (2022), PLOS Biology • Pérez-Granados (2023), "BirdNET: applications, performance, pitfalls and future opportunities," Ibis • BirdNET-Analyzer model documentation (birdnet-team.github.io)