Free · No upload · 100% browser · InstantDetects Suno · Udio · MusicGen · ElevenLabs

Is This Song AI Generated?
Free AI Music Detector

Drop any song. We analyze 5 acoustic signals — high-frequency artifacts, spectral flatness, timing consistency, harmonic-to-noise ratio, and dynamic range — entirely in your browser. Results in seconds. Free, no signup, no upload.

5 DSP signals · per-signal breakdown · honest verdict · 70-85% accuracy range

Detects output fromSunoUdioMurekaElevenLabsMusicGenMusicFX

The 5 acoustic signals — explained in detail

Every measurement happens in your browser via JavaScript DSP. No model download, no server round-trip. Each signal returns an independent score from 0 (definitely human) to 1 (definitely AI), then we combine them with calibrated weights into a single verdict.

01

High-frequency artifact density

What it measures

How unusual is the energy distribution above 8 kHz?

Why it works

AI music generators use neural vocoders that introduce a 'metallic sheen' between 8-16 kHz. The harmonic spacing in this band has measurable patterns not found in natural recordings. Suno output is particularly identifiable here.

Reading the score

Higher score = more AI-like high-frequency character. Score above 0.7 strongly suggests AI vocoder output.

02

Spectral flatness variance

What it measures

How much does the spectral character change over time?

Why it works

Human-recorded music transitions between sections — verses vs choruses, acoustic vs electric, dry vs reverb-heavy — and the spectral character shifts accordingly. AI music tends to have less variation because the model converges on consistent acoustic characteristics across a song.

Reading the score

Lower variance score = more AI-like (suspiciously constant). High variance score is human-like.

03

Onset timing consistency

What it measures

Do note onsets have natural human micro-timing?

Why it works

Real musicians have consistent average tempo with ±5-15ms variation per note. Drum machines have ~0ms variation. AI music sits in an uncanny middle: not perfectly quantized, but more consistent than human performers. We measure inter-onset interval variance and compare to the expected human distribution.

Reading the score

Too consistent = AI-like. Too random = drum machine or click track. Natural variation = human.

04

Harmonic-to-noise ratio at vocal frequencies

What it measures

How clean is the harmonic content in the 200-4000 Hz range?

Why it works

AI vocals have specific HNR signatures — typically higher than natural vocals because the generator outputs idealized harmonics with less breath noise, sibilance, and consonant artifacts than humans naturally produce.

Reading the score

Unusually high HNR = AI-like. Natural HNR with breath/sibilance noise = human. Less reliable on instrumentals.

05

Crest factor / dynamic range

What it measures

How much does the amplitude vary between peaks and average?

Why it works

AI mixes tend to be more compressed than human mixes. The model learns to output a consistent loudness profile because that's what most training data sounds like. Human mixes have more peak-to-RMS variation, especially in genres like classical, jazz, and acoustic music.

Reading the score

Lower dynamic range = more AI-like (heavy compression baked in). Wider dynamic range = human-like.

How AI music differs from human music — measurable signals

Eight measurable acoustic differences between human-recorded and AI-generated music. Our detector samples five of these (the most reliable in browser-based DSP) and combines them into a single verdict.

Acoustic featureHuman-recordedAI-generated
High-frequency character (>8kHz)Natural decay, breath noise, variedMetallic sheen, unusual harmonic spacing
Spectral character varianceShifts between sectionsSuspiciously constant
Onset timing variance±5-15ms human micro-timingToo consistent (uncanny middle)
Vocal harmonic-to-noise ratioMid HNR with breath / sibilanceUnusually high HNR (no breath noise)
Crest factor / dynamic rangeWide variation, especially in dynamics genresHeavily compressed (learned from training data)
Reverb / room cuesNatural acoustic reflectionsOften dry or unnaturally consistent
Stereo width consistencyVaries by sectionConstant width throughout
Vocal formant transitionsSmooth glides between phonemesSubtle discontinuities at phoneme boundaries

Frequently asked questions

Is this song AI generated? How can I tell?

AI-generated music has measurable acoustic differences from human-recorded music: unusual energy distribution above 8 kHz (the 'metallic sheen' from AI vocoders), too-consistent timing (no human micro-timing variations), low spectral flatness variance (AI character changes less over time), and lower dynamic range. Drop your song into our free detector and we'll measure all 5 signals in your browser.

How accurate is this AI music detector?

Honest answer: no AI music detector is 100% accurate. State-of-the-art ML classifiers reach 85-95% on clean test sets but degrade on real-world audio. Our statistical detector is in the 70-85% accuracy range — useful as a screening signal, not a verdict. The tool shows you the breakdown of each signal so you can interpret rather than just trust a single number. False positives (real music flagged as AI) are most common with heavily auto-tuned vocal tracks and extremely compressed modern pop.

What AI music tools does this detect?

The detector picks up acoustic artifacts common to most current AI music generators: Suno (all versions), Udio, Mureka, MusicGen, ElevenLabs Music, MusicFX, and similar. The signals it measures (high-freq spectral envelope, timing consistency, dynamic range) are characteristic of AI vocoder output regardless of which model produced it. It's less reliable on tools that use diffusion-only audio synthesis with no vocoder step.

Why is my song flagged as AI when I made it myself?

Common false-positive triggers on human-recorded music: heavy auto-tune (compresses pitch variance similarly to AI vocals), modern pop production (heavily compressed dynamic range), 100% MIDI-programmed instruments (no human timing variance), and certain electronic genres (EDM with sample-based percussion looks like AI to the timing-consistency signal). Look at the per-signal breakdown — if only the timing signal flagged but the rest were clean, it's likely a false positive.

Does my audio leave my browser?

No. The audio file stays on your device. Decoding is done by the Web Audio API. All 5 signal analyses run locally with JavaScript DSP — no upload, no server processing, no log. The only network call is to load the page itself.

Why does the high-frequency analysis matter?

AI music generators use neural vocoders that decode audio from a learned latent space. Vocoders introduce specific high-frequency artifacts — typically a 'metallic sheen' between 8 kHz and 16 kHz that has unusual harmonic spacing not found in natural recordings. Our signal #1 measures this directly by comparing the spectral envelope above 8 kHz to expected human-recording statistics.

What are the 5 signals exactly?

Signal 1: High-frequency artifact density — unusual energy patterns above 8 kHz from AI vocoder output. Signal 2: Spectral flatness variance — does the spectral character change naturally over time, or is it suspiciously constant? Signal 3: Onset timing consistency — do note onsets have natural human micro-timing variation, or are they too regular? Signal 4: Harmonic-to-noise ratio in vocal range — AI vocals have specific HNR signatures. Signal 5: Crest factor / dynamic range — AI mixes are often more compressed than human mixes.

Will this catch Suno songs that have been mastered?

Mostly yes. Mastering changes loudness, EQ, and stereo width but doesn't remove the underlying AI vocoder fingerprint above 8 kHz or the unnaturally consistent timing. Mastering may slightly compress the dynamic-range signal, but the other 4 signals remain. Heavy mastering with aggressive saturation and compression can produce false negatives in some cases.

Can I use this to prove a song was AI-generated for legal purposes?

No. This is a screening tool with statistical accuracy in the 70-85% range. For legal use cases (copyright disputes, distribution disputes, content authenticity claims) you need a forensic-grade audio analysis from a qualified expert. Our tool is appropriate for: deciding which tracks to investigate further, screening submissions, satisfying personal curiosity. Not appropriate for: courts, contract enforcement, official disclosures.

How long does the analysis take?

About 5-15 seconds for a typical 3-minute song. The pipeline decodes the audio, computes a spectrogram, runs all 5 signal analyses in parallel, and presents results. There's no model download, no server round-trip, and no warmup — just JavaScript DSP. Larger files take proportionally longer (a 10-minute song may take ~30 seconds).

Does this work on instrumentals?

Partially. 4 of the 5 signals work on any audio (high-freq artifacts, spectral flatness, timing consistency, dynamic range). The harmonic-to-noise ratio signal is most informative on vocal tracks; on instrumentals it's still computed but contributes less weight. Result: instrumental analysis is slightly less accurate but still useful.

What if I disagree with the verdict?

Look at the per-signal breakdown the tool shows you. If only one signal flagged but the others were clean, the verdict is weak. If 4 of 5 signals flagged, the verdict is strong. This tool is designed to be transparent — every measurement is shown to you. Don't trust a single number; trust the pattern across signals. And remember: no detector is 100% accurate.

Related tools

Master your track free