Drop any song. We analyze 5 acoustic signals — high-frequency artifacts, spectral flatness, timing consistency, harmonic-to-noise ratio, and dynamic range — entirely in your browser. Results in seconds. Free, no signup, no upload.
5 DSP signals · per-signal breakdown · honest verdict · 70-85% accuracy range
Every measurement happens in your browser via JavaScript DSP. No model download, no server round-trip. Each signal returns an independent score from 0 (definitely human) to 1 (definitely AI), then we combine them with calibrated weights into a single verdict.
What it measures
How unusual is the energy distribution above 8 kHz?
Why it works
AI music generators use neural vocoders that introduce a 'metallic sheen' between 8-16 kHz. The harmonic spacing in this band has measurable patterns not found in natural recordings. Suno output is particularly identifiable here.
Reading the score
Higher score = more AI-like high-frequency character. Score above 0.7 strongly suggests AI vocoder output.
What it measures
How much does the spectral character change over time?
Why it works
Human-recorded music transitions between sections — verses vs choruses, acoustic vs electric, dry vs reverb-heavy — and the spectral character shifts accordingly. AI music tends to have less variation because the model converges on consistent acoustic characteristics across a song.
Reading the score
Lower variance score = more AI-like (suspiciously constant). High variance score is human-like.
What it measures
Do note onsets have natural human micro-timing?
Why it works
Real musicians have consistent average tempo with ±5-15ms variation per note. Drum machines have ~0ms variation. AI music sits in an uncanny middle: not perfectly quantized, but more consistent than human performers. We measure inter-onset interval variance and compare to the expected human distribution.
Reading the score
Too consistent = AI-like. Too random = drum machine or click track. Natural variation = human.
What it measures
How clean is the harmonic content in the 200-4000 Hz range?
Why it works
AI vocals have specific HNR signatures — typically higher than natural vocals because the generator outputs idealized harmonics with less breath noise, sibilance, and consonant artifacts than humans naturally produce.
Reading the score
Unusually high HNR = AI-like. Natural HNR with breath/sibilance noise = human. Less reliable on instrumentals.
What it measures
How much does the amplitude vary between peaks and average?
Why it works
AI mixes tend to be more compressed than human mixes. The model learns to output a consistent loudness profile because that's what most training data sounds like. Human mixes have more peak-to-RMS variation, especially in genres like classical, jazz, and acoustic music.
Reading the score
Lower dynamic range = more AI-like (heavy compression baked in). Wider dynamic range = human-like.
Eight measurable acoustic differences between human-recorded and AI-generated music. Our detector samples five of these (the most reliable in browser-based DSP) and combines them into a single verdict.
| Acoustic feature | Human-recorded | AI-generated |
|---|---|---|
| High-frequency character (>8kHz) | Natural decay, breath noise, varied | Metallic sheen, unusual harmonic spacing |
| Spectral character variance | Shifts between sections | Suspiciously constant |
| Onset timing variance | ±5-15ms human micro-timing | Too consistent (uncanny middle) |
| Vocal harmonic-to-noise ratio | Mid HNR with breath / sibilance | Unusually high HNR (no breath noise) |
| Crest factor / dynamic range | Wide variation, especially in dynamics genres | Heavily compressed (learned from training data) |
| Reverb / room cues | Natural acoustic reflections | Often dry or unnaturally consistent |
| Stereo width consistency | Varies by section | Constant width throughout |
| Vocal formant transitions | Smooth glides between phonemes | Subtle discontinuities at phoneme boundaries |
AI-generated music has measurable acoustic differences from human-recorded music: unusual energy distribution above 8 kHz (the 'metallic sheen' from AI vocoders), too-consistent timing (no human micro-timing variations), low spectral flatness variance (AI character changes less over time), and lower dynamic range. Drop your song into our free detector and we'll measure all 5 signals in your browser.
Honest answer: no AI music detector is 100% accurate. State-of-the-art ML classifiers reach 85-95% on clean test sets but degrade on real-world audio. Our statistical detector is in the 70-85% accuracy range — useful as a screening signal, not a verdict. The tool shows you the breakdown of each signal so you can interpret rather than just trust a single number. False positives (real music flagged as AI) are most common with heavily auto-tuned vocal tracks and extremely compressed modern pop.
The detector picks up acoustic artifacts common to most current AI music generators: Suno (all versions), Udio, Mureka, MusicGen, ElevenLabs Music, MusicFX, and similar. The signals it measures (high-freq spectral envelope, timing consistency, dynamic range) are characteristic of AI vocoder output regardless of which model produced it. It's less reliable on tools that use diffusion-only audio synthesis with no vocoder step.
Common false-positive triggers on human-recorded music: heavy auto-tune (compresses pitch variance similarly to AI vocals), modern pop production (heavily compressed dynamic range), 100% MIDI-programmed instruments (no human timing variance), and certain electronic genres (EDM with sample-based percussion looks like AI to the timing-consistency signal). Look at the per-signal breakdown — if only the timing signal flagged but the rest were clean, it's likely a false positive.
No. The audio file stays on your device. Decoding is done by the Web Audio API. All 5 signal analyses run locally with JavaScript DSP — no upload, no server processing, no log. The only network call is to load the page itself.
AI music generators use neural vocoders that decode audio from a learned latent space. Vocoders introduce specific high-frequency artifacts — typically a 'metallic sheen' between 8 kHz and 16 kHz that has unusual harmonic spacing not found in natural recordings. Our signal #1 measures this directly by comparing the spectral envelope above 8 kHz to expected human-recording statistics.
Signal 1: High-frequency artifact density — unusual energy patterns above 8 kHz from AI vocoder output. Signal 2: Spectral flatness variance — does the spectral character change naturally over time, or is it suspiciously constant? Signal 3: Onset timing consistency — do note onsets have natural human micro-timing variation, or are they too regular? Signal 4: Harmonic-to-noise ratio in vocal range — AI vocals have specific HNR signatures. Signal 5: Crest factor / dynamic range — AI mixes are often more compressed than human mixes.
Mostly yes. Mastering changes loudness, EQ, and stereo width but doesn't remove the underlying AI vocoder fingerprint above 8 kHz or the unnaturally consistent timing. Mastering may slightly compress the dynamic-range signal, but the other 4 signals remain. Heavy mastering with aggressive saturation and compression can produce false negatives in some cases.
No. This is a screening tool with statistical accuracy in the 70-85% range. For legal use cases (copyright disputes, distribution disputes, content authenticity claims) you need a forensic-grade audio analysis from a qualified expert. Our tool is appropriate for: deciding which tracks to investigate further, screening submissions, satisfying personal curiosity. Not appropriate for: courts, contract enforcement, official disclosures.
About 5-15 seconds for a typical 3-minute song. The pipeline decodes the audio, computes a spectrogram, runs all 5 signal analyses in parallel, and presents results. There's no model download, no server round-trip, and no warmup — just JavaScript DSP. Larger files take proportionally longer (a 10-minute song may take ~30 seconds).
Partially. 4 of the 5 signals work on any audio (high-freq artifacts, spectral flatness, timing consistency, dynamic range). The harmonic-to-noise ratio signal is most informative on vocal tracks; on instrumentals it's still computed but contributes less weight. Result: instrumental analysis is slightly less accurate but still useful.
Look at the per-signal breakdown the tool shows you. If only one signal flagged but the others were clean, the verdict is weak. If 4 of 5 signals flagged, the verdict is strong. This tool is designed to be transparent — every measurement is shown to you. Don't trust a single number; trust the pattern across signals. And remember: no detector is 100% accurate.
Related tools
Suno Sample Detector
Already know it's AI? Check whether the lyrics accidentally copy a real song.
Suno Audio Fixer
Made a Suno track? Strip the metallic AI sheen this detector would otherwise flag.
Free AI Mastering
Master your AI track for streaming. Doesn't hide AI signals but levels the loudness.
AI Lyric Video Maker
Make a TikTok-ready lyric video from your AI-generated track.
Spotify Canvas Maker
Auto 8-second Spotify Canvas from any AI track, sized to spec.
Suno Prompt Guide
Write better Suno prompts to produce more natural-sounding output.