Vocal Remover Vocals BPM Finder BPM Key Finder Key Track ID ID
30 min remaining

Get more Cloud Assist minutes
Pricing

Music Analysis
Technology

TuneLab computes tempo, key, mood, and structure directly from the audio waveform — not cached metadata from a decade ago. Custom deep learning models analyse everything from lo-fi recordings to complex polyrhythms and heavily compressed masters. Every model runs identically on your device and via API.

Raw Audio Decode Spectral Analysis Neural Inference Sequence Modeling Structured Results
9.3M
Tracks Resolved
652K+
DJ Sets Analysed
±0.1
BPM Precision
32
Free Audio Tools
Layer 1 — The Engine
Beat Detection

Float-Precision Tempo. Not a Rounded Integer.

TuneLab feeds audio through a multi-resolution spectral analysis stage, then into a recurrent neural architecture with long-term temporal context that outputs beat and downbeat probabilities across the full track.

A probabilistic sequence model then finds the globally optimal beat sequence — no greedy peak-picking, no dropped beats. The result: float-precision BPM (±0.1), half/double-tempo alternatives, and per-beat timestamps. Handles syncopation, polyrhythms, tempo changes, and even tracks with no clear downbeat.

±0.1 BPM precision Sub-beat timestamps Tempo change detection Syncopation-aware
MULTI-RESOLUTION SPECTRAL ANALYSIS BEAT ACTIVATION (NEURAL OUTPUT) BEATS ▼ DOWNBEAT 124.03 BPM
Key Detection

Pitch-Class Spectral Analysis. 82.1% on GiantSteps.

A custom pitch-class spectrogram feeds a convolutional network trained on pitch-class spectrograms for 24-class key classification (12 major + 12 minor). No external signal processing libraries — the entire spectral frontend is purpose-built for key detection.

Tested against the GiantSteps-MTG benchmark (the standard dataset for key detection evaluation), TuneLab achieves 82.1% accuracy — significantly outperforming classical template-matching approaches. Camelot wheel mapping is computed automatically from the 24-class output.

82.1% GiantSteps accuracy 24-class output Camelot mapping Zero external dependencies
1 2 3 4 5 6 7 8 9 10 11 12 B A PITCH-CLASS SPECTROGRAM → NEURAL CLASSIFIER C C# D D# E F F# G G# A A# B C major · 8B 82.1% acc NEURAL KEY CLASSIFIER · 24-CLASS
Mood & Genre Intelligence

Six Features. Computed from Audio. Not Metadata.

A transformer-based audio embedding model, pre-trained on 100K+ annotated tracks, produces a rich spectral representation from a 30-second clip. Specialised regression heads fine-tuned on curated reference data then extract six continuous features.

Energy, danceability, happiness, acousticness, instrumentalness, speechiness — each genuinely computed from the audio waveform, not looked up in a database. Holdout metrics are published because TuneLab's numbers hold up to scrutiny.

6 audio features Transformer backbone Curated training data Published holdout metrics
AUDIO EMBEDDING → FEATURE HEADS 30s AUDIO EMBED transformer SPECTRAL Energy 0.82 Danceability 0.74 Happiness 0.65 Acousticness 0.12 Instrumental 0.88 Speechiness 0.05 HOLDOUT METRICS (10K TEST SET) Energy r=0.88 Dance r=0.82 Happy r=0.78 Acoustic r=0.91 Instrum r=0.87 Speech r=0.84 Pearson correlation against ground truth. Disclosed because TuneLab's numbers hold up.
Song Structure

TuneLab Knows Where the Drop Is.

A self-similarity analysis across the full track detects structural repetition, then a novelty-based boundary detector locates transitions between sections. Hierarchical clustering labels each segment: intro, verse, chorus, drop, breakdown, outro.

Every section comes with start/end timestamps and a confidence score — the kind of structural data that rhythm games, VJ tools, and DJ software need but lost when Spotify deprecated their /audio-analysis endpoint.

Full-track analysis Section labels Timestamped boundaries Confidence scores
STRUCTURAL SELF-SIMILARITY ANALYSIS SECTIONS intro verse chor verse chor 0:00 1:24 2:58 3:42 // API response {"label": "chorus",  "start_s": 84.2,  "end_s": 112.7,  "confidence": 0.91} NOVELTY CURVE ▲ structural boundaries
Stem Separation

Two Models. Four Stems. GPU-Accelerated.

An ensemble of state-of-the-art transformer and hybrid architectures powers TuneLab's separation pipeline. Two-stem vocal isolation produces studio-quality splits in roughly 30 seconds; full 4-stem separation (vocals, drums, bass, other) completes in approximately 45 seconds.

Both run on dedicated GPU containers with automatic scaling, custom-optimised for high-throughput inference. Results are stored as presigned URLs with 24-hour expiry. No audio is retained beyond that window.

Transformer ensemble (2-stem) Hybrid architecture (4-stem) GPU-accelerated ~30s processing
4-STEM SEPARATION (HYBRID ARCHITECTURE) MIX (INPUT) VOCALS DRUMS BASS OTHER GPU · ~45s

Active Inference vs Cached Metadata

Capability Spotify Audio Features (deprecated) TuneLab
Method Cached lookup from ~2014 models Active inference from source waveform
Model era Echo Nest (~2014), frozen 2022–2024, continuously updated
Tempo precision Rounded integer (116) Float-precision (116.01), ±0.1 BPM
Key accuracy Undisclosed 82.1% on GiantSteps-MTG benchmark
Beat grid Deprecated with /audio-analysis Float-precision timestamps, per-beat
Song structure Deprecated with /audio-analysis Section labels + timestamps + confidence
Status 403 Forbidden (Nov 2024) Production API with published changelog
Layer 2 — The Graph
Data & Intelligence

9.3M Tracks Resolved. Growing with Every Query.

TuneLab maintains a continuously growing catalog of acoustic data across three tiers — each serving a different layer of the intelligence stack. Cache misses trigger real-time DSP and are cached permanently.

DATA GRAPH — THREE TIERS CANONICAL CATALOG 9.3M cross-platform IDs Streaming · Video · Databases ISRC · Cross-platform IDs 6.8M with BPM + key DJ PERFORMANCE ARCHIVE 652K+ sets analysed 9.8M track entries 2.3M distinct tracks Co-occurrence + transitions ACOUSTIC VERIFICATION 96K+ tracks verified 3,862 sets processed 3,445 distinct artists 11 identification platforms CONTINUOUSLY GROWING · CROSS-REFERENCED · VERIFIED
Canonical Catalog
9.3M cross-platform IDs
Cross-platform IDs across major streaming services, video platforms, and music databases — linked and resolved. 6.8M tracks with pre-computed BPM, key, and Camelot notation.
DJ Performance Archive
652K+ sets
9.8M track entries across 2.3M distinct tracks. BPM progression, harmonic transitions, and co-occurrence data from the largest analysed DJ set corpus available.
Acoustic Verification
96K+ audio-verified tracks
3,862 DJ sets processed through a multi-provider identification cascade across 11 platforms. 3,445 distinct artists. Ground truth from live audio, not metadata.
Layer 3 — Applications
Real-Time Sync

Live Beatmatching. Two Streams. 10ms Tick.

A digital phase-locked loop continuously tracks tempo and phase across two live audio streams, correcting pitch in real time to keep them beatmatched. The control loop runs at a 10ms tick rate with ±2% pitch bend range — fast enough to lock onto tempo drift within seconds.

This is the engine behind the live radio demo — two independent internet radio streams, beatmatched automatically, mixed live in the browser. No pre-analysis, no metadata. Just the audio signal.

10ms control loop ±2% pitch bend Live stream input Live demo →
DIGITAL PHASE-LOCKED LOOP (DPLL) STREAM A · 122.4 BPM STREAM B · 122.1 BPM (DRIFTING) PHASE ERROR (ms) DRIFT DETECTED LOCKED PITCH CORRECTION +0.12% BEND STABLE
Mix & Transition Analysis

WASM-Accelerated Pipeline. Real Transition Forensics.

TuneLab's mix analyser runs a multi-stage WASM-accelerated pipeline on actual audio: multiple independent detection methods work in concert through a voting system to locate transitions with sub-second precision. Phase-locked beat analysis then measures drift in milliseconds at each one.

Multiple genre profiles — each with calibrated tolerances and scoring weights — produce a composite assessment: technical precision, harmonic compatibility, energy flow, and EQ quality, graded A to F with a fully transparent breakdown.

Every score is derived from the waveform itself — not from metadata lookups or pre-computed averages. Every transition is timestamped. Every grade is decomposed into sub-scores with transparent weighting. The full methodology is published, and results are reproducible: same audio in, same analysis out.

Multi-detector voting Phase sequence lock Genre-calibrated scoring WASM-accelerated
MULTI-DETECTOR TRANSITION VOTING DETECTOR A — SPECTRAL DETECTOR B — LOW-FREQUENCY DETECTOR C — TEMPORAL TRANSITION 1 TRANSITION 2 T1: drift 8ms · bass 2 bars Camelot 8B→9B · dist 1 Technical 87 · Grade A- T2: drift 3ms · bass 1 bar Camelot 9B→9A · dist 1 Technical 94 · Grade A
Specialist Tools

Purpose-Built for Problems Others Ignore.

Four tools that solve problems no one else addresses — each built on custom spectral analysis, WASM-accelerated pipelines, and proprietary detection algorithms. All processing runs on your device.

AI artifact detection Transcode detection Transition forensics Audio restoration WASM-accelerated
SPECIALIST TOOLS FIX AI AUDIO Harmonic ratio · group delay deviation · adaptive filtering LOSSLESS CHECKER LOSSLESS ✓ Nyquist boundary analysis · 16/20/22 kHz rolloff detection MIX ANALYZER GRADE: A- multi-stage WASM pipeline · per-transition forensics AUDIO REPAIR DENOISE DE-REVERB RESTORE BEFORE AFTER Spectral estimation · gating · envelope reshaping · de-reverb BUILT FOR WHAT'S NEXT
Audio to MIDI
Polyphonic pitch detection that transcribes audio to MIDI notation. Handles overlapping notes, chord voicings, and percussive transients. Studio-grade accuracy from a neural network trained on professional multitrack sessions.
AI Stem Splitter
Neural source separation isolating vocals, drums, bass, and instruments from any mix. WebGPU-accelerated inference achieves near-real-time processing with lossless WAV output at the original sample rate.
Acapella Extractor
Vocal isolation tuned for clean extraction — preserves breath, vibrato, and reverb tails without the ghostly artifacts common in consumer tools. The same separation engine used by the stem splitter, optimized for vocal clarity.
Mix Analyzer
Full DJ mix assessment — automated transition detection, harmonic compatibility scoring, BPM consistency tracking, and energy flow analysis. Genre-calibrated mix quality scoring across 90-minute mixes. Zero upload.
Fix AI Audio
Targets generative-AI artifacts — metallic sheen, phase smearing, spectral discontinuities from tools like Suno and Udio. Adaptive filtering repairs the audio while preserving musical content.
Audio Repair Clinic
Multi-stage restoration — noise reduction, de-reverberation, echo cancellation, and transient recovery. Each stage independent or chained. Diagnoses the problem and routes to the right repair tool.
Lossless Audio Checker
Detects lossy-to-lossless transcodes by analyzing spectral energy distribution. Identifies the hard frequency ceiling that betrays re-encoded files — a proprietary detection method built in-house.
Client-Side Processing

32 Audio Tools. Zero Uploads. No Limits.

Every tool runs directly on your device — custom WASM DSP pipelines compiled from AssemblyScript, ONNX Runtime for neural inference, and the Web Audio API for real-time signal routing. SharedArrayBuffer enables true multi-threaded processing with near-native throughput.

No file ever leaves your machine. No account required. No processing caps. From BPM detection and key analysis to loudness metering and chord recognition — the full pipeline executes locally, with results computed from your actual audio waveform.

32 tools WASM-accelerated ONNX Runtime Zero uploads SharedArrayBuffer
LOCAL DEVICE PROCESSING YOUR DEVICE AUDIO 44.1kHz WEB AUDIO DECODE WASM DSP PIPELINE ONNX RT INFERENCE WORKER THREADS (SharedArrayBuffer) BPM · KEY · BEATS MIX ANALYSIS CHORD · PITCH LOUDNESS VOCAL SEPARATION RESULTS bpm: 124.03 key: 8B · A minor lufs: -14.2 chords: Am → F → C → G FILES NEVER LEAVE YOUR DEVICE 32 TOOLS
Audio Intelligence API

Real Analysis. Not Cached Metadata.

A REST API built on the same DSP and neural inference that powers the tools. Synchronous responses for lookups and analysis — no polling loops, no per-request surprises. Cache hits return in under 100ms; cache misses trigger real-time compute and respond in 1–5 seconds.

Universal track resolution maps any platform ID to all known cross-references plus verified acoustic features — one request, every major streaming service resolved. Beat grids with float-precision timestamps, audio embeddings for custom ML, song structure with section labels, and DJ mix intelligence combining co-occurrence data from 652K+ analysed sets.

REST API Sync responses Audio embeddings 9.3M resolved tracks Universal resolver
AUDIO INTELLIGENCE API REQUEST GET /v1/resolve/{id} GET /v1/beatgrid/{id} GET /v1/embedding/{id} GET /v1/similar/{id} POST /v1/analyze SYNC 200 · RESPONSE { bpm: 124.03, key: "A minor", camelot: "8A", energy: 0.84, beats: [0.48, 0.97…] } ENDPOINT CATEGORIES LOOKUPS resolve · beatgrid · compat <100ms ANALYSIS bpm · key · mood · structure 1–5s PROCESSING stems · master · MIDI async DJ MIX INTELLIGENCE Transition-aware similarity · Co-occurrence from 652K+ sets · BPM + Camelot compat · 96K tracks
Autonomous Systems

The Engine Powers Working Products.

AI DJ Radio — running since 2020. 24/7/365 continuous mix of underground electronic music, with automated track selection driven by energy curves, harmonic compatibility, and genre coherence. No human intervention. No pre-programmed playlists. The engine described on this page was built to make this work.

Coming soon — a real-time audio platform. Powered by TuneLab.

24/7 since 2020 More coming soon Live demo →
AUTONOMOUS SYSTEMS — BUILT ON TUNELAB AI DJ RADIO 24/7 continuous mix ENERGY CURVE COMING SOON Real-time audio platform SHARED INFRASTRUCTURE Beat Detection Real-Time Sync Data Graph Track Resolution Same engine. Different applications. LISTEN LIVE

Engineering Principles

Private by Design

  • Most tools process entirely on your device — audio never leaves your machine
  • Cloud Assist uploads are deleted immediately after processing — no retention, no listening
  • Accounts are only required for cloud processing — everything else works without sign-up

Deterministic & Reproducible

  • Every score is decomposed into sub-scores with transparent weighting — no black-box numbers
  • Same audio in, same analysis out, every time — results are deterministic
  • Holdout metrics and benchmark results disclosed on this page — TuneLab's numbers hold up to scrutiny

Build With Real Music Analysis

32 free audio analysis tools running on your device. A production API with 9.3M resolved tracks. Float-precision tempo, verified key detection, and features computed from the actual waveform.

Unlock Cloud Assist

Offload heavy processing to secure private GPUs. Free account, 30 min/day.

or