user_id int64 | stimuli_id string | stimuli_type int64 | answer int64 | correct int64 | know_speaker int64 | age float64 | gender float64 | first_language float64 | num_stimuli_seen float64 |
|---|---|---|---|---|---|---|---|---|---|
26 | 1_F07 | 1 | 1 | 1 | 1 | 3 | 1 | 0 | null |
26 | 4_M12_M15B | 4 | 1 | 0 | 0 | 3 | 1 | 0 | 1 |
26 | 4_F11_F14B | 4 | 0 | 1 | 0 | 3 | 1 | 0 | 2 |
26 | 2_M29C | 2 | 0 | 0 | 0 | 3 | 1 | 0 | 3 |
26 | 4_M28_M27D | 4 | 0 | 1 | 0 | 3 | 1 | 0 | 4 |
26 | 2_M46E | 2 | 0 | 0 | 0 | 3 | 1 | 0 | 5 |
26 | 4_M29_M27E | 4 | 1 | 0 | 0 | 3 | 1 | 0 | 6 |
26 | 6_M14B_M11B_002 | 6 | 1 | 0 | 0 | 3 | 1 | 0 | null |
26 | 6_M06B_M10B_022 | 6 | 1 | 0 | 0 | 3 | 1 | 0 | null |
26 | 6_M09A_M06A_054 | 6 | 1 | 0 | 0 | 3 | 1 | 0 | null |
26 | 2_M24B | 2 | 1 | 1 | 0 | 3 | 1 | 0 | 7 |
26 | 4_M17_M16E | 4 | 0 | 1 | 1 | 3 | 1 | 0 | 8 |
7,884 | 4_F11_F13C | 4 | 0 | 1 | 1 | 2 | 0 | 0 | 1 |
7,884 | 5_M50_M49D | 5 | 0 | 1 | 0 | 2 | 0 | 0 | null |
7,884 | 6_M42A_M44A_051 | 6 | 0 | 1 | 0 | 2 | 0 | 0 | null |
7,884 | 6_F12A_F13A_084 | 6 | 0 | 1 | 0 | 2 | 0 | 0 | null |
7,884 | 6_M07A_M08A_067 | 6 | 0 | 1 | 1 | 2 | 0 | 0 | null |
7,884 | 6_M01B_M02B_011 | 6 | 0 | 1 | 1 | 2 | 0 | 0 | null |
7,884 | 6_F29B_F26B_007 | 6 | 0 | 1 | 0 | 2 | 0 | 0 | null |
7,884 | 6_M09B_M07B_066 | 6 | 0 | 1 | 1 | 2 | 0 | 0 | null |
7,884 | 4_M50_M46D | 4 | 0 | 1 | 1 | 2 | 0 | 0 | 2 |
7,884 | 6_M05A_M01A_079 | 6 | 1 | 0 | 0 | 2 | 0 | 0 | null |
7,884 | 4_M27_M26C | 4 | 0 | 1 | 0 | 2 | 0 | 0 | 3 |
44,047 | 6_M49B_M50B_078 | 6 | 1 | 0 | 0 | 3 | 1 | 1 | null |
44,047 | 5_M03_M02C | 5 | 0 | 1 | 0 | 3 | 1 | 1 | null |
44,047 | 6_M30A_M26A_042 | 6 | 0 | 1 | 0 | 3 | 1 | 1 | null |
44,047 | 6_M44A_M41A_073 | 6 | 1 | 0 | 0 | 3 | 1 | 1 | null |
44,047 | 6_M22B_M23B_057 | 6 | 1 | 0 | 0 | 3 | 1 | 1 | null |
44,529 | 6_M45A_M43A_014 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
44,529 | 6_F24A_F22A_068 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
44,529 | 6_F29B_F26B_070 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
44,529 | 6_F19B_F16B_037 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
44,529 | 6_F37A_F39A_022 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
44,529 | 6_M20B_M19B_066 | 6 | 1 | 0 | 0 | 2 | 1 | 0 | null |
48,476 | 6_M08A_M09A_086 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
48,476 | 4_F01_F02E | 4 | 0 | 1 | 0 | 2 | 1 | 1 | 1 |
48,476 | 6_F28A_F29A_024 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
48,476 | 6_M20A_M16A_094 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
52,359 | 6_F15A_F13A_005 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
52,359 | 5_M43_M45C | 5 | 0 | 1 | 0 | 2 | 1 | 1 | null |
52,359 | 6_M06B_M10B_028 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
52,359 | 5_M44_M41B | 5 | 0 | 1 | 0 | 2 | 1 | 1 | null |
52,359 | 2_F38E | 2 | 0 | 0 | 0 | 2 | 1 | 1 | 1 |
53,341 | 6_F48B_F50B_049 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F33A_F35A_068 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F41A_F44A_033 | 6 | 0 | 1 | 1 | 2 | 1 | 1 | null |
53,341 | 6_M16A_M19A_060 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M42B_M41B_090 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 4_F45_F41B | 4 | 0 | 1 | 0 | 2 | 1 | 1 | 1 |
53,341 | 4_F12_F13E | 4 | 0 | 1 | 0 | 2 | 1 | 1 | 2 |
53,341 | 6_M25A_M21A_053 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F08B_F07B_063 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M15B_M11B_022 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F18A_F19A_052 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 2_F50C | 2 | 0 | 0 | 0 | 2 | 1 | 1 | 3 |
53,341 | 6_M05B_M04B_074 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M13A_M11A_052 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M07A_M10A_025 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 2_F10D | 2 | 0 | 0 | 0 | 2 | 1 | 1 | 4 |
53,341 | 2_M09C | 2 | 1 | 1 | 0 | 2 | 1 | 1 | 5 |
53,341 | 6_M42A_M42A_100 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M23B_M25B_048 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M40B_M39B_032 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M16B_M18B_038 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 4_F43_F44D | 4 | 0 | 1 | 0 | 2 | 1 | 1 | 6 |
53,341 | 6_M19A_M20A_031 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M04B_M05B_002 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M26B_M30B_027 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M14A_M13A_070 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M36A_M37A_013 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F15A_F13A_025 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F17B_F18B_078 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M21A_M24A_091 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 2_M09E | 2 | 1 | 1 | 0 | 2 | 1 | 1 | 7 |
53,341 | 6_F37B_F39B_061 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_F13A_F15A_064 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M33B_M35B_050 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 6_M02A_M05A_086 | 6 | 1 | 0 | 0 | 2 | 1 | 1 | null |
53,341 | 1_F26 | 1 | 1 | 1 | 0 | 2 | 1 | 1 | null |
53,341 | 2_M11B | 2 | 0 | 0 | 0 | 2 | 1 | 1 | 8 |
53,341 | 6_M21B_M25B_043 | 6 | 0 | 1 | 0 | 2 | 1 | 1 | null |
53,937 | 6_F23B_F24B_046 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 6_M23B_M25B_075 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 6_M37B_M36B_049 | 6 | 1 | 0 | 1 | 7 | 2 | 2 | null |
53,937 | 4_F48_F49C | 4 | 0 | 1 | 0 | 7 | 2 | 2 | 1 |
53,937 | 6_M01B_M05B_084 | 6 | 1 | 0 | 1 | 7 | 2 | 2 | null |
53,937 | 6_M14B_M13B_089 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 1_M30 | 1 | 0 | 0 | 1 | 7 | 2 | 2 | null |
53,937 | 6_F21B_F22B_076 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 6_M07A_M06A_057 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 6_M18B_M16B_093 | 6 | 0 | 1 | 0 | 7 | 2 | 2 | null |
53,937 | 2_F14E | 2 | 0 | 0 | 0 | 7 | 2 | 2 | 2 |
54,223 | 6_M32B_M35B_050 | 6 | 0 | 1 | 0 | 2 | 1 | 0 | null |
54,223 | 4_F01_F02E | 4 | 0 | 1 | 1 | 2 | 1 | 0 | 1 |
54,870 | 2_M34B | 2 | 1 | 1 | 0 | 2 | 0 | 1 | 1 |
54,870 | 2_F41D | 2 | 0 | 0 | 0 | 2 | 0 | 1 | 2 |
54,870 | 6_M44B_M43B_072 | 6 | 0 | 1 | 0 | 2 | 0 | 1 | null |
54,870 | 6_F10B_F07B_086 | 6 | 1 | 0 | 0 | 2 | 0 | 1 | null |
54,870 | 3_F38C | 3 | 1 | 1 | 0 | 2 | 0 | 1 | null |
54,870 | 6_M39A_M36A_018 | 6 | 0 | 1 | 0 | 2 | 0 | 1 | null |
VIPBench: A Human-Aligned Benchmark for Voice Identity Perception in the Age of Voice Cloning
VIPBench is a benchmark of 124,876 same/different identity judgments from 1,290 English-speaking listeners on 9,800 voice pairs spanning 100 demographically-stratified speakers. Pairs cover three stimulus families: real recordings, AI voice clones generated by a state-of-the-art TTS system, and continuously morphed voices.
The benchmark evaluates whether speaker-embedding and speech-representation models align with human voice-identity perception, providing a perceptual evaluation target distinct from metadata-label speaker identification.
Anonymized release for NeurIPS 2026 Evaluations & Datasets Track double-blind review. Author identities and permanent URLs will be added at camera-ready.
Dataset summary
| Item | Count |
|---|---|
| Speakers | 100 (50 M / 50 F, 5 sociophonetic groups, 2 age brackets) |
| Reference audio clips | 100 (one per speaker) |
| Comparison audio clips | 9,800 (98 per speaker) |
| Voice pairs | 9,800 |
| Listener judgments | 124,876 |
| Listeners | 1,290 |
| Median judgments per pair | 10 (range 8 to 92) |
| Stimulus types | 6 (real same/different, AI clones, voice morphs) |
| Pre-extracted speaker embeddings | 10 models |
| Per-layer SSL embeddings | 5 models |
Supported tasks
The benchmark defines four evaluation tasks:
- Predict listener agreement rate (continuous regression). Predict
P(same)per pair. Metric: Pearson r, R^2 against the human consensus, bounded by the Spearman-Brown noise ceiling rho_SB = 0.705. - Human-aligned binary verification. Classify pairs against the human majority vote. Metrics: AUC (ranking) and Platt-calibrated ECE (calibration).
- Representational similarity (RSA). Spearman correlation between human and model representational dissimilarity matrices, with a Mantel permutation test.
- Real-to-synthetic transfer. Whether a predictor fit on real-speech pairs still works on voice clones and morphs.
A 10-fold gender-balanced speaker-level cross-validation protocol prevents speaker leakage.
Dataset structure
data/
speakers.csv # 100 rows: speaker id, name, group, gender, age
stimuli.csv # 9,800 rows: per-pair aggregates (P(same), votes, type)
participant_responses.csv # 124,876 rows: per-judgment records
stimuli_interpol.csv # 8,100 rows: morph-trajectory metadata for Type 6
audio/
reference/ # 100 *R.wav (16 kHz mono)
comparison/ # 9,800 *.wav
embeddings/
rawnet3.npz, ecapa_tdnn.npz, titanet.npz, xvector.npz, resemblyzer.npz,
wav2vec2.npz, hubert.npz, wavlm.npz, xlsr.npz, whisper.npz
layers/ # per-layer (mean-pooled) for SSL models
wav2vec2.npz, hubert.npz, wavlm.npz, xlsr.npz, whisper.npz
samples/ # 5-speaker quick-look subset (~150 MB)
code/ # 10 extraction scripts + analysis notebook + reproduce.sh
docs/ # annotation protocol, schemas, model table, reproduction
For column-level dictionaries see docs/data_dictionary.md. For the six stimulus types see docs/stimulus_types.md. For the listening-study protocol see docs/annotation_protocol.md.
Embedding format
Each .npz is a key-value store keyed by audio basename without the .wav extension (e.g., M01R, 1_F01, 4_M12_M15B). Values are numpy arrays of shape (embedding_dim,) for the 10 main embeddings and (num_layers, embedding_dim) for the per-layer bundles. The 9,900 keys cover 100 references plus 9,800 comparisons.
Pairing reference and comparison
Each row of data/stimuli.csv represents one voice pair. The reference column gives the reference speaker ID (e.g., M01) and the id column gives the stimulus identifier of the comparison clip (e.g., 1_M01, 4_M12_M15B). The pairing rule is:
| You want | Reference clip | Comparison clip |
|---|---|---|
| Audio file | data/audio/reference/{row.reference}R.wav |
data/audio/comparison/{row.id}.wav |
| Embedding key | {row.reference}R (e.g., M01R) |
{row.id} (e.g., 1_M01) |
Cosine-similarity scoring against P(same):
import numpy as np, pandas as pd
from sklearn.metrics.pairwise import cosine_similarity
stim = pd.read_csv('data/stimuli.csv')
emb = dict(np.load('data/embeddings/ecapa_tdnn.npz'))
stim['cos'] = stim.apply(
lambda r: cosine_similarity(
emb[f'{r.reference}R'].reshape(1, -1),
emb[r.id].reshape(1, -1)
)[0, 0],
axis=1,
)
stim['p_same'] = stim['same_vote'] / stim['num_response']
print(stim[['cos', 'p_same']].corr()) # Pearson r against listener consensus
Quick start
pip install -r requirements.txt
cd code && bash reproduce.sh # ~10 min from cached embeddings
To re-extract embeddings from the audio (~24 CPU-hours plus ~1 GPU-hour for Whisper), see docs/reproduction.md.
Source data and collection
- Speakers. 100 English-speaking US celebrities stratified across 5 sociophonetic groups (1 = New York City English, 2 = Southern American English, 3 = African American English, 4 = Latino English, 5 = Asian American English) x 2 genders x 2 age brackets (1 = under 45, 2 = 55 or older), 5 speakers per cell.
- Reference audio. Clips selected from publicly available recordings (interviews, podcasts).
- Voice clones. Generated with Cartesia (a state-of-the-art TTS system) seeded from a natural source clip of the speaker being cloned. The variant letter in the stimulus ID identifies the seed: a Type 3 clone shares its seed clip with the comparison clip of the matched Type 2 pair, and a Type 5 clone shares its seed with the matched Type 4 pair (e.g., the clone in
3_F01Bis seeded from the same F01B source clip used as the comparison in2_F01B). - Voice morphs. For each of the 100 reference speakers, the latent voice representation of the reference speaker is interpolated toward each of 4 within-group comparison speakers (matched on sociophonetic group, age group, and gender), at 2 distinct recordings per comparison speaker, sampled at 10 morph scales between 0 and 1, plus 1 anchor at scale 1. This yields 4 x 2 x 10 + 1 = 81 Type-6 stimuli per reference speaker (8,100 total). Generated using the voice-morphing feature of the same Cartesia TTS system.
- Listeners. 1,290 adult English-speaking participants recruited via the Centaur AI platform under an IRB-approved protocol. Consent followed the platform's standard pipeline.
Each pair received at least 8 judgments; real-speech pairs (Types 1, 2, 4) received more coverage than synthetic pairs to give tighter consensus estimates on the real-speech reference distribution.
Considerations for use
Personally identifying information
The dataset names public-figure speakers because the celebrity-stratified design is integral to the benchmark and source recordings are already public. Listener identifiers in participant_responses.csv are pseudonymized integers tied to no external account.
Biases and limitations
- English-speaking listener pool, US-dialect speakers. Cross-language perception is not measured.
- 100 speakers limits statistical power for some subgroup contrasts (20 speakers per sociophonetic group).
- Studio-quality audio. In-the-wild conditions (noise, codec compression, telephony) are not represented.
- The operational target is a population consensus, appropriate for ambiguous stimuli where any absolute identity label would itself be probabilistic.
Responsible use
The benchmark measures model-human alignment at the evaluation level. We do not release clone-generation recipes or adversarial training targets. Voice-cloning systems that better align with human perception could inform adversarial use; the same alignment knowledge also strengthens defenses (perception-aligned identity models can flag clones that metadata-based verification accepts).
License
- Dataset (audio, judgments, metadata, embeddings): Creative Commons Attribution-NonCommercial 4.0 International (CC-BY-NC 4.0). See
LICENSE. - Code (scripts, notebook): MIT License. See
LICENSE-CODE. - Pretrained model weights (loaded by extraction scripts): each baseline retains its original license; see
docs/model_table.md.
Commercial use of the audio, judgments, or derived embeddings is not permitted under this license.
Citation
To be filled in at camera-ready.
@inproceedings{vipbench2026,
title = {VIPBench: A Human-Aligned Benchmark for Voice Identity Perception in the Age of Voice Cloning},
author = {Anonymous},
booktitle = {Advances in Neural Information Processing Systems Datasets and Benchmarks},
year = {2026},
note = {Anonymized for double-blind review.}
}
Files
README.md(this file): dataset card.LICENSE: CC-BY-NC 4.0 full text.LICENSE-CODE: MIT full text for scripts.croissant.json: MLCommons Croissant 1.0 metadata (core + Responsible AI fields).DATASHEET.md: Datasheet for Datasets (Gebru et al. 2021).CHANGELOG.md: version history.CITATION.cff: machine-readable citation.requirements.txt: pinned Python dependencies.data/,samples/,code/,docs/: see structure section above.
- Downloads last month
- 83