anonymous-submission000's picture
Upload dataset
3362622 verified
metadata
dataset_info:
  features:
    - name: index
      dtype: int64
    - name: audio
      dtype:
        audio:
          sampling_rate: 16000
    - name: subset
      dtype: string
    - name: speaker
      dtype: string
    - name: label
      dtype: string
    - name: original_name
      dtype: string
  splits:
    - name: train
      num_bytes: 24343526
      num_examples: 887
  download_size: 22452898
  dataset_size: 24343526
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
license: cc-by-4.0
tags:
  - audio
  - animal-vocalization
  - birdsong
  - zebra-finch
  - perceptual-similarity
  - benchmark
  - zero-shot
  - vocsim
  - avian-perceptual-judgment
  - audio-perceptual-judgment
size_categories:
  - n<1K
pretty_name: VocSim - Avian Perception Alignment

Dataset Card for VocSim - Avian Perception Alignment

Dataset Description

This dataset is used in the VocSim benchmark paper, specifically designed to evaluate how well neural audio embeddings align with biological perceptual judgments of similarity. It utilizes data from Zandberg et al. (2024), which includes recordings of zebra finch (Taeniopygia guttata) song syllables and results from behavioral experiments (probe and triplet tasks) measuring the birds' perception of syllable similarity.

The dataset allows researchers to:

  1. Extract features/embeddings from the song syllables using various computational models.
  2. Compute pairwise distances between these embeddings.
  3. Compare the resulting computational similarity matrices against the avian perceptual judgments recorded in the accompanying probes.csv and triplets.csv files.

This facilitates the development and benchmarking of audio representations that better capture biologically relevant acoustic features.

Included Files:

  • Hugging Face Dataset object containing audio file paths and metadata.
  • probes.csv: Contains results from perceptual probe trials (sound_id, left, right, decision, etc.). Filtered to include only rows where all mentioned audio files exist.
  • triplets.csv: Contains results from perceptual triplet trials (Anchor, Positive, Negative, diff, etc.). Filtered to include only rows where all mentioned audio files exist.
  • missing_audio_files.txt (optional): Lists identifiers from the original CSVs for which no corresponding audio file was found.

Dataset Structure

Data Instances

A typical example in the dataset looks like this:

{
  'audio': {'path': '/path/to/datasets/avian_perception/wavs/ZF_M_123_syllable_A.wav', 'array': array([-0.00024414, -0.00048828, ...,  0.00024414], dtype=float32), 'sampling_rate': 16000},
  'subset': 'avian_perception',
  'index': 42,
  'speaker': 'ZF_M_123',
  'label': 'ZF_M_123', # Label is set to speaker ID for this dataset
  'original_name': 'ZF_M_123_syllable_A.wav' # Identifier as used in CSVs
}

Citation Information

If you use this dataset in your work, please cite both the VocSim benchmark paper and the original source data paper:

@unpublished{vocsim2025,
  title={VocSim: Zero-Shot Audio Similarity Benchmark for Neural Embeddings},
  author={Anonymous},
  year={2025},
  note={Submitted manuscript}
}

@article{zandberg2024bird,
  author = {Zandberg, Lies and Morfi, Veronica and George, Julia M. and Clayton, David F. and Stowell, Dan and Lachlan, Robert F.},
  title = {Bird song comparison using deep learning trained from avian perceptual judgments},
  journal = {PLoS Computational Biology},
  volume = {20},
  number = {8},
  year = {2024},
  month = {aug},
  pages = {e1012329},
  doi = {10.1371/journal.pcbi.1012329},
  publisher = {Public Library of Science}
}