PredANN++ (Entropy, ctx16) — Encoder-only checkpoint

Model description

This model is a finetuned EEG encoder for song identification from 3-second EEG segments (10-class classification) on the NMED‑T dataset.

The encoder is trained with multitask pretraining (masked prediction) using MusicGen Entropy features, followed by finetuning with cross-entropy on song ID.

  • Input: EEG (128 channels, 125 Hz, 3 seconds)
  • Output: Song ID logits (10 classes)

Intended use

  • Research use for EEG-based music recognition
  • Comparing the effect of predictive-information features (Entropy) vs acoustic features

Not intended use

  • Medical diagnosis
  • Any clinical decision making
  • Commercial usage without verifying upstream non-commercial licenses

Training data

  • NMED‑T (Naturalistic Music EEG Dataset – Tempo), 10 songs, 20 subjects, trial=1.

Training procedure (high-level)

  1. Multitask pretraining: encoder-decoder masked prediction of Entropy tokens (50% masking)
  2. Finetuning: encoder-only training for Song ID classification

Evaluation

The repository contains an evaluation script:

  • codes_3s/analysis/evaluate.py

License and upstream dependencies (IMPORTANT)

This checkpoint is trained using features computed with:

  • MusicGen model weights (CC-BY-NC 4.0) :contentReference[oaicite:13]{index=13}

Therefore, this checkpoint is distributed as CC-BY-NC 4.0.

Citation

If you use this model, please cite:

  • The PredANN++ paper (to be added upon publication)
  • NMED‑T dataset paper
  • Audiocraft / MusicGen

Acknowledgements

We borrow and adapt code from multiple repositories. Please see:

  • THIRD_PARTY_NOTICES.md in the GitHub repository.
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support