-
✅ Ensure app.py audio input sounds like training data
-
✅ Use a huggingface transformer with the dataset
-
Verify that the training spectrogram matches the predict spectrogram
-
Count number of example misses in dataset loading
-
Verify windowing and jitter params in Song Dataset
-
Create an attention-based network
-
✅ Increase parameter count in network
-
Verify that labels really match what is on the music4dance site
-
✅ Read the Medium series about audio DL
-
double check _rectify_duration
-
✅ Filter out songs that have only one vote
-
✅ Download songs from Best Ballroom
-
✅ fix nan values
-
Try higher mels (224) and more ffts (2048)
-
Verify random sample of dataset outputs by hand.
-
Train with non music data and add a non music category
-
Add back class weights
-
Add back multi label classification