DreamSound is a python package for sonic deep dream generation. See our paper, a working colab, and a web version with examples at https://fdch.github.io/dreamsound
Please cite us:
@inproceedings{DreamSound2021,
title={DreamSound: Deep Activation Layer Sonification},
author={Cámara Halac, Federico and Delgadino, Matías},
year={2021},
booktitle={Presented at the 27th International Conference on Auditory Display (ICAD 2021)},
publisher={International Community on Auditory Display },
}
Inspired by the DeepDream project, DreamSound plays a sound file to Yamnet, a pre-trained neural network, and Yamnet returns a dreamed sound.
Internally, DreamSound takes the gradients of a class from the pre-trained yamnet model and filters them with an original sound with some combination technique.
Head to this Google Colab for a quick example on how to get started with the module. An old version can be accessible in this other Google Colab. Yet an older version is here, which goes hand in hand with an early paper we did on the matter of Convolutional Neural Network Activation Layer Sonification, or what we called DreamSound.
Dreamsound depends on the following pip packages you can pip install
:
requests
numpy
matplotlib
IPython
librosa
tensorflow
soundfile
First, install the dependencies
python3 -m pip install -r requirements.txt
Install the dreamsound package using pip:
python3 -m pip install dreamsound
The pip project is hosted at PyPi: https://pypi.org/project/dreamsound/
NOTE: you need to download the yamnet model before importing the dreamsound module. Please, continue reading.
Create a directory for your project and relocate there.
mkdir dream_test
cd dream_test
The Yamnet Downloader file does not come with the pip distribution. However, it is distributed on this repository. If you do not want to clone this repository, simply do: curl -O https://raw.githubusercontent.com/fdch/dreamsound/main/yamnet_downloader.py
on this same directory, and run:
python3 yamnet_downloader.py
Alternatively, you can get Yamnet yourself, crudely like this:
git clone https://github.com/fdch/models.git models
mv models/research/audioset/yamnet/* .
rm -rf models
curl -O https://storage.googleapis.com/audioset/yamnet.h5
You must have the yamnet model on the same directory. Now, you can import the dreamsound module and use the class. This code loads some files from disk and passes them to the DreamSound
class from the dreamsound
module. This looks something like this:
>>> import dreamsound
INFO:tensorflow:Enabling eager execution
INFO:tensorflow:Enabling v2 tensorshape
INFO:tensorflow:Enabling resource variables
INFO:tensorflow:Enabling tensor equality
INFO:tensorflow:Enabling control flow v2
>>> ds = dreamsound.DreamSound(["../audio/original.wav", "../audio/cat.wav"])
Loading audio files...
Done.
I have now 2 audio files in memory.
Using last layer.
Yamnet loaded, using layer:activation_1
Dreamer started.
There are two types of filtering, auto or targetted:
Filter the first audio with it's dreamed self
>>> ds(audio_index=0)
Running step 0, class: Whistling...
...
Writing ./audio/Whistle-9-orig.wav...
Writing ./audio/Whistle-9-diff.wav...
Writing ./audio/Whistle-9-filt.wav...
Writing ./audio/Whistle-9-hard.wav...
Writing ./audio/Whistle-9-grad.wav...
Filter the first with a dreamed target
>>> ds(audio_index=0, tgt=1)
Target class: Animal...
Running step 0, class: Whistling...
Running step 1, class: Whistle...
Running step 2, class: Whistle...
Running step 3, class: Whistle...
Running step 4, class: Whistle...
Running step 5, class: Whistle...
Running step 6, class: Whistle...
Running step 7, class: Whistle...
Running step 8, class: Flute...
Running step 9, class: Whistle...
Writing ./audio/Whistle-9-orig-tgt-Animal.wav...
Writing ./audio/Whistle-9-diff-tgt-Animal.wav...
Writing ./audio/Whistle-9-filt-tgt-Animal.wav...
Writing ./audio/Whistle-9-hard-tgt-Animal.wav...
Writing ./audio/Whistle-9-grad-tgt-Animal.wav...
Finally, you can pass no arguments to continue filtering recursively
>>> ds()
Target class: Animal...
Running step 10, class: Whistle...
Running step 11, class: Whistle...
Running step 12, class: Wind instrument, woodwind instrument...
Running step 13, class: Wind instrument, woodwind instrument...
Running step 14, class: Flute...
Running step 15, class: Flute...
Running step 16, class: Wind instrument, woodwind instrument...
Running step 17, class: Whistle...
Running step 18, class: Whistle...
Running step 19, class: Music...
Writing ./audio/Music-19-orig-tgt-Animal.wav...
Writing ./audio/Music-19-diff-tgt-Animal.wav...
Writing ./audio/Music-19-filt-tgt-Animal.wav...
Writing ./audio/Music-19-hard-tgt-Animal.wav...
Writing ./audio/Music-19-grad-tgt-Animal.wav...
You can change any of the following before or after calling the class:
sr = 22050
max_dur = 10
patch_hop = 0.1
win_length = 2048
hop_length = 128
pad_end = False
loss_power = 0.001
plot_every = 10
figsize = (10,8)
top_db = 80.0
step_size = 0.95
output_type = 3
steps = 10
threshold = 1e-07
classid = None
maxloss = True
elapsed = 0
recurse = False
target = None
power = 1.0
audio_dir = "./audio/"
image_dir = "./image/"
Fede Camara Halac (https://github.com/fdch) Matias Delgadino (https://github.com/zaytam)
YamNet AudioSet