This is a TensorFlow implementation of the WaveNet generative neural network architecture for audio generation.
The WaveNet architecture directly generates a raw audio waveform, and shows excellent results in TTS and general audio generation (see the DeepMind blog post and paper for examples).
The network is a model of the conditional probability to generate the next sample in the audio waveform, given all previous samples and possibly additional parameters. It is constructed from a stack of causal dilated layers, each of which is a dilated convolution (convolution with holes), which only accesses the current and past audio samples.
The network is implemented in the file wavenet.py
.
This implementation is currently still under ๐ง
TODO:
- Add step-by-step generation of audio samples (and make them available in TensorBoard)
- Add conditioning on extra parameters (tagged speech, speaker, etc.)
- Generate example outputs
The VCTK corpus is currently used.
In order to train the network, you need to download the corpus and unpack it in the same directory as the main.py
script.
Then, execute
python main.py
to train the network.
You can see documentation on the settings by by running
python main.py --help
Disclaimer: This repository is not affiliated with DeepMind or Google in any way.