Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

Running pretrained model and Generate.ipynb giving 0 second audio outputs #20

Open
nishkakhendry opened this issue Jul 26, 2021 · 2 comments

Comments

@nishkakhendry
Copy link

I am working on Ubuntu 20.04 with jupyter notebooks on my laptop. As I don't have a graphics card, I changed the decoder definition from NVWavenetGenerator to WavenetGenerator and removed .cuda() from the code to make it runnable locally. The code runs without a problem but the output is a bunch of 0 second audio clips. I have tried debugging by placing disp() in the forward of the encoder and decoder but have not yet fixed the problem. Any help or ideas would be appreciated. Thanks in advance!
image

@nishkakhendry nishkakhendry changed the title Generated outputs using pretrained model and Generate.ipynb giving 0 second audio outputs Running pretrained model and Generate.ipynb giving 0 second audio outputs Jul 26, 2021
@DanileRond
Copy link

I got the same problem. Anyone solved it?

@xistva02
Copy link

xistva02 commented Feb 19, 2023

Yes. This happens when you use WavenetGenerator decoder (WavenetGenerator) but keep the parameters of NVWavenetGenerator from the ipython script. Try to modify decoder from "decoder = NVWavenetGenerator(decoder, rate // 20, batch_size, 3)" to "decoder = WavenetGenerator(decoder, batch_size, wav_freq=rate)". The inference is much slower but it seems to work without needing of wavenet compilation via setup.py or make.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants