We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
the command i executed : python inference.py -c nuwave2_02_16_13_epoch=629.ckpt -i 5dk.mp3 --sr 16
here results
D:\86 se courses youtube kanali\upsample audio>python inference.py -c nuwave2_02_16_13_epoch=629.ckpt -i 5dk.mp3 --sr 16 Traceback (most recent call last): File "D:\86 se courses youtube kanali\upsample audio\inference.py", line 115, in <module> wav_recon, wav_list = model.inference(wav_l, band, args.steps, noise_schedule) File "C:\Python399\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "D:\86 se courses youtube kanali\upsample audio\lightning_model.py", line 50, in inference signal, recon = self.model.denoise_ddim(signal, wav_l, band, logsnr_t, logsnr_s) File "D:\86 se courses youtube kanali\upsample audio\diffusion.py", line 54, in denoise_ddim noise = self.model(y, y_l, band, norm_nlogsnr) File "C:\Python399\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl return forward_call(*input, **kwargs) File "D:\86 se courses youtube kanali\upsample audio\model.py", line 215, in forward x, skip_connection = layer(x, band, noise_level) File "C:\Python399\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl return forward_call(*input, **kwargs) File "D:\86 se courses youtube kanali\upsample audio\model.py", line 173, in forward y_l, y_g = self.ffc1(y_l, y_g, band) # STFC File "C:\Python399\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl return forward_call(*input, **kwargs) File "D:\86 se courses youtube kanali\upsample audio\model.py", line 152, in forward out_xl = self.convl2l(x_l) + self.convg2l(x_g) File "C:\Python399\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl return forward_call(*input, **kwargs) File "C:\Python399\lib\site-packages\torch\nn\modules\conv.py", line 313, in forward return self._conv_forward(input, self.weight, self.bias) File "C:\Python399\lib\site-packages\torch\nn\modules\conv.py", line 309, in _conv_forward return F.conv1d(input, weight, bias, self.stride, torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 3.43 GiB (GPU 0; 12.00 GiB total capacity; 10.41 GiB already allocated; 0 bytes free; 10.42 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
The text was updated successfully, but these errors were encountered:
failed with same error on 1080-ti (12GB)
but was able to run on gcp w/ A100 GPU, seems to require at least 20GB (for audio ~90seconds)
Sorry, something went wrong.
No branches or pull requests
the command i executed : python inference.py -c nuwave2_02_16_13_epoch=629.ckpt -i 5dk.mp3 --sr 16
here results
The text was updated successfully, but these errors were encountered: