-
Notifications
You must be signed in to change notification settings - Fork 881
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] ValueError: 'cuda' is not a valid DistributedType #802
Comments
We released darts version 0.17.0 yesterday. Passing Can try it instead with below and let us know if it works?
|
@dennisbader thanks for the prompt reply. This did the trick in getting the GPU used, but I guess because of PyTorch Lightning something else also broke and I now get this error:
I tried passing |
Hey @gsamaras and thanks for that. This is indeed a bug and happens when using ReduceLROnPlateau -> Lightning-AI/pytorch-lightning#4454 We will fix this soon. For now, you can either use the model without ReduceLROnPlateau or downgrade darts to version 0.16.1. |
No that's fine, I can do it. Thanks again! |
Darts 0.17.1 was released, which fixes both the |
@dennisbader indeed I was able to have this working. I also checked that the documentation was updated, thanks! May I ask if I'll be able to simply use a TPU like:
or it's something that darts won't seamlessly handle (like in the GPU case)? I don't know if TPUs can work with local data (which do not live in the Google cloud to be honest). PS: As a side note: After upgrading darts to 0.17.1, historical_forecasts() take a significant amount of time (40 minutes for < 3.500 data points), while with darts 0.16.1 that would take just a few minutes. I'll investigate further though and open a new issue if needed. |
I think that should work, but if you are on Colab, PyTorch lightning (which Darts relies on), requires taking an extra step to make TPUs work: https://pytorch-lightning.readthedocs.io/en/stable/advanced/tpu.html#colab-tpus I'll close this issue for now as the GPU issue is solved. Don't hesitate to open a new one if you spot other issues. |
Describe the bug
Suddenly, upon relaunching my notebook I wasn't able to train N-Beats on GPU and got
ValueError: 'cuda' is not a valid DistributedType
, without me changing anything in the code.To Reproduce
Install like this in a Jupyter Notebook:
!pip install 'u8darts[torch]
and then try to train any model in GPU, e.g. an N-Beats model like this:
which gives the error:
My instance has a GPU:
Expected behavior
Training in GPU should be possible.
System (please complete the following information):
Could it be that something with the dependency on Torch is happening?
Additional context
Related: #801
The text was updated successfully, but these errors were encountered: