-
I've been banging my head against this for a couple days and I'm still coming up empty. I have multiple modules and multiple GPUs, however this sequence continues to fail. I've narrowed it down to being a problem with the model. I can load it fresh from a file each loop and the error no longer exists. (Pseudocode) ModuleA.to(cuda:1) The error is consistently in the loss output.backward() call if that helps. This error doesn't happen if I load the module from a file each loop. Input data is not the issue, the model isn't correctly switching devices. I've tried many different combinations of code and have tried directly moving from cuda:0 to cuda:1 without luck. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
This was resolved in a recent PR. Bug was found and fixed. Thanks devs. |
Beta Was this translation helpful? Give feedback.
This was resolved in a recent PR. Bug was found and fixed. Thanks devs.