You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I am a graduate student. I tried to execute my project based on torch_ACA solver in multiple gpus with DP wrapper recently. However, I found that the performance will decrease much compared with that in a single gpu, while the naive backpropagation still works. Can you give me some instructions or possible reasons?
The text was updated successfully, but these errors were encountered:
I believe it's because there's no proper error tolerance or grad reduce operation in the case of data parallel if you have a distributed setup. Sorry I did not wrote that since I don't have much machine.
Hello, I am a graduate student. I tried to execute my project based on torch_ACA solver in multiple gpus with DP wrapper recently. However, I found that the performance will decrease much compared with that in a single gpu, while the naive backpropagation still works. Can you give me some instructions or possible reasons?
The text was updated successfully, but these errors were encountered: