-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Docker gpu mode not working #19
Comments
Of note, the server doesn't have the NUMA notes, but the output was identical otherwise (this log is from a WSL instance). |
Hello, PS: As far as I remember (at least while training) we needed a GPU with at least 11GB VRAM to run bertax. So I would try the changes discussed above on the A30 first if possible. :) |
Hi! Thank you for the response. First, I am spinning up the docker container and setting the entrypoint to bash From the container I confirmed I was on debian 11 x86_64. Then I check the tensorflow version
This confirms that I should need CUDnn 8.0 and CUDA 11.0 I have tried to install both by manual means and using Conda, however I keep getting similar output as the above logs (with some variation depending on the version of CUDA, [I tested up to 11.3]. I could not successfully install CUDA by manual means. Here are things I've tried: Conda:
Manual:The CUDA install page doesn't provide a setup for debian 11 until 11.5, so I was attempting to install CUDA 11.5 Before installing CUDA, I setup add-apt-repository
Then installing gnupg2 Next, I followed the network install instructions for my platform and architecture here. There's a snag at this point:
So I switched to the local runfile
Which also fails. I have hit a wall and am unsure how to proceed. In the meantime I'll keep trying configurations. Thank you and kind regards. |
I tested on a server with an A30 GPU and a laptop with an RTX 3060.
I believe I followed all steps in the setup guide.
The text was updated successfully, but these errors were encountered: