Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU memory consuming. #44

Open
arranclo opened this issue May 24, 2024 · 0 comments
Open

GPU memory consuming. #44

arranclo opened this issue May 24, 2024 · 0 comments

Comments

@arranclo
Copy link

Thank you very much for your contributions, which have been of great help to my research at present. In the process of using this framework for training, I observed a sudden increase in memory usage during validation after training the first epoch (in my case, about 10g of memory was used during training the first epoch, and 17g of memory was used during validation after the first epoch, which did not change thereafter). I speculate that 10g of memory was used for training and 7g of memory was used for validation, but the memory used for training was not released during the validation inference process, resulting in the above phenomenon. Since I haven't thoroughly researched the source code, I'm not sure if this is a necessary operation for the framework, but this issue may lead to meaningless memory usage (for example, a 24GB graphics card can only be treated as a 17GB graphics card under this training/validation framework). And I found that the training script doesn't seem to support multi graphics card separation training, which can to some extent solve the problem of insufficient graphics card memory.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant