-
Notifications
You must be signed in to change notification settings - Fork 206
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA Out of Memory Error on a 32GB GPU when Running trainer.py for Fine-Tuning #141
Comments
Hi What is your bs and your config.yaml? And can you try with more GPUs? |
Hello, thank you for your response.
Would this configuration be sufficient for the training process with this dataset size? Thank you for your assistance! |
Hi. I forget the exact hardware requirement for fine-tuning DynamiCrafter-1024 model. It may be hard for a single 32GB GPU... Multiple 32GB GPUs would be possible. |
Thank you! |
Hi, I'm encountering a
Configuration:
What I’ve Tried:
Despite these efforts, the error persists. Any insights into why this might be happening, or additional suggestions to troubleshoot? |
Hi Did you try to fine-tune the DynamiCrafter-512 model and check the memory usage? |
Hi,I try to |
Hello. The issue persists; what is going on? |
Description:
Hello, I encountered a
torch.cuda.OutOfMemoryError
while fine-tuning a model usingtrainer.py
. My setup includes only a single GPU with 32GB of memory, and the error occurs even at the beginning of training.modify trainer.py:
Error Message:
This error occurred at:
Steps to Reproduce:
trainer.py
with a single 32GB GPU.My Setup:
Solutions Tried:
Thank you for your assistance!
The text was updated successfully, but these errors were encountered: