-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Minimum VRAM requirements #62
Comments
I also only have 4090, waiting for an answer |
When 4090 graphics card is running, 0.5B model can run, 1.5B, 3B model,GPU= 4, CUDA OMM,GPU=8, memory overflow, my 4090 8 card is like this |
I was able to run the 0.5B model on an RTX 4090 for a few mins before Ray errored out with OOM. (RAM, not VRAM. I'm running on a machine with 32G RAM). For the 1.5B model, I don't think you can get it to run on the 4090 (unless you can figure out how to quantize to 8-bit or lower). Tried to reproduce it on an A100 40G and failed with GPU OOM. So now I'm trying A100 80G, and training seems to be progressing ok. Peak GPU mem usage is ~64G: |
Yes, you can use GRPO instead of PPO. Here are my results. (#5 (comment)) |
I found that my Flash atteh 2 didn't work how can i solve it |
Can I train the model with a single RTX4090? Is the single GPU mentioned in the readme an H100 or A100 for training, which means I need at least 80G VRAM?
The text was updated successfully, but these errors were encountered: