-
Notifications
You must be signed in to change notification settings - Fork 5.8k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
ValueError: Attempting to unscale FP16 gradients. #6442
Comments
You'll need to cast the torch.float16 to torch.float32 for trainable parameters. Refer to this code block: https://github.com/huggingface/diffusers/blob/main/examples/text_to_image/train_text_to_image_lora_sdxl.py#L632-L641 |
Excuse me, what should I do then? |
Should have been fixed by: #6231. Could you please pull in the latest changes? |
I'm using the latest pull and using the |
Cc: @SunMarc |
This is on top of our head. We need tp solve #6510 first and then it should be a breeze. |
Any progress or temporary solution? |
@sayakpaul why do the trainable params need to be in float32? Is this a new requirement with the integration of PEFT or has this always been the case? I can't remember needing to do this in the past. |
That's the case. If you remember, we never manually casted the |
I am trying to resume training a lora in sdxl but when I try to resume it gives an error
ValueError: Attempting to unscale FP16 gradients.
It works the first time but when I resume training it gives me that error
The text was updated successfully, but these errors were encountered: