You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There's an issue when training Llava model on latest transformers (4.49.0.dev0) where requires_grad_pre_hook failed on the CLIPEncoder -> specifically this module model.base_model.model.vision_tower.vision_model.encoder
Upon inspection, this is because the inputs on the hook is empty. This means this is a larger issue because CLIPEncoder never receive any input.
I tested bypassing the hook error, and the training result is very bad
The text was updated successfully, but these errors were encountered:
There's an issue when training
Llava
model on latest transformers (4.49.0.dev0) whererequires_grad_pre_hook
failed on theCLIPEncoder
-> specifically this modulemodel.base_model.model.vision_tower.vision_model.encoder
Upon inspection, this is because the
inputs
on the hook is empty. This means this is a larger issue becauseCLIPEncoder
never receive any input.I tested bypassing the hook error, and the training result is very bad
The text was updated successfully, but these errors were encountered: