Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

requires_grad_pre_hook on Llava #44

Open
Erland366 opened this issue Feb 14, 2025 · 0 comments
Open

requires_grad_pre_hook on Llava #44

Erland366 opened this issue Feb 14, 2025 · 0 comments

Comments

@Erland366
Copy link
Contributor

There's an issue when training Llava model on latest transformers (4.49.0.dev0) where requires_grad_pre_hook failed on the CLIPEncoder -> specifically this module model.base_model.model.vision_tower.vision_model.encoder

Upon inspection, this is because the inputs on the hook is empty. This means this is a larger issue because CLIPEncoder never receive any input.

I tested bypassing the hook error, and the training result is very bad

Image

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant