Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Does xtuner support DPO for InternVL? #943

Open
fabriceyhc opened this issue Oct 7, 2024 · 1 comment
Open

Does xtuner support DPO for InternVL? #943

fabriceyhc opened this issue Oct 7, 2024 · 1 comment

Comments

@fabriceyhc
Copy link

I am trying to do a custom DPO fine-tuning for internvl_v2_internlm2_2b_lora_finetune, but the default config is oriented towards vanilla supervised fine-tuning with images. I tried to compare / incorporate changes from internlm2_chat_1_8b_dpo_full but am running into some issues with the dataset formats supported.

Is this something that xtuner actually supports at the moment?

@hhaAndroid
Copy link
Collaborator

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants