You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
But the file greedy_decode.py seems missing. can you please provide the file?
Regarding DPO training
In the paper, it is mentioned that the training was done on a single NVIDIA A40 GPU.
I am currently working on a remote server that has two NVIDIA A40 GPUs with 48 GB of Cuda memory each.
But when I ran the commands given in Step 2.2: Train SFT model with DPO objective, then I received the out-of-memory error. When I made changes to the following variables in dpo-1.yaml, then only I was able to train:
However, I want to clarify that the version of the trl and transformers library mentioned in the requirements file was not working somehow. For me, these versions worked:
trl==0.13.0
transformers==4.46.0
The text was updated successfully, but these errors were encountered:
For the Missing File, greedy_decode.py is actually from a previous version and it is deprecated now. You may use generate.py to generate the pseudo-labels.
Regarding DPO training, I think we don't mention the training is performed on single A40 GPU in our paper. Actually, we use a 8*A40 GPU server to perform the training. We only mention single A40 GPU in our Figure 4, which is used to illustrate the inference speedup.
Hi, I have a similar question. I run the code on 4 A6000 GPUs (each GPU has 48GB memory). I want to know how to change the config files so that I can train the model.
So far, I’ve tried reducing per_device_train_batch_size in sft-0.yaml and setting num_processes in the fsdp.yaml file to 4, but I’m still encountering out-of-memory errors. Are there any other parameters that should be adjusted to reduce the memory usage during training?
Additionally, as far as I know, the model seems to have been full fine-tuned. I’m curious why methods like LoRA weren’t used for fine-tuning instead. Thank you!
Hi
Missing File
I am trying to run the Step 2.3 i.e Sampling pseudo-labels from DPO model using the command
But the file
greedy_decode.py
seems missing. can you please provide the file?Regarding DPO training
In the paper, it is mentioned that the training was done on a single NVIDIA A40 GPU.
I am currently working on a remote server that has two NVIDIA A40 GPUs with 48 GB of Cuda memory each.
But when I ran the commands given in Step 2.2: Train SFT model with DPO objective, then I received the out-of-memory error. When I made changes to the following variables in dpo-1.yaml, then only I was able to train:
However, I want to clarify that the version of the
trl
andtransformers
library mentioned in the requirements file was not working somehow. For me, these versions worked:The text was updated successfully, but these errors were encountered: