-
Notifications
You must be signed in to change notification settings - Fork 27k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
split single_gpu and multi_gpu #17083
Conversation
The documentation is not available anymore as the PR was closed or merged. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for your PR! Did you launch it as a trial to see if it works? I see the following should be completed, on line 303:
[setup, run_tests_gpu, run_examples_gpu, run_pipelines_tf_gpu, run_pipelines_torch_gpu, run_all_tests_torch_cuda_extensions_gpu]
You are right, that line should be changed. I haven't launched it (just tried with a dummy example). I will launch it now. |
You can launch it with only 1-2 models in each run, for example by updating this line:
to
This way you'll test the full behavior without having 12-hour long iterations. |
It took sometime, but the run looks good. https://github.com/huggingface/transformers/actions/runs/2276209307 |
a9c25d1
to
493b384
Compare
Looks good, thanks @ydshieh! |
* split single_gpu and multi_gpu * update needs in send_result Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
* split single_gpu and multi_gpu * update needs in send_result Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
* split single_gpu and multi_gpu * update needs in send_result Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
What does this PR do?
Fix the scheduled CI issue caused by the 256 limits (jobs generated from matrix).
Note that the workflow run page has a graph that has no single-gpu and multi-gpu on it. But on the left side, the job names have matrix mentioned.