-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
LSTM training is super slow on GPU #34
Comments
Thank you for reporting this @phgilde . Are you running this script on Windows or WSL? |
@PatriceVignola I'm running this on windows |
We've implemented the single-step/block-based LSTM/GRU/RNN ops, but these are really better suited to CPU architectures. Models typically use the multi-step cuDNN ops when executing on a GPU device. It's not unsurprising that there's some more work here to make DML perform better with recurrent networks. |
@phgilde What GPU you're running this with? You mentioned standard tensorflow and that your config is with Intel HD graphics. Is this training script running on CPU? |
I have same problem with 4750u amd apu , also gpu load not even %1-2 |
@onurberkay What does |
This training loop takes more than a second per epoch using tensorflow-directml but a fraction of a second with standard tensorflow.
It actually doesnt work at all (error is NaN after a couple of iterations) but I already opened another Issue for that.
Code:
System: Intel i5-7200U with Intel HD graphics 620
The text was updated successfully, but these errors were encountered: