You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the case of Halide, I remember having a discussion with Halide people, and if I remember, Halide can support LSTMs if you know the number of steps you want to unroll the LSTM (i.e., extent of your time loop is known at compile time), but the trick of using a time loop will not work if you don't know the size at compile time.
I would be very interested in a RNN implementation example using Tiramisu.
Unfortunately due to #217 I cannot explore that myself at the moment.
Also in the tiramisu.github.io website, you claim that Halide cannot represent RNNs but time is just another loop/tensor dimension. Case in point, this seems to be an actual implementation of LSTM in Halide: https://github.com/misaka-10032/Halstm/blob/master/src/layer/lstm.cpp.
One thing I would be very interested in is the wavefront optimisation on stacked RNNs as mentionned in Nvidia's blog post in optimization 3.
They even provide the Cuda source code that can serve as a reference benchmark.
The text was updated successfully, but these errors were encountered: