Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Proper convolution mode for fast GPU processing #350

Merged
merged 1 commit into from
Oct 7, 2023

Conversation

nshmyrev
Copy link
Contributor

@nshmyrev nshmyrev commented Oct 5, 2023

Default convolution mode is very slow, as a result sherpa-onnx-gpu is much slower than CPU. With this fix the decoding speed is ok.

See also

microsoft/onnxruntime#12880 (comment)

https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#cudnn_conv_algo_search

@csukuangfj
Copy link
Collaborator

Thank you for your first contribution!

@csukuangfj csukuangfj merged commit c12286f into k2-fsa:master Oct 7, 2023
XiaYucca pushed a commit to XiaYucca/sherpa-onnx that referenced this pull request Jan 9, 2025
# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants