We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
I compiled the latest code in this repo with cuBLAS support as described in the README
It doesn't seem to be utilizing my 1070 although main is running in nvidia-smi
llama_model_load_internal: [cublas] offloading 0 layers to GPU llama_model_load_internal: [cublas] total VRAM used: 0 MB
What am I missing here?
The text was updated successfully, but these errors were encountered:
Add the option -ngl 10, for upload 10 layers to your gpu memory
-ngl 10
Sorry, something went wrong.
No branches or pull requests
I compiled the latest code in this repo with cuBLAS support as described in the README
It doesn't seem to be utilizing my 1070 although main is running in nvidia-smi
What am I missing here?
The text was updated successfully, but these errors were encountered: