Advice for GPU/cuda in an electron app #173
Replies: 3 comments 4 replies
-
The latest beta includes both CUDA and Vulkan prebuilt binaries for Windows and Linux, so you can utilize the GPU without compiling anything on the user’s machine. In the next few betas I plan to improve the ‘getLlama’ method to automatically detect whether CUDA or Vulkan is supported on the user machine and enable them automatically, to always utilize the GPU support by default without any configuration necessary. At the moment, to fully utilize the GPU on non-Mac machines, you’d have to enable CUDA or Vulkan and pass I also plan to make it automatically offload as much layers that could be fitted into the VRAM of the GPUs automatically by default in one of the next few betas. |
Beta Was this translation helpful? Give feedback.
-
Thanks for your response. Just a couple of clarifying questions:
|
Beta Was this translation helpful? Give feedback.
-
Any update on this @giladgd ? |
Beta Was this translation helpful? Give feedback.
-
Hi there, thanks for this great project! I use it here: https://github.com/reorproject/reor
I was wondering if you had any advice for the best approach to using the GPU (across OS) in the context of an electron app. So far, all I've implemented is to send 100 layers to the gpu on arm macs and there are users asking for gpu support on windows and linux and was interested in how you'd approach this - particularly along the lines of whether I should build node llama cpp with cuda or not.
The issues are these if you're curious: reorproject/reor#31 & reorproject/reor#79
Beta Was this translation helpful? Give feedback.
All reactions