-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
It would be cool if one could access self-hosted generative models (Llama, WizardLLM, etc.) #15
Comments
I've forked it and added some of that functionality. Its not complete, so I'm not making a PR for it yet, and I'm not sure he wants the changes :P |
It'd be great if we could host and run our own models with Ollama (https://github.com/ollama/ollama). @minosvasilias I have just seen that you have several Godot code finetunes from about a year ago! :D Would you recommend them over newer models like LLaMA 3? Are you planning on finetuning any of the recently launched models? (and maybe adding it to the Ollama model library?) I'm somewhat baffled that the official Godot devs aren't supporting this initiative more actively ^^U By lowering the entrance barrier and enabling more people to code in Godot, we could all benefit! |
That's amazing @Efoi! Will follow and star your fork - even if the PR won't be accepted here, I'll just use the fork. |
Fixes minosvasilias#15 Add support for self-hosted generative models Llama and WizardLLM. * **Add Llama model support:** - Create `addons/copilot/Llama.gd` script. - Implement `_get_models`, `_set_model`, `_send_user_prompt` functions. - Define constants and helper functions for Llama model. * **Add WizardLLM model support:** - Create `addons/copilot/WizardLLM.gd` script. - Implement `_get_models`, `_set_model`, `_send_user_prompt` functions. - Define constants and helper functions for WizardLLM model. * **Update Copilot.gd:** - Modify `populate_models` function to include Llama and WizardLLM models. - Add Llama and WizardLLM to the `models` dictionary. * **Update CopilotUI.tscn:** - Add Llama and WizardLLM nodes to the `LLMs` node. - Set the script for Llama and WizardLLM nodes to the respective new scripts. - Add signal connections for Llama and WizardLLM nodes. --- For more details, open the [Copilot Workspace session](https://copilot-workspace.githubnext.com/minosvasilias/godot-copilot/issues/15?shareId=XXXX-XXXX-XXXX-XXXX).
Either over the network via Llamacpp/Kobold AI interfaces or running locally on your machine.
With the subscription prices being as they are, hosting your own model can be cheaper.
The text was updated successfully, but these errors were encountered: