-
Notifications
You must be signed in to change notification settings - Fork 26
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Discord? #5
Comments
Hey, any ideas and contributions are appreciated! If you do want a private chat though, you can reach me on Discord at |
Just got a chance to try out your model, and the results seem promising. Here is a sample of a conversation I had with your model. It grasps basic GDScript terminology and syntax, but it seems to be setup wrong maybe in my webui. what is the best way to use this model? I am using the 7B model locally on the oobabooga webui. |
The model is finetuned on an instruct-dataset similar to stanford-alpaca and similar models. This means all samples confirm to a specific prompting template, which in the case of godot-dodo is:
I have not tested using the model without that format, and am not familiar with how oobabooga sets things up. So if you want to reproduce the model performance as i evaluated it, you will need to follow the exact prompting template above. You can do this via Google Colab using the Jupyter notebook linked in the readme: https://colab.research.google.com/github/minosvasilias/godot-dodo/blob/main/demo/inference_demo.ipynb |
That looks sensible, though again not sure how exactly they format the context. However, godot-dodo models are unlikely to generate full scripts anyway if you're looking for that. The training dataset is split into individual methods, and the model therefore learns to implement the instructions within the scope of a single method. It will rarely, if ever, exceed that scope. |
I know this is usually unnecessary, but I'd like to help out. Do you have a discord?
I have some ideas about datasets, training that I think could be very useful. I'm also a GDScript veteran at this point.
I've successfully and commercially used GPT2 models in 2019 - so I have experience with datasets.
The text was updated successfully, but these errors were encountered: