-
-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
support chatml-function-calling via llama-cpp #4
Conversation
i would prefer to merge smaller changes rather than having so many items in this PR. also you wont need to be merging and rebasing that frequently. I am working on the markdown fix because the current interpreter.py have some legacy code that is broken and its not really working well with some models right now. I hope to get this done hopefully today |
See the EPIC ticket :) |
sounds good, i'll split them up! |
So it's ok to merge now? |
not yet, i haven't done any regression testing yet, ill let you know soon |
I want to merge this #6 as it's actually fixing the behaviour of code colorization and ive tested it with several models. the code is still not fully cleaned but its much simpler and at least it works as expected. is it ok for you for me to merge it? |
Yeah go ahead, I'll deal with any merge conflicts when this is ready |
ok this works now, merged with your changes. Not getting any good results with mistral, but that's more of a model problem. Hopefully functionary will be a lot better |
Ready to merge to start experimenting with it? I think enabling vectordb is important for the auto mode, as well as extending the doc/data stuff to give more contextual hints about how to achieve things and which commands use |
uh "This branch cannot be rebased due to conflicts" |
Also, note that commit messages must be capitalized, and i would recommend you to squash the commits |
ok should be good now |
yeah, to the extent of if you How do you envision structuring the RAG data? What's the best "use case -> commands" documentation/reference that we can shove into the vectordb? If we just put the r2 docs in there, i don't think we'd get the any good results for high level queries like "solve this crackme" or "what's the password" |
We may probably update to the latest mistral models in the -M output (i filled a ticket for this). ive tested your code and it's not really working well. not sure if there's a way to debug what the model is doing internally to trace what's going on.. maybe via About vectordb, what it does is to send the user prompt to the database and the database returns a list of sentences that can be prepended to the query for contextual information. this data can provide instructions to perform actions or information about the answer the user is looking for. so its transparent to the user and works across all models. i think integrating this in the auto mode will help a lot in the local results too. |
Yeah ive gotten it to work only like once by luck. I'll add some debugging. About the vectors, yeah, I know how RAG works, I mean what text are you thinking of putting in there, do you have examples? |
chatml-function-calling
via llama-cpp