-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Standardized prompting metadata #774
Comments
There is already |
@ggerganov Those seem to cover about the same area, but don't Jinja templates require pre-processing? Can those be used directly as input to -p in llama.cpp, for example? I didn't notice them anywhere in Ollama Modelfiles, either. I am not sure if / how multiple types of templates should be supported. Maybe something like
Or, more flexible approach, just allow any set of types and templates via prompting.templates map. Just an idea, but I feel like standardizing this could make experimenting with different types of models a bit easier. |
The model files are converted from Because HuggingFace uses When the models are converted from You can get a feel for how a I hope this provides some clarity into the current implementation. As an aside, I would've personally used a Mapping, not Jinja2. I have no idea what the rationale behind this was, the consequence is that we're all stuck dealing with it now. |
It would be nice to have standardized prompting metadata defined within GGUF files.
Currently when importing GGUF model to tools like Ollama it's necessary to explicitely provide prompting metadata - like the template and stopping sequences specific to given model, sometimes also default system message. It would be useful to include most commonly used prompting parameters in GGUF specification, so it could be read and used by applications like llama.cpp or Ollama.
My proposal, based on Ollama Modelfile definition:
The text was updated successfully, but these errors were encountered: