-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
model: llama3.2 #26
Comments
already supported |
Model responds with garbage: ❯ cortex-nightly run https://huggingface.co/hugging-quants/Llama-3.2-3B-Instruct-Q8_0-GGUF/blob/main/llama-3.2-3b-instruct-q8_0.gguf
Validating download items, please wait..
Start downloading: llama-3.2-3b-instruct-q8_0.gguf
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 1138 100 1138 0 0 3942 0 --:--:-- --:--:-- --:--:-- 3951
100 3263M 100 3263M 0 0 20.6M 0 0:02:37 0:02:37 --:--:-- 25.3M
Model Llama-3.2-3B-Instruct-Q8_0-GGUF downloaded successfully!
Starting server ...
Host: 127.0.0.1 Port: 3928
Server started
Model loaded!
Inorder to exit, type `exit()`
> hiya
+"(($>'9%A69G,87>6.#1C3F+0.>.B$;(,2032,+?FF'3**5"*&>69FCGHF((>*F252+'HB2%C<#!;'39,GE&#>?+'97F+,4GE?"1H+%+-?3A3*,+#C070+F:*)2*?260)B?"F)DB+F')63A6+;G$=F$H)6&/4E4"7"6.(&31+B-0A(*#;!&1C2//G-65%*=,5.2D>A6=B$2=<D417(%74'!?,2>FG/.)9&178;73=D<6?''10F/;C%')(*GC!-0!60=D1HA0AG(4E;8*.>0&-*H4E)3"462965,48!&=7H*E+E9=(9A6#.3""5##7HG8#A;81$F<B%1.;73*0#,7GC&9:HH.6(%G+"-"D>72)5#C"E'6:&=C,F19&3=./(06'F$A'F-.CB8/>DF5A*!2"!5?&)$#/.D)E.2&E"6>C/-:CB!0BE5F!8H)9H><?D%3-
,>?AG'=)+757+H'EE=F#$G#6+55)*73=?&^C |
Hi @0xSage, can you share the I tested in my machine and it works well. I just need to update the ctx_len to fit with 16GB Vram. |
Hi @gabrielle-ong, This is model.yml for llama 3.2
|
Model conversion - Llama3.2 - CompleteChecklist (LlamaCpp only)
Out of scope
|
Note: @0xSage 11b is a vision model which we dont support with the current model.yml - I checked with Alex. |
Please make sure this is tracked in the Model Kanban - we should set it as a goal to support all the models. |
Model Requests
https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct
and 11b
Which formats?
The text was updated successfully, but these errors were encountered: