Skip to content

Pull requests: ggml-org/llama.cpp

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

llava : fix occasional undefined behavior crash examples
#9078 opened Aug 18, 2024 by jart Loading…
2 of 4 tasks
ggml : make GeLU faster and more accurate on CPU ggml changes relating to the ggml tensor library for machine learning Review Complexity : High Generally require indepth knowledge of LLMs or GPUs
#8878 opened Aug 5, 2024 by jart Loading…
Introduce ggml_syncthreads() performance Speed related topics Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level
#7455 opened May 22, 2024 by jart Loading…
Check for llama_get_logits_ith() errors android Issues specific to Android examples Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level server
#7448 opened May 21, 2024 by jart Loading…
Clamp out of range values in K quantizer bugfix fixes an issue or bug model Model specific Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level
#6888 opened Apr 25, 2024 by jart Draft
llamafile : improve moe prompt eval speed on cpu enhancement New feature or request ggml changes relating to the ggml tensor library for machine learning Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level
#6840 opened Apr 23, 2024 by jart Loading…
ProTip! Mix and match filters to narrow down what you’re looking for.