-
Notifications
You must be signed in to change notification settings - Fork 11.4k
Issues: ggml-org/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Eval bug: HIP: llama.cpp server locks up when running multiple instances on the same gpu
bug-unconfirmed
#12991
opened Apr 17, 2025 by
pl752
Eval bug: Quad P40 unable to run 70B models on recent releases
bug-unconfirmed
#12990
opened Apr 17, 2025 by
iraqigeek
Feature Request: Add kv-quant fa kernel variants for head sizes other than 128
enhancement
New feature or request
#12989
opened Apr 17, 2025 by
pl752
4 tasks done
Misc. bug: Potential memory leak in backend registry
bug-unconfirmed
#12986
opened Apr 16, 2025 by
dorpxam
Feature Request: Make chat sessions possible with multi model cli tools
#12982
opened Apr 16, 2025 by
domasofan
Feature Request: multi model cli tools: Convert submitted images to best size and format for model
#12981
opened Apr 16, 2025 by
domasofan
Misc. bug: Vulkan performance depends on thread priority
bug-unconfirmed
#12976
opened Apr 16, 2025 by
jeffbolznv
Misc. bug: llama-server speculative decoding not as performant as llama-speculative-simple
bug-unconfirmed
#12968
opened Apr 16, 2025 by
hjc4869
Why does /ggml/CMakeLists.txt add_subdirectory(examples)?
bug-unconfirmed
#12963
opened Apr 15, 2025 by
Please-just-dont
Misc. bug: Excessive power draw on the second GPU in dual RTX 3090 setup when idle
bug-unconfirmed
#12958
opened Apr 15, 2025 by
goodglitch
Eval bug: Deepseek V2 Lite no longer working with Vulkan (assert fail during tg)
bug-unconfirmed
#12956
opened Apr 15, 2025 by
stduhpf
Feature Request: Improve model load time when using the RPC backend
enhancement
New feature or request
#12954
opened Apr 15, 2025 by
rgerganov
4 tasks done
Eval bug: OpenAI incompatible image handling in server multimodal
bug-unconfirmed
#12947
opened Apr 15, 2025 by
kerlion
Misc. bug: llama-server "terminate called after throwing an instance of 'std::runtime_error'"
bug-unconfirmed
#12939
opened Apr 14, 2025 by
guertsen
Misc. bug: The llama-server not read the "--keep" param that user input in the cli
bug-unconfirmed
#12927
opened Apr 13, 2025 by
ZUIcat
Feature Request: Support for Apriel-5B-Instruct
enhancement
New feature or request
#12926
opened Apr 13, 2025 by
Abdulhanan535
4 tasks done
Misc. bug: gguf-my-repo doesn't work - [Errno 2] No such file or directory: './llama.cpp/llama-quantize'
bug-unconfirmed
#12925
opened Apr 13, 2025 by
rnckp
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.