Skip to content

Issues: ggml-org/llama.cpp

examples : add configuration presets
#10932 opened Dec 21, 2024 by ggerganov
Open 3
changelog : libllama API
#9289 opened Sep 3, 2024 by ggerganov
Open 9
changelog : llama-server REST API
#9291 opened Sep 3, 2024 by ggerganov
Open 14
Beta
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Feature Request: Add kv-quant fa kernel variants for head sizes other than 128 enhancement New feature or request
#12989 opened Apr 17, 2025 by pl752
4 tasks done
Feature Request: Improve model load time when using the RPC backend enhancement New feature or request
#12954 opened Apr 15, 2025 by rgerganov
4 tasks done
Model conversion issue
#12941 opened Apr 14, 2025 by Eucliwoodprpr
Feature Request: Support for Apriel-5B-Instruct enhancement New feature or request
#12926 opened Apr 13, 2025 by Abdulhanan535
4 tasks done
ProTip! Find all open issues with in progress development work with linked:pr.