Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Issues: ggml-org/llama.cpp

examples : add configuration presets
#10932 opened Dec 21, 2024 by ggerganov
Open 5
changelog : libllama API
#9289 opened Sep 3, 2024 by ggerganov
Open 9
changelog : llama-server REST API
#9291 opened Sep 3, 2024 by ggerganov
Open 15
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

mtmd : (WIP) add ultravox audio input examples python python script changes
#13623 opened May 18, 2025 by ngxson Draft
server : separate the notion of position and KV tokens, remove prompt truncation breaking change Changes that break ABIs, APIs, file formats, or other forms of backwards compatibility. examples python python script changes server
#13576 opened May 15, 2025 by ngxson Loading…
Update python verions examples python python script changes server
#13574 opened May 15, 2025 by robbiemu Loading…
sycl : reviewing the backend documentation documentation Improvements or additions to documentation examples SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language
#13544 opened May 14, 2025 by Alcpz Loading…
Fix build on OpenBSD examples
#13541 opened May 14, 2025 by percypiper Loading…
[SYCL] Overcoming workaround for mmap() allocation on Windows and remove useless wait examples ggml changes relating to the ggml tensor library for machine learning SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language
#13482 opened May 12, 2025 by s-Nick Loading…
Webui dynamic config examples server
#13429 opened May 10, 2025 by ServeurpersoCom Loading…
Add mistral-chat-7b preset for llama-server examples
#13348 opened May 7, 2025 by vahedshaik Loading…
mtmd : add vision support for llama 4 documentation Improvements or additions to documentation examples help wanted Extra attention is needed python python script changes
#13282 opened May 3, 2025 by ngxson Loading…
kv-cache : add SWA support examples server
#13194 opened Apr 29, 2025 by ggerganov Loading…
15 of 22 tasks
llama : try loading tensors with pre-computed hashes Apple Metal https://en.wikipedia.org/wiki/Metal_(API) examples ggml changes relating to the ggml tensor library for machine learning Kompute https://github.com/KomputeProject/kompute/ Nvidia GPU Issues specific to Nvidia GPUs SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language Vulkan Issues specific to the Vulkan backend
#13106 opened Apr 25, 2025 by rgerganov Loading…
[sync #10544] llama/ggml: add LLM training support examples ggml changes relating to the ggml tensor library for machine learning testing Everything test related
#13105 opened Apr 25, 2025 by ggerganov Draft
1 task
threading: support for GGML_SCHED_PRIO_LOW, update thread info on Windows to avoid throttling examples ggml changes relating to the ggml tensor library for machine learning
#12995 opened Apr 17, 2025 by max-krasnyansky Loading…
llama-tts : input from stdin examples
#12890 opened Apr 11, 2025 by marcoStocchi Loading…
Support for OuteTTS 1.0 examples python python script changes
#12794 opened Apr 7, 2025 by edwko Draft
ProTip! Exclude everything labeled bug with -label:bug.
Morty Proxy This is a proxified and sanitized view of the page, visit original site.