-
Notifications
You must be signed in to change notification settings - Fork 9.7k
Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Feature Request: Adding support for Ternary DiT models
enhancement
New feature or request
#10334
opened Nov 16, 2024 by
Lucky-Lance
4 tasks done
Bug: "GPU + CUDA + VRAM + Shared Memory (UMA)" slower then "CPU + RAM"?
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10330
opened Nov 16, 2024 by
kripper
Bug: CI failing because of windows-latest-cmake-sycl
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10327
opened Nov 16, 2024 by
FirstTimeEZ
Bug: All SYCL builds since b3987 or so are unstable
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10323
opened Nov 15, 2024 by
0xDEADFED5
Bug: Using llama_batch_init+add+free instead of llama_batch_get_one() permanently slows down llama_decode significantly
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10322
opened Nov 15, 2024 by
Nekotekina
Bug: llama-gbnf-validator parses grammar but gets a seg fault when validating an input string against the grammar
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10321
opened Nov 15, 2024 by
nissenbenyitskhak
Feature Request: Add OLMo November 2024
enhancement
New feature or request
#10316
opened Nov 15, 2024 by
2015aroras
4 tasks done
Bug: I am unable to use llama_cli interactively
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10297
opened Nov 14, 2024 by
phalexo
Feature Request: shared tokens in batches with New feature or request
logits = true
enhancement
#10295
opened Nov 14, 2024 by
Lyrcaxis
4 tasks done
Bug: Build failure with GGML_VULKAN=1 GGML_HIPBLAS=1
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10284
opened Nov 14, 2024 by
nullref
Feature Request: A method to load all model layers into VRAM, then with the remaining VRAM load context active context, and overlow into system ram
enhancement
New feature or request
#10283
opened Nov 13, 2024 by
tigert2173
4 tasks done
Bug: I use qwen2_7b_instruc Python llama. cp/convert_cf_to_gguf. py error
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#10273
opened Nov 13, 2024 by
cxs86621
Feature Request: adderALL
enhancement
New feature or request
#10265
opened Nov 12, 2024 by
jam54186
4 tasks done
Feature Request: Tencent-Hunyuan-Large (Text Generation)
enhancement
New feature or request
#10263
opened Nov 12, 2024 by
jhofseth
4 tasks done
Bug: docker sample usage will always trigger unhealty container status
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10262
opened Nov 12, 2024 by
lilunxm12
Bug: CANN: Inference result garbled
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10252
opened Nov 11, 2024 by
feichenchina
Bug: In interactive chat mode (LLaMa 3.1 70B) sometimes llama.cpp fills in the user's side of the conversation.
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10249
opened Nov 10, 2024 by
vfhbg
web UI : support syntax highlighting
enhancement
New feature or request
good first issue
Good for newcomers
help wanted
Extra attention is needed
#10246
opened Nov 10, 2024 by
slaren
Feature Request: Flash Attention 3
enhancement
New feature or request
#10245
opened Nov 10, 2024 by
hg0428
4 tasks done
Bug: server GET /props request return json with chat_template with last char replaced by \x00
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10235
opened Nov 9, 2024 by
kks-imt
Bug: Server Slows Down Significantly Over Time, Requires Frequent Reboots (RX 7900 XT)
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10227
opened Nov 9, 2024 by
tigert2173
bge-multilingual-gemma2:ERROR:hf-to-gguf:Model Gemma2Model is not supported
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#10215
opened Nov 8, 2024 by
hellozjj
Bug: not support langchain v0.3 to use tools
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#10214
opened Nov 8, 2024 by
lee249876293
Feature Request: Support Airllm
enhancement
New feature or request
#10202
opened Nov 7, 2024 by
kbocock-krg
4 tasks done
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.