-
Notifications
You must be signed in to change notification settings - Fork 10.9k
Issues: ggml-org/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Misc. bug: llama-cli llama_backend_free may not free all the gpu memory
bug-unconfirmed
#12057
opened Feb 25, 2025 by
GaoXiangYa
Misc. bug: cannot scroll to right side when input too long
bug-unconfirmed
#12054
opened Feb 24, 2025 by
gnusupport
Compile bug: llama.cpp-b4749/ggml/src/ggml-cpu/ggml-cpu-quants.c:5141:26: error: initialization of ‘uint32_t *’ {aka ‘unsigned int *’} from incompatible pointer type ‘const uint8_t (*)[12]’ {aka ‘const unsigned char (*)[12]’} [-Wincompatible-pointer-types]
bug-unconfirmed
#12050
opened Feb 24, 2025 by
Arniiiii
Eval bug: TikTokenTokenizer has no attribute vocab
bug-unconfirmed
#12044
opened Feb 24, 2025 by
zhanghui-china
Eval bug: Error when converting moonlight from bf16 to q4km
bug-unconfirmed
#12040
opened Feb 23, 2025 by
qiyuxinlin
Misc. bug: --no-context-shift OR --context-shift ?
bug-unconfirmed
#12038
opened Feb 23, 2025 by
simonchen
Misc. bug: Web-UI now unusably slow - over network or locally.
bug-unconfirmed
#12026
opened Feb 22, 2025 by
clort81
Eval bug: unknown pre-tokenizer type: 'deepseek-r1-qwen'
bug-unconfirmed
#12021
opened Feb 22, 2025 by
wr131
Misc. bug: Concurrency Limitation: Only 6 Inferences Run Simultaneously When Setting
--parallel
> 6
bug-unconfirmed
#12013
opened Feb 21, 2025 by
karanotsingyu
[CANN] Compile bug: no matching function for call to 'CastIntrinsicsImpl' Ascend NPU issues specific to Ascend NPUs
#12010
opened Feb 21, 2025 by
Cikaros
Eval bug: does llama.cpp support Intel AMX instruction? how to enable it
bug-unconfirmed
#12003
opened Feb 21, 2025 by
montagetao
Misc. bug: add tool_calls id in response in server
bug-unconfirmed
#11992
opened Feb 21, 2025 by
henryclw
Feature Request: add Kernel level verbose option
enhancement
New feature or request
#11985
opened Feb 20, 2025 by
0400H
4 tasks done
Misc. bug: llama-cli '--log-disable' parameter omits response
bug-unconfirmed
#11983
opened Feb 20, 2025 by
nmandic78
Eval bug: CANNOT LINK EXECUTABLE "./llama-cli": library "libomp.so" not found: needed by main executable
bug-unconfirmed
#11979
opened Feb 20, 2025 by
Krallbe68
GGML to GGUF FAIL Quantized tensor bytes per row (5120) is not a multiple of Q2_K type size (84)
#11976
opened Feb 20, 2025 by
chokoon123
tensor 'blk.25.ffn_down.weight' has invalid ggml type 42 (NONE)
bug-unconfirmed
#11975
opened Feb 20, 2025 by
evaninf
Misc. bug: Sporadic MUL_MAT Failures in test-backend-ops for Nvidia backend
bug-unconfirmed
#11972
opened Feb 20, 2025 by
ShanoToni
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.