Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Is Infini-attention support possible?
enhancement
New feature or request
#7213
opened May 11, 2024 by
sdmorrey
Native Intel IPEX-LLM Support
enhancement
New feature or request
#7190
opened May 10, 2024 by
iamhumanipromise
Build error at server.cpp: undefined reference to `json_schema_to_grammar
bug-unconfirmed
#7189
opened May 10, 2024 by
jarviszeng-zjc
third party applications are overwhelmingly slow for subsequent prompt evaluation compared to examples/main and examples/server
enhancement
New feature or request
#7185
opened May 9, 2024 by
khimaros
4 tasks done
BF16 prompt processing has half the performance compared to F16 and F32 von AMD Ryzen Embedded V3000 (Zen 3)
enhancement
New feature or request
#7182
opened May 9, 2024 by
lemmi
llamacpp --prompt-cache-all < -- more than a year passed and still is not fully implemented
enhancement
New feature or request
#7179
opened May 9, 2024 by
mirek190
selects too many cores by default on orange pi 5 (2x slower)
bug-unconfirmed
#7176
opened May 9, 2024 by
calculatortamer
Should we add an autolabeler for PR?
devops
improvements to build systems and github actions
enhancement
New feature or request
help wanted
Extra attention is needed
#7174
opened May 9, 2024 by
mofosyne
Add support for mistral Dutch and Armenian models: Tweeties/tweety-7b-dutch-v24a and Tweeties/tweety-7b-armenian-v24a
enhancement
New feature or request
#7170
opened May 9, 2024 by
JohnClaw
Support for Consistency Large Language Models?
enhancement
New feature or request
#7168
opened May 9, 2024 by
unoexperto
how can i modify the setting,make it answer in Chinese by default
enhancement
New feature or request
#7167
opened May 9, 2024 by
LiangZeFenglzf
Add metadata override and also generate dynamic default filename when converting gguf
enhancement
New feature or request
help wanted
Extra attention is needed
need feedback
Testing and feedback with results are needed
#7165
opened May 9, 2024 by
mofosyne
Looking for help for using llama.cpp with Phi3 model and LoRA
bug-unconfirmed
#7164
opened May 9, 2024 by
SHIMURA0
Gibberish response from server and main exits on M1 macstudio ultra with gpu (cpu ok)
bug-unconfirmed
#7159
opened May 9, 2024 by
jrozentur
Impact of bf16 on Llama 3 8B perplexity?
enhancement
New feature or request
#7148
opened May 8, 2024 by
jim-plus
4 tasks done
error: implicit declaration of function ‘vld1q_s8_x4’; did you mean ‘vld1q_s8_x2’?
bug-unconfirmed
#7147
opened May 8, 2024 by
CaptainOfHacks
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.