Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Question] Can not get chat CLI working, throwing error after cloning model
question
Question about the usage
#2339
opened May 14, 2024 by
BeytoA
[Question] Deployment of Pruned Models
question
Question about the usage
#2338
opened May 14, 2024 by
qianjyM
Could not find org.apache.tvm:tvm-android:0.1.0.
question
Question about the usage
#2333
opened May 13, 2024 by
viaowp
[Question] Parallel computations using multiple streams?
question
Question about the usage
#2332
opened May 13, 2024 by
taegeonum
[Bug] InternalError: Check failed: (res == VK_SUCCESS) is false: Vulkan Error, code=-4: VK_ERROR_DEVICE_LOST
bug
Confirmed bugs
#2328
opened May 11, 2024 by
aaaaaad333
[Tracking] Create a CPU Compatible PagedKVCache
status: tracking
Tracking work in progress
#2325
opened May 11, 2024 by
tqchen
1 task
[Tracking] Sentence Embedding Model
status: tracking
Tracking work in progress
#2324
opened May 11, 2024 by
tqchen
1 task
[Bug] mlc_llm package failed once, and i cant run it again
bug
Confirmed bugs
#2323
opened May 11, 2024 by
CallMeTkt
[Feature Request] Medusa support
feature request
New feature or request
#2319
opened May 10, 2024 by
EmilioZhao
[Bug] Support multiple "system" messages in REST API
bug
Confirmed bugs
#2311
opened May 10, 2024 by
bayley
[Bug] mlc-llm not working, tvm check returns none
bug
Confirmed bugs
#2301
opened May 9, 2024 by
CallMeTkt
[Bug] REST server doesn't work on V100 (SM70) - cudaErrorNoKernelImageForDevice (but chat works)
bug
Confirmed bugs
#2296
opened May 8, 2024 by
bayley
[Model Request] can we get Aryanne/Calypso-3B-alpha-v2-gguf
new-models
#2293
opened May 7, 2024 by
Louis654
Prebuilt StableLM 1.6B model compilation not working
bug
Confirmed bugs
#2283
opened May 6, 2024 by
saurav-pwh-old
[Bug] llama2 7b android compilation is giving "Can only handle constant size stack allocation for now" error
bug
Confirmed bugs
#2282
opened May 6, 2024 by
Ramees025
[Bug] Check failed: (args.size() == initial_indices_orig.size()) is false
bug
Confirmed bugs
#2276
opened May 5, 2024 by
jpf888
Phi-3 mini 4k instruct with MICROSOFT's quantization
help wanted
Looking for community help
new-models
#2273
opened May 4, 2024 by
federicoparra
[Question] What models actually work with function calling?
question
Question about the usage
#2267
opened May 2, 2024 by
dtkettler
[Bug] Confirmed bugs
system-lib-prefix
would be cleared if device
is not strictly android
while mlc_llm compile
bug
#2255
opened Apr 30, 2024 by
BitCircuit
[Bug] Confirmed bugs
mlc_llm chat
throws errors for model mlc-ai/Qwen1.5-1.8B-Chat-q4f16_1-MLC
bug
#2254
opened Apr 30, 2024 by
BodhiHu
[Question] Omniquant. (AFAIK) scores best for Q. Methods, why no adoption? In any case, is per-tensor quant. best for Mixtral/MoE models?
question
Question about the usage
#2247
opened Apr 29, 2024 by
BuildBackBuehler
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-04-14.