[Bug] mlc_llm chat
throws errors for model mlc-ai/Qwen1.5-1.8B-Chat-q4f16_1-MLC
#2254
Labels
bug
Confirmed bugs
mlc_llm chat
throws errors for model mlc-ai/Qwen1.5-1.8B-Chat-q4f16_1-MLC
#2254
馃悰 Bug
Hello,
HF://mlc-ai/Qwen1.5-1.8B-Chat-q4f16_1-MLC
seems to be incomplete:missing
max_batch_size
inmlc-chat-config.json
;no tokenizers found under
Qwen1.5-1.8B-Chat-q4f16_1-MLC/
And above two missings will cause
mlc_llm chat ...
throw errors.To Reproduce
Steps to reproduce the behavior:
Expected behavior
Environment
conda
, source): yespip
, source):python -c "import tvm; print('\n'.join(f'{k}: {v}' for k, v in tvm.support.libinfo().items()))"
, applicable if you compile models):Additional context
The text was updated successfully, but these errors were encountered: