-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] Cannot run finetuned model of Mistral 7B with Confirmed bugs
mlc_llm convert_weights
with "data did not match any variant of untagged enum ModelWrapper"
bug
#3023
opened Nov 12, 2024 by
pjyi2147
Speculative mode for the LLaMA 3.1 70B model
question
Question about the usage
#3022
opened Nov 11, 2024 by
shahizat
Support of heterogeneous devices
feature request
New feature or request
#3018
opened Nov 11, 2024 by
musram
[Bug] flutter 跟安卓原生交互,调用engine.chatCompletion 就会发生anr
bug
Confirmed bugs
#3017
opened Nov 10, 2024 by
tdd102
[Question] How to show model progress download on WebLLM Javascript SDK?
question
Question about the usage
#3014
opened Nov 8, 2024 by
DenisSergeevitch
[Question] How to export MLCChat .apk with weight bundled/included?
question
Question about the usage
#3013
opened Nov 6, 2024 by
lifelongeeek
[Bug] Speculative decoding doesn't work on Vulkan (AMD iGPU)
bug
Confirmed bugs
#3011
opened Nov 4, 2024 by
SkyHeroesS
[Question] Android app issue
question
Question about the usage
#3010
opened Nov 4, 2024 by
j0h0k0i0m
[Bug] Llama-3.1-70B-Instruct-q3f16_1-MLC model running across two GPUs with tensor_parallel_shards=2
bug
Confirmed bugs
#3004
opened Oct 31, 2024 by
shahizat
[Bug] Misalignment of Llama3.2 chat template
bug
Confirmed bugs
#3002
opened Oct 31, 2024 by
Hzfengsy
[Question] Error running prep_emcc_deps.sh - 'tvm/runtime/object.h' file not found
question
Question about the usage
#3001
opened Oct 30, 2024 by
Big-Boy-420
[Question] Which models do you recommend for compiling on Mac Intel chip, metal gpu?
question
Question about the usage
#2999
opened Oct 28, 2024 by
RINO-GAELICO
[Bug] Llama 3.2 3B and 1B on MLC are significantly slower than Llama 3.1 8B (L40s, fp16)
bug
Confirmed bugs
#2997
opened Oct 26, 2024 by
chrisreese-if
[Question] TVM error on Mac Intel chip, Metal accelerator
question
Question about the usage
#2995
opened Oct 24, 2024 by
RINO-GAELICO
[Bug] Concurrent requests are being run sequentially on AMD MI60 (gfx906)
bug
Confirmed bugs
#2992
opened Oct 22, 2024 by
Said-Akbar
Problems encountered while doing Android SDK
question
Question about the usage
#2986
opened Oct 18, 2024 by
whyshifht
[Bug] some questions about build Android app with mlc
bug
Confirmed bugs
#2985
opened Oct 17, 2024 by
ecccccsgo
[Bug] Running Quick Start Example in Windows gives Error: Confirmed bugs
'MLCEngine' object has no attribute '_ffi'
bug
#2979
opened Oct 15, 2024 by
jackuh105
[Bug] Llama example inference using Vulkan gives build error
bug
Confirmed bugs
#2977
opened Oct 14, 2024 by
asfarkTii
[Question] Fail to pip install mlc-llm
question
Question about the usage
#2974
opened Oct 13, 2024 by
xiboliyaxiangjiaojun
Previous Next
ProTip!
Follow long discussions with comments:>50.