-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Issues: mlc-ai/mlc-llm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] WARNING: Skipping mlc-ai-nightly as it is not installed
bug
Confirmed bugs
#3063
opened Dec 13, 2024 by
vfdff
[Feature Request] Streamed [DONE] response in RestAPI should have token data
feature request
New feature or request
#3061
opened Dec 10, 2024 by
TNT3530
[Bug] Infinite loop after generate token length near context_windows_size/chunk_prefill_size.
bug
Confirmed bugs
#3057
opened Dec 6, 2024 by
gesanqiu
[Bug] Binary was created using {relax.Executable} but a loader of that name is not registered.
bug
Confirmed bugs
#3055
opened Dec 3, 2024 by
LLIo6oH
[Bug] gemma-2-27b-it-q4f16_1-MLC output the incorrect content.
bug
Confirmed bugs
#3054
opened Dec 1, 2024 by
rankaiyx
[Bug] Still Experiencing 'Error: Using LLVM 19.1.3 with Confirmed bugs
-mcpu=apple-latest
is not valid in -mtriple=arm64-apple-macos
, using default -mcpu=generic
'
bug
#3053
opened Dec 1, 2024 by
BuildBackBuehler
[Question] How to get runtime stats in serve mode?
question
Question about the usage
#3052
opened Dec 1, 2024 by
rankaiyx
[Feature Request] Embeddings support for iOS
feature request
New feature or request
#3050
opened Nov 29, 2024 by
jondeandres
[Bug] Android Llama-3.2-3B-Instruct-q4f16_0-MLC init failed
bug
Confirmed bugs
#3048
opened Nov 26, 2024 by
tdd102
[Bug] issue while compiling FP8 dataype of mlc-ai/Llama-3.1-8B-Instruct-fp8-MLC
bug
Confirmed bugs
#3047
opened Nov 26, 2024 by
Vinaysukhesh98
[Bug][iOS/Swift SDK] Multiple image input to vision models will throw error from TVM
bug
Confirmed bugs
#3044
opened Nov 22, 2024 by
Neet-Nestor
[Question] Does MLC_LLM MLCEngine have an equivalent API for Question about the usage
llm.generate
in VLLM or SGLang?
question
#3034
opened Nov 17, 2024 by
pjyi2147
KV cache offloading to CPU RAM
feature request
New feature or request
#3033
opened Nov 17, 2024 by
shahizat
[Feature Request] Add vision model flag to model record
feature request
New feature or request
#3031
opened Nov 16, 2024 by
Neet-Nestor
[Bug] flutter 跟安卓原生交互,调用engine.chatCompletion 就会发生anr
bug
Confirmed bugs
#3017
opened Nov 10, 2024 by
tdd102
[Bug] Speculative decoding doesn't work on Vulkan (AMD iGPU)
bug
Confirmed bugs
#3011
opened Nov 4, 2024 by
SkyHeroesS
[Question] Android app issue
question
Question about the usage
#3010
opened Nov 4, 2024 by
j0h0k0i0m
[Bug] Llama-3.1-70B-Instruct-q3f16_1-MLC model running across two GPUs with tensor_parallel_shards=2
bug
Confirmed bugs
#3004
opened Oct 31, 2024 by
shahizat
[Bug] Misalignment of Llama3.2 chat template
bug
Confirmed bugs
#3002
opened Oct 31, 2024 by
Hzfengsy
[Question] Error running prep_emcc_deps.sh - 'tvm/runtime/object.h' file not found
question
Question about the usage
#3001
opened Oct 30, 2024 by
Big-Boy-420
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.