-
Notifications
You must be signed in to change notification settings - Fork 44
Issues: LLMServe/DistServe
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
How to independently measure the performance of the Prefill phase and the Decode phase?
#50
opened Nov 19, 2024 by
J1nLo
offline/online serving is stuck at fetching files in paraworkers
#48
opened Oct 16, 2024 by
hyuenmin-choi
How to use tensorrt-llm as the inference backend for DistServe
#45
opened Sep 18, 2024 by
GGBond8488
fail to run examples/offline.py , unable to download the model to reproduce
#35
opened Aug 6, 2024 by
William12github
Generating max_num_tokens.csv for Different Hardware Environments
#30
opened Jul 28, 2024 by
village-way
codellama34b ttft延迟问题
question
Further information is requested
#19
opened Jul 2, 2024 by
sitabulaixizawaluduo
Decode Wrong Token
help wanted
Extra attention is needed
#16
opened Jun 18, 2024 by
sitabulaixizawaluduo
Offline.py LLMEngine.__init__() missing 1 required positional argument: 'simulator_config'
help wanted
Extra attention is needed
#15
opened Jun 14, 2024 by
fivebamboo694
How difficult will adding Llama 3 support be?
enhancement
New feature or request
#12
opened Jun 13, 2024 by
kalradivyanshu
decoder.embed_tokens.weight.pt not found
help wanted
Extra attention is needed
#10
opened Jun 11, 2024 by
llx-08
ProTip!
Exclude everything labeled
bug
with -label:bug.