-
-
Notifications
You must be signed in to change notification settings - Fork 6.7k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug]: Calling /wake_up after /sleep and then sending a request leads to improper LLM response
bug
Something isn't working
#16234
opened Apr 8, 2025 by
akshayqylis
1 task done
[Usage]: Multiple Models on Same Port
usage
How to use vllm
#16232
opened Apr 8, 2025 by
dipta007
1 task done
[Feature]: Support Pipeline Parallelism on Llama-4-Maverick-17B-128E
feature request
New feature or request
#16231
opened Apr 8, 2025 by
Edwinhr716
1 task done
[Bug]: failed to load deepseek-r1 AWQ quantization on CPU
bug
Something isn't working
#16230
opened Apr 8, 2025 by
spaceater
1 task done
[Bug]: Cannot register new model
bug
Something isn't working
#16228
opened Apr 8, 2025 by
amogkam
1 task done
Llama4 Pretrained Model support
feature request
New feature or request
#16227
opened Apr 8, 2025 by
jerrickzchen
1 task done
[Usage]: Model loading failure with Llama-4-Maverick-17B when using torch.compile
usage
How to use vllm
#16223
opened Apr 8, 2025 by
ehartford
1 task done
[Bug]: benchmark_throughput.py not working with data-parallelism
bug
Something isn't working
#16222
opened Apr 8, 2025 by
kartikx
1 task done
[RFC]: Changes to support attention + quant fusion
RFC
#16220
opened Apr 7, 2025 by
ProExpertProg
1 task done
[Installation]: Missing dev wheel for vllm 0.8.3 and 0.8.2
installation
Installation problems
#16217
opened Apr 7, 2025 by
magdyksaleh
1 task done
[Usage]: Llama4 tool parser
usage
How to use vllm
#16214
opened Apr 7, 2025 by
dhruvmullick
1 task done
[Bug]: Issue loading AWQ model from GCS on Vertex AI using Custom Container. How to use a custom model with LLM from GCS.
bug
Something isn't working
#16208
opened Apr 7, 2025 by
nandhiniramanan5
1 task done
[Usage]: Llama-3.1-8B-Instruct
usage
How to use vllm
#16207
opened Apr 7, 2025 by
sbuliyax
1 task done
[Installation]: the --mount option requires BuildKit
installation
Installation problems
#16205
opened Apr 7, 2025 by
mohdrashid1
1 task done
[Bug]: github-action bot have a error
bug
Something isn't working
#16204
opened Apr 7, 2025 by
lengrongfu
1 task done
[Bug]: OOM when serve Gemma3-AWQ
bug
Something isn't working
#16199
opened Apr 7, 2025 by
hahmad2008
1 task done
[Bug]: Not able to deploy Llama-4-Scout-17B-16E-Instruct on vllm-openai v0.8.3
bug
Something isn't working
#16197
opened Apr 7, 2025 by
rabaja
1 task done
[Bug]: Models converted to GGUF don't seem to be able to do tool calling
bug
Something isn't working
#16195
opened Apr 7, 2025 by
RH-steve-grubb
1 task done
[Usage]: How can I quickly obtain the number of prompt tokens containing multimodal data?
help wanted
Extra attention is needed
multi-modality
Related to multi-modality (#4194)
usage
How to use vllm
#16191
opened Apr 7, 2025 by
yansh97
1 task done
[Bug]: Mistral tool parser failed to parse function calling
bug
Something isn't working
#16190
opened Apr 7, 2025 by
hahmad2008
1 task done
[Bug]: Huge memory overhead with V1 (multiprocessing) when handling several multimodal inputs
bug
Something isn't working
#16185
opened Apr 7, 2025 by
p88h
1 task done
[Bug]: memory usage is greater than expected
bug
Something isn't working
#16184
opened Apr 7, 2025 by
wjdsg
1 task done
[Bug]: Deepseek reasoning and guided_json no longer works
bug
Something isn't working
#16182
opened Apr 7, 2025 by
supersteves
1 task done
[Doc]: v1 supported model list?
documentation
Improvements or additions to documentation
#16179
opened Apr 7, 2025 by
Aganlengzi
1 task done
Previous Next
ProTip!
Follow long discussions with comments:>50.