-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug]: Cannot use FlashAttention-2 backend for head size 72.
bug
Something isn't working
#12656
opened Feb 2, 2025 by
jasstionzyf
1 task done
[Usage]: A problem when use llm.generate() for several times in one LLM case
usage
How to use vllm
#12651
opened Feb 1, 2025 by
KleinMoretti07
1 task done
[Installation]: how to create envs.py file for build on CPU machine?
installation
Installation problems
#12649
opened Feb 1, 2025 by
d95776
1 task done
[Bug]: Assertion Error When Using moe_wna16
bug
Something isn't working
#12647
opened Feb 1, 2025 by
LagPixelLOL
1 task done
[Bug]: When the dataset is not a JSON file in benchmark_prioritization.py, the number of Tuple elements in the constructed requests is wrong
bug
Something isn't working
#12645
opened Feb 1, 2025 by
Accelerator1996
1 task done
[Bug]: shape is invalid for input of size
bug
Something isn't working
#12633
opened Jan 31, 2025 by
kagurazakasanae
1 task done
[Bug]: stack trace for "Watchdog caught collective operation timeout"
bug
Something isn't working
#12625
opened Jan 31, 2025 by
remusao
1 task done
[Feature]: Only apply Guided/Structured grammar after reasoning steps in Reasoning models
feature request
structured-output
#12619
opened Jan 31, 2025 by
cksac
1 task done
[Performance]: Weird Sliding Window Attention Profiling Results
performance
Performance-related issues
#12616
opened Jan 31, 2025 by
tilmto
1 task done
[Bug]: VLLM (0.7.0) will report gpu missing on the hosting node in Ray
bug
Something isn't working
#12614
opened Jan 31, 2025 by
pcpLiu
1 task done
[Bug]: GPU memory usage gradually increases.
bug
Something isn't working
#12610
opened Jan 31, 2025 by
loxs123
1 task done
[Doc]: apparently missing git clone in "Build wheel from source" non-GPU installation documentation
documentation
Improvements or additions to documentation
#12590
opened Jan 30, 2025 by
kikoreis
1 task done
[Bug]: [V1] New v1 engine does not support n>1?
bug
Something isn't working
#12584
opened Jan 30, 2025 by
m-harmonic
1 task done
[Bug][TPU]: Non-deterministic behaviour
bug
Something isn't working
tpu
Related to Google TPUs
#12580
opened Jan 30, 2025 by
Akshat-Tripathi
1 task done
[Usage]: Does DeepSeek-R1 1.58-bit Dynamic Quant work on VLLM?
usage
How to use vllm
#12573
opened Jan 30, 2025 by
shimmyshimmer
1 task done
[Bug]: Ray/vLLM RuntimeError: HIP error: invalid device ordinal
bug
Something isn't working
#12572
opened Jan 30, 2025 by
etiennemlb
1 task done
[Bug]: vllm container does not set LD_LIBRARY_PATH correctly
bug
Something isn't working
#12559
opened Jan 29, 2025 by
csiefer2
1 task done
[Feature]: static distribution of vllm
feature request
#12556
opened Jan 29, 2025 by
paigeadelethompson
1 task done
[Bug]: Engine fails to start when running Qwen2.5 Deepseek r1
bug
Something isn't working
#12554
opened Jan 29, 2025 by
JamesDConley
1 task done
[Bug]: Distilled DeepSeek Models do not work with guided_json
bug
Something isn't working
structured-output
#12548
opened Jan 29, 2025 by
FilippoBoni1921
1 task done
[Bug]: DeepseekR1 model load fails with weights tied error
bug
Something isn't working
#12541
opened Jan 29, 2025 by
gdagur
1 task done
[Performance]: V1 higher memory usage
performance
Performance-related issues
v1
#12529
opened Jan 28, 2025 by
wedobetter
1 task done
[Bug]: Error After Model Load in vllm 0.7.0 (No Issue in vllm 0.6.6)
bug
Something isn't working
#12515
opened Jan 28, 2025 by
denddyprod
1 task done
[Feature]: Support torch.distributed as the runtime for multi-node inference
feature request
#12511
opened Jan 28, 2025 by
gaocegege
1 task done
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.