-
Notifications
You must be signed in to change notification settings - Fork 43
Pull requests: triton-inference-server/vllm_backend
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
feat: Add
GPU_DEVICE_IDS parameter support for pinning multi-GPU models to specific GPUs
#125
opened Apr 13, 2026 by
pskiran1
Member
Loading…
feat: Report more vllm metrics
enhancement
New feature or request
#92
opened May 13, 2025 by
Pavloveuge
Loading…
3 of 10 tasks
Add support for priority in vllm backend
#88
opened Apr 24, 2025 by
TheCodeWrangler
Loading…
2 of 5 tasks
Build: Trigger CI for new vllm_backend Triton releases
#49
opened Jul 23, 2024 by
nvda-mesharma
Collaborator
•
Draft
ProTip!
Exclude everything labeled
bug with -label:bug.