Skip to content

Issues: kubernetes-sigs/inference-perf

Beta
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Publish inference-perf image on GHCR
#52 opened Apr 21, 2025 by surajssd
Ephemeral Port Exhaustion
#50 opened Apr 10, 2025 by Bslabe123
Logger
#48 opened Mar 31, 2025 by Bslabe123
Consolidate perf testing tools
#23 opened Feb 19, 2025 by kfswain
Add Kubernetes Orchestration Library for Model Server Deployment and Benchmarking kind/feature Categorizes issue or PR as related to a new feature. priority/important-soon Must be staffed and worked on either currently, or very soon, ideally in time for the next release.
#22 opened Feb 13, 2025 by wangchen615
[Feature] Add a model server client for Triton using TensorRT-LLM lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.
#18 opened Feb 3, 2025 by achandrasekar
[Feature] Add a client to get model server metrics lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.
#17 opened Feb 3, 2025 by achandrasekar
Proposal: Inference-perf loadgen component to be based on Grafana k6 load testing tool lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.
#2 opened Jan 20, 2025 by SachinVarghese
ProTip! no:milestone will show everything without a milestone.