Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Benchmark Performance to measure response times for Inference #450

Open
ratnopamc opened this issue Feb 24, 2024 · 0 comments
Open

Benchmark Performance to measure response times for Inference #450

ratnopamc opened this issue Feb 24, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@ratnopamc
Copy link
Collaborator

Community Note

  • Please vote on this issue by adding a 👍 reaction to the original issue to help the community and maintainers prioritize this request
  • Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
  • If you are interested in working on this issue or have submitted a pull request, please leave a comment

What is the outcome that you are trying to reach?

Use a tool to load test and benchmark Inference latency, throughput, response times to scale Pod and create new nodes at load.

Describe the solution you would like

Use a bench marking tool like fmbt for this purpose.

Describe alternatives you have considered

Additional context

@vara-bonthu vara-bonthu added the enhancement New feature or request label Feb 26, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants