Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Provide comprehensive inference benchmarking for LLM backends (vLLM, TensorRT-LLM, TGI) #3868

Open
yuzisun opened this issue Aug 18, 2024 · 0 comments

Comments

@yuzisun
Copy link
Member

yuzisun commented Aug 18, 2024

/kind feature

Describe the solution you'd like
Provide comprehensive benchmarking results for inference runtimes (vLLM, TensorRT-LLM and TGI) in a KServe setup.

  • Latency (TOFT)
  • Throughput (Token generation rate)

Factors:

  • Quantization (AWQ/GPTQ)
  • Number of concurrent virtual users
  • Feature enablement for prefix caching etc

GPUs:

  • A100 80GB GPU
  • H100 80GB GPU

Models:

  • llama3.1 8B
  • llama3.1 70B 4bit quantization
  • llama3.1 405B FP8

Dataset for testing

Anything else you would like to add:
[Miscellaneous information that will assist in solving the issue.]

Links to the design documents:
[Optional, start with the short-form RFC template to outline your ideas and get early feedback.]
[Required, use the longer-form design doc template to specify and discuss your design in more detail]

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant