Skip to content

Show input and output length on vLLM dashboard #6992

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 2 commits into from
Aug 12, 2025

Conversation

huydhn
Copy link
Contributor

@huydhn huydhn commented Aug 9, 2025

Fixes #6974

Input and output lengths are new dimensions on the dashboard that needs to be displayed after pytorch/pytorch-integration-testing#42. This PR also cleans up some old TODO code path for vLLM dashboard.

Testing

Different input and output lengths are showing up correctly now with their benchmark results on the preview

@huydhn huydhn requested review from BoyuanFeng and yangw-dev August 9, 2025 01:33
Copy link

vercel bot commented Aug 9, 2025

@huydhn is attempting to deploy a commit to the Meta Open Source Team on Vercel.

A member of the Team first needs to authorize it.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Aug 9, 2025
Copy link

vercel bot commented Aug 9, 2025

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Project Deployment Preview Updated (UTC)
torchci Ready Preview Aug 12, 2025 7:42pm

@BoyuanFeng
Copy link

Thanks for the pr!

Is it possible to add in/out to the name? E.g., "meta-llama/Llama-4-Scout-17B-16E-Instruct / tp4 / qps_10 / 200 / 200" -> "meta-llama/Llama-4-Scout-17B-16E-Instruct / tp4 / qps_10 / input-200 / output-200":
image

Also, curious what is the setting here?
image

@huydhn
Copy link
Contributor Author

huydhn commented Aug 12, 2025

Thanks for the pr!

Is it possible to add in/out to the name? E.g., "meta-llama/Llama-4-Scout-17B-16E-Instruct / tp4 / qps_10 / 200 / 200" -> "meta-llama/Llama-4-Scout-17B-16E-Instruct / tp4 / qps_10 / input-200 / output-200": image

Oh, they're already part of the name, you can see them there in the format / 200 / 200. But let me make them clearer with their prefix, i.e. / input_200 / output_200, otherwise, it's unclear which is which

@huydhn
Copy link
Contributor Author

huydhn commented Aug 12, 2025

Also, curious what is the setting here?

For latency and thoughput benchmarks, input and output length aren't set directly, but from what I see input_len is 32 and output_len is 128, probably the default values somewhere, I could show them in the chart

@huydhn
Copy link
Contributor Author

huydhn commented Aug 12, 2025

One last note here is that input and output lengths are set in latency and serving benchmarks, but not throughput. I try to look into the default that it uses, but there is none for the ShareGPT dataset https://github.com/vllm-project/vllm/blob/main/vllm/benchmarks/datasets.py#L402. Few other datasets like Random or Sonnet define input and output lengths, but not ShareGPT. I will keep these fields empty for throughput benchmark for now.

@huydhn huydhn merged commit f9f933a into pytorch:main Aug 12, 2025
5 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Track input and output length on vLLM dashboard
3 participants