Skip to content

Conversation

tukwila
Copy link
Contributor

@tukwila tukwila commented Sep 9, 2025

Summary

From docs i noticed that benchmark showcase is needed

image

this is one example.

Details

  • [ ]

Test Plan

Related Issues

  • Resolves #

  • "I certify that all code in this PR is my own, except as noted below."

Use of AI

  • Includes AI-assisted code completion
  • Includes code generated by an AI application
  • Includes AI-generated tests (NOTE: AI written tests should have a docstring that includes ## WRITTEN BY AI ##)

@tukwila tukwila force-pushed the example1 branch 4 times, most recently from 3a81383 to 89b0893 Compare September 9, 2025 20:35
Signed-off-by: guangli.bao <[email protected]>
@sjmonson
Copy link
Collaborator

Rejecting this for a few reasons:

Performance benchmarking against third-party services strongly discouraged. Results can be biased by network variability, rate limits, and other external factors. Additionally it is often against ToS to benchmark a service. For example, looking through the d.run ToS I found:

4.1.6 [...] using automated tools, scripts, or other means to send large quantities of requests or consume network bandwidth that severely impacts the d.run or other users' normal use. We reserve the right to limit your service access due to violations of these fairness principles, including but not limited to suspending or terminating services.

Finally, we do not want to promote any AI service vendor in our official documentation. We're open to bug fixes and enablement for hosted AI services, but GuideLLM is intended for use with your own infrastructure; other use-cases will not be officially endorsed or supported.

@sjmonson sjmonson closed this Sep 11, 2025
@tukwila tukwila deleted the example1 branch September 12, 2025 02:42
@tukwila
Copy link
Contributor Author

tukwila commented Sep 12, 2025

Rejecting this for a few reasons:

Performance benchmarking against third-party services strongly discouraged. Results can be biased by network variability, rate limits, and other external factors. Additionally it is often against ToS to benchmark a service. For example, looking through the d.run ToS I found:

4.1.6 [...] using automated tools, scripts, or other means to send large quantities of requests or consume network bandwidth that severely impacts the d.run or other users' normal use. We reserve the right to limit your service access due to violations of these fairness principles, including but not limited to suspending or terminating services.

Finally, we do not want to promote any AI service vendor in our official documentation. We're open to bug fixes and enablement for hosted AI services, but GuideLLM is intended for use with your own infrastructure; other use-cases will not be officially endorsed or supported.

If i adopt vllm-simulator in one docker container [https://llm-d.ai/docs/architecture/Components/inference-sim] to start up one vllm backend server to describe benchmark process, is that ok?

@sjmonson
Copy link
Collaborator

Yes!! That sounds perfect.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants