@@ -66,20 +66,20 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
6666 "mount_location" : " /models" ,
6767 "volume_size_in_gbs" : 500 ,
6868 "include" : [
69- " new_example_sglang .yaml" ,
69+ " offline_sglang_example .yaml" ,
7070 " NousResearch/Meta-Llama-3.1-8B"
7171 ]
7272 }
7373 ],
7474 "output_object_storage" : [
7575 {
7676 "bucket_name" : " inference_output" ,
77- "mount_location" : " /mlcommons_output " ,
77+ "mount_location" : " /benchmarking_output " ,
7878 "volume_size_in_gbs" : 200
7979 }
8080 ],
8181 "recipe_container_command_args" : [
82- " /models/new_example_sglang .yaml"
82+ " /models/offline_sglang_example .yaml"
8383 ],
8484 "recipe_replica_count" : 1 ,
8585 "recipe_container_port" : " 8000" ,
@@ -93,7 +93,7 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
9393```
9494
9595---
96- ### [ Sample Blueprint (Job Mode for Offline vLLM Inference)] ( dhttps://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_deployment_sglang .json )
96+ ### [ Sample Blueprint (Job Mode for Offline vLLM Inference)] ( dhttps://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_deployment_vllm .json )
9797
9898``` json
9999{
@@ -116,7 +116,7 @@ Notes : Make sure your output object storage is in the same tenancy as your stac
116116 "output_object_storage" : [
117117 {
118118 "bucket_name" : " inference_output" ,
119- "mount_location" : " /mlcommons_output " ,
119+ "mount_location" : " /benchmarking_output " ,
120120 "volume_size_in_gbs" : 200
121121 }
122122 ],
@@ -161,7 +161,7 @@ experiment_name: "sglang-bench-doc-test-new"
161161run_name : " llama3-8b-sglang-test"
162162
163163
164- save_metrics_path : /mlcommons_output /benchmark_output_llama3_sglang.json
164+ save_metrics_path : /benchmarking_output /benchmark_output_llama3_sglang.json
165165
166166` ` `
167167## [Sample Config File - 2 vLLM (` offline_vllm_example.yaml`)](https://github.com/oracle-quickstart/oci-ai-blueprints/blob/offline-inference-benchmark/docs/sample_blueprints/offline-inference-infra/offline_vllm_example.yaml)
@@ -194,7 +194,7 @@ distributed_executor_backend: mp
194194mlflow_uri: http://mlflow-benchmarking.corrino-oci.com:5000
195195experiment_name: test-bm-suite-doc
196196run_name: llama3-vllm-test
197- save_metrics_path: /mlcommons_output /benchmark_output_llama3_vllm.json
197+ save_metrics_path: /benchmarking_output /benchmark_output_llama3_vllm.json
198198
199199` ` `
200200
0 commit comments