@@ -82,7 +82,7 @@ async def generate_python_vllm_output(
8282 return python_vllm_output
8383
8484
85- def prepare_vllm_baseline_outputs (
85+ async def prepare_vllm_baseline_outputs (
8686 export_file = "vllm_baseline_output.pkl" , prompts = PROMPTS , guided_generation = None
8787):
8888 """
@@ -93,13 +93,12 @@ def prepare_vllm_baseline_outputs(
9393 llm_engine = AsyncLLMEngine .from_engine_args (AsyncEngineArgs (** VLLM_ENGINE_CONFIG ))
9494 python_vllm_output = []
9595 for i in range (len (prompts )):
96- python_vllm_output .extend (
97- asyncio .run (
98- generate_python_vllm_output (
99- prompts [i ], llm_engine , guided_generation = guided_generation
100- )
101- )
96+ output = await generate_python_vllm_output (
97+ prompts [i ], llm_engine , guided_generation = guided_generation
10298 )
99+ if output :
100+ python_vllm_output .extend (output )
101+
103102 with open (export_file , "wb" ) as f :
104103 pickle .dump (python_vllm_output , f )
105104
@@ -240,7 +239,7 @@ def tearDown(self):
240239 )
241240 FLAGS = parser .parse_args ()
242241 if FLAGS .generate_baseline :
243- prepare_vllm_baseline_outputs ()
242+ asyncio . run ( prepare_vllm_baseline_outputs () )
244243 exit (0 )
245244
246245 if FLAGS .generate_guided_baseline :
@@ -249,10 +248,12 @@ def tearDown(self):
249248 "backend" : "outlines" ,
250249 }
251250 guided_generation = GuidedDecodingParams (** guided_decoding_params )
252- prepare_vllm_baseline_outputs (
253- export_file = "vllm_guided_baseline_output.pkl" ,
254- prompts = GUIDED_PROMPTS ,
255- guided_generation = guided_generation ,
251+ asyncio .run (
252+ prepare_vllm_baseline_outputs (
253+ export_file = "vllm_guided_baseline_output.pkl" ,
254+ prompts = GUIDED_PROMPTS ,
255+ guided_generation = guided_generation ,
256+ )
256257 )
257258 exit (0 )
258259
0 commit comments