@@ -194,6 +194,7 @@ async def generate_text(self, prompt: Prompt):
194
194
# priority=QueuePriority.GENERATE_TEXT,
195
195
# start_timestamp=start_timestamp,
196
196
)
197
+ logger .info (f"generated text: { text } " )
197
198
# return text[0]
198
199
return text
199
200
@@ -351,13 +352,9 @@ async def predict(self, model: str, prompt: Union[Prompt, List[Prompt]]) -> Unio
351
352
logger .info (f"set modelID: { item } " )
352
353
logger .info (f"search model key { modelID } " )
353
354
if isinstance (prompt , Prompt ):
354
- results = await asyncio .gather (
355
- * (await asyncio .gather (* [self ._models [modelID ].generate_text .remote (prompt )]))
356
- )
355
+ results = await asyncio .gather (* [self ._models [modelID ].generate_text .remote (prompt )])
357
356
elif isinstance (prompt , list ):
358
- results = await asyncio .gather (
359
- * (await asyncio .gather (* [self ._models [modelID ].batch_generate_text .remote (prompt )]))
360
- )
357
+ results = await asyncio .gather (* [self ._models [modelID ].batch_generate_text .remote (prompt )])
361
358
else :
362
359
raise Exception ("Invaid prompt format." )
363
360
logger .info (f"{ results } " )
@@ -418,9 +415,7 @@ async def query(self, *args) -> Dict[str, Dict[str, Any]]:
418
415
else :
419
416
prompts = args [0 ]
420
417
logger .info (f"ExperimentalDeployment query.prompts { prompts } " )
421
- results = await asyncio .gather (
422
- * (await asyncio .gather (* [self ._model .generate_text .remote (Prompt (prompt = prompts , use_prompt_format = False ))]))
423
- )
418
+ results = await asyncio .gather (* [(self ._model .generate_text .remote (Prompt (prompt = prompts , use_prompt_format = False )))])
424
419
logger .info (f"ExperimentalDeployment query.results { results } " )
425
420
results = results [0 ]
426
421
return results .generated_text
0 commit comments