Skip to content

Commit f774197

Browse files
authored
update log (#77)
1 parent 00ec9b3 commit f774197

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

llmserve/backend/llm/pipelines/llamacpp/llamacpp_pipeline.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -246,13 +246,14 @@ def streamGenerate(self, prompt: str, **generate_kwargs) -> Generator[str, None,
246246
logger.info(f"chat generate_kwargs: {generate_kwargs}")
247247
output = self.model.create_chat_completion(messages=inputs[0], stream=True, **generate_kwargs)
248248
for chunk in output:
249-
logger.info(f'LlamaCppPipeline -> create_chat_completion -> Yield -> "{chunk}" -> "{type(chunk)}"')
249+
# logger.info(f'LlamaCppPipeline -> create_chat_completion -> Yield -> "{chunk}" -> "{type(chunk)}"')
250250
delta = chunk['choices'][0]['delta']
251251
val = ''
252252
if 'role' in delta:
253253
val = ''
254254
elif 'content' in delta:
255255
val = delta['content']
256+
logger.info(f'LlamaCppPipeline -> create_chat_completion -> Yield -> "{val}"')
256257
yield val
257258
else:
258259
logger.info(f"generate_kwargs: {generate_kwargs}")

0 commit comments

Comments
 (0)