Skip to content

Commit 1d81d33

Browse files
fix path params issue, make interface consistent (#78)
1 parent f774197 commit 1d81d33

File tree

2 files changed

+6
-6
lines changed

2 files changed

+6
-6
lines changed

llmserve/backend/llm/pipelines/llamacpp/llamacpp_pipeline.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -242,6 +242,7 @@ def streamGenerate(self, prompt: str, **generate_kwargs) -> Generator[str, None,
242242
logger.info(f"stream generate_kwargs: {generate_kwargs}")
243243
logger.info(f"model inputs: {inputs}")
244244
generate_kwargs.pop('stopping_sequences', None)
245+
generate_kwargs.pop('echo', None)
245246
if chat_completion:
246247
logger.info(f"chat generate_kwargs: {generate_kwargs}")
247248
output = self.model.create_chat_completion(messages=inputs[0], stream=True, **generate_kwargs)

llmserve/backend/server/app.py

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -353,19 +353,18 @@ async def metadata(self, model: str) -> Dict[str, Dict[str, Any]]:
353353
async def models(self) -> List[str]:
354354
return list(self._models.keys())
355355

356-
@app.post("/run/stream")
357-
def streamer(self, data: dict) -> StreamingResponse:
358-
logger.info(f"data: {data}")
359-
logger.info(f'Got stream -> body: {data}, keys: {self._models.keys()}')
360-
prompt = data.get("prompt")
361-
model = data.get("model")
356+
@app.post("/{model}/run/stream")
357+
def streamer(self, model: str, prompt: Union[Prompt, List[Prompt]]) -> StreamingResponse:
358+
logger.info(f"url: {model}, keys: {self._models.keys()}")
359+
362360
modelKeys = list(self._models.keys())
363361
modelID = model
364362
for item in modelKeys:
365363
logger.info(f"_reverse_prefix(item): {_reverse_prefix(item)}")
366364
if _reverse_prefix(item) == model:
367365
modelID = item
368366
logger.info(f"set stream model id: {item}")
367+
369368
logger.info(f"search stream model key: {modelID}")
370369
return StreamingResponse(self.streamer_generate_text(modelID, prompt), media_type="text/plain")
371370

0 commit comments

Comments
 (0)