Skip to content

Commit 0869543

Browse files
keep removing deprecated stuff (#69)
1 parent f3d2847 commit 0869543

File tree

3 files changed

+5
-32
lines changed

3 files changed

+5
-32
lines changed

llmserve/backend/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
from llmserve.backend.server.run import llm_application, llm_server
1+
from llmserve.backend.server.run import llm_server
22
from llmserve.backend.server.run import run as serve_model
33

4-
__all__ = ["llm_application", "llm_server", "serve_model"]
4+
__all__ = ["llm_server", "serve_model"]

llmserve/backend/server/app.py

Lines changed: 3 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -159,7 +159,6 @@ async def reconfigure(
159159
@property
160160
def max_batch_size(self):
161161
return (self.args.model_config.generation.max_batch_size if self.args.model_config.generation else 1)
162-
# return 1
163162

164163
@property
165164
def batch_wait_timeout_s(self):
@@ -194,30 +193,12 @@ async def generate_text(self, prompt: Prompt):
194193
with async_timeout.timeout(GATEWAY_TIMEOUT_S):
195194
text = await self.generate_text_batch(
196195
prompt,
197-
# [prompt],
198-
# priority=QueuePriority.GENERATE_TEXT,
199196
# start_timestamp=start_timestamp,
200197
)
201198
logger.info(f"generated text: {text}")
202199
# return text[0]
203200
return text
204201

205-
# no need anymore, will be delete soon
206-
async def generate(self, prompt: Prompt):
207-
time.time()
208-
logger.info(prompt)
209-
logger.info(self.get_max_batch_size())
210-
logger.info(self.get_batch_wait_timeout_s())
211-
with async_timeout.timeout(GATEWAY_TIMEOUT_S):
212-
text = await self.generate_text_batch(
213-
prompt,
214-
# [prompt],
215-
# priority=QueuePriority.GENERATE_TEXT,
216-
# start_timestamp=start_timestamp,
217-
)
218-
return text
219-
# return text[0]
220-
221202
@app.post("/batch", include_in_schema=False)
222203
async def batch_generate_text(self, prompts: List[Prompt]):
223204
logger.info(f"batch_generate_text prompts: {prompts} ")
@@ -229,7 +210,6 @@ async def batch_generate_text(self, prompts: List[Prompt]):
229210
*[
230211
self.generate_text_batch(
231212
prompt,
232-
# priority=QueuePriority.BATCH_GENERATE_TEXT,
233213
# start_timestamp=start_timestamp,
234214
)
235215
for prompt in prompts
@@ -333,20 +313,22 @@ def __init__(self, models: Dict[str, DeploymentHandle], model_configurations: Di
333313
async def predict(self, model: str, prompt: Union[Prompt, List[Prompt]]) -> Union[Dict[str, Any], List[Dict[str, Any]], List[Any]]:
334314
logger.info(f"url: {model}, keys: {self._models.keys()}")
335315
modelKeys = list(self._models.keys())
336-
# model = _replace_prefix(model)
316+
337317
modelID = model
338318
for item in modelKeys:
339319
logger.info(f"_reverse_prefix(item): {_reverse_prefix(item)}")
340320
if _reverse_prefix(item) == model:
341321
modelID = item
342322
logger.info(f"set modelID: {item}")
343323
logger.info(f"search model key {modelID}")
324+
344325
if isinstance(prompt, Prompt):
345326
results = await asyncio.gather(*[self._models[modelID].generate_text.remote(prompt)])
346327
elif isinstance(prompt, list):
347328
results = await asyncio.gather(*[self._models[modelID].batch_generate_text.remote(prompt)])
348329
else:
349330
raise Exception("Invaid prompt format.")
331+
350332
logger.info(f"{results}")
351333
return results[0]
352334

llmserve/backend/server/run.py

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -133,15 +133,6 @@ def llm_experimental(args: Union[str, LLMApp, List[Union[LLMApp, str]]]):
133133
return (ExperimentalDeployment.bind(deployment, model), serve_conf) # pylint:disable=no-member
134134

135135

136-
def llm_application(args):
137-
"""This is a simple wrapper for LLM Server
138-
That is compatible with the yaml config file format
139-
140-
"""
141-
serve_args = ServeArgs.parse_obj(args)
142-
return llm_server(serve_args.models)[0]
143-
144-
145136
def run(models: Union[LLMApp, str], appname: str = None, port: int = DEFAULT_HTTP_PORT):
146137
"""Run the LLM Server on the local Ray Cluster
147138

0 commit comments

Comments
 (0)