From d5abdb0d265e02e72f5d05111497901dbaf956ac Mon Sep 17 00:00:00 2001 From: lyer-liu <55565795+lyer-liu@users.noreply.github.com> Date: Mon, 12 Jun 2023 17:15:01 +0800 Subject: [PATCH] Update custom_model_handler.py MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 接口访问都需要带这几个参数,调整后该参数为可选项 --- paddlenlp/server/handlers/custom_model_handler.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/paddlenlp/server/handlers/custom_model_handler.py b/paddlenlp/server/handlers/custom_model_handler.py index c5a133b8ef19..b6b466d5089f 100644 --- a/paddlenlp/server/handlers/custom_model_handler.py +++ b/paddlenlp/server/handlers/custom_model_handler.py @@ -26,9 +26,9 @@ def __init__(self): def process(cls, predictor, tokenizer, data, parameters): max_seq_len = 128 batch_size = 1 - if "max_seq_len" not in parameters: + if "max_seq_len" in parameters: max_seq_len = parameters["max_seq_len"] - if "batch_size" not in parameters: + if "batch_size" in parameters: batch_size = parameters["batch_size"] text = None if "text" in data: @@ -97,9 +97,9 @@ def __init__(self): def process(cls, predictor, tokenizer, data, parameters): max_seq_len = 128 batch_size = 1 - if "max_seq_len" not in parameters: + if "max_seq_len" in parameters: max_seq_len = parameters["max_seq_len"] - if "batch_size" not in parameters: + if "batch_size" in parameters: batch_size = parameters["batch_size"] text = None if "text" in data: