Skip to content

Commit 0fb3244

Browse files
committed
fix text version and add precision in readme
1 parent 6d4a116 commit 0fb3244

File tree

3 files changed

+38
-8
lines changed

3 files changed

+38
-8
lines changed

README.md

Lines changed: 30 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -281,9 +281,37 @@ gcloud beta artifacts repositories create $ARTIFACT_REGISTRY_NAME \
281281
--repository-format=docker \
282282
--location=$REGION
283283
```
284-
3. Create AI Platform version model using the docker image of the torchserve API package.
285284
```bash
286-
gcloud beta ai-platform versions create $VERSION_NAME --region=$REGION --model=$MODEL_NAME --machine-type=n1-standard-4 --image=$DOCKER_IMAGE_PATH --ports=$PORT --health-route=$HEALTH_ROUTE --predict-route=$PREDICT_ROUTE
285+
docker build -t $REGION-docker.pkg.dev/$GCP_PROJECT_ID/$ARTIFACT_REGISTRY_NAME/$DOCKER_NAME:$DOCKER_TAG . -f TextDockerfile
286+
```
287+
```bash
288+
docker push $REGION-docker.pkg.dev/$GCP_PROJECT_ID/$ARTIFACT_REGISTRY_NAME/$DOCKER_NAME:$DOCKER_TAG
289+
```
290+
291+
3. [Optional] Run your docker locally and try to send a prediction
292+
293+
```bash
294+
docker run -d -p 8080:8080 --name local_imdb $REGION-docker.pkg.dev/$GCP_PROJECT_ID/$ARTIFACT_REGISTRY_NAME/$DOCKER_NAME:$DOCKER_TAG
295+
```
296+
```bash
297+
curl -X POST -H "Content-Type: application/json" -d '["this was a bad movie"]' 127.0.0.1:8080/predictions/fastai_model
298+
```
299+
If everything is working okay, you should receive a response from the server in your console
300+
```bash
301+
[
302+
{
303+
"Categories": "0",
304+
"Tensor": [
305+
0.9999990463256836,
306+
9.371918849865324e-07
307+
]
308+
}
309+
]
310+
```
311+
312+
4. Create AI Platform version model using the docker image of the torchserve API package.
313+
```bash
314+
gcloud beta ai-platform versions create $VERSION_NAME --region=$REGION --model=$MODEL_NAME --image=$REGION-docker.pkg.dev/$GCP_PROJECT_ID/$ARTIFACT_REGISTRY_NAME/$DOCKER_NAME:$DOCKER_TAG --ports=8080 --health-route=/ping --predict-route=/predictions/fastai_model
287315
```
288316

289317

TextDockerfile

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,16 +15,16 @@ COPY requirements.txt /home/model-server/
1515
RUN pip install --upgrade pip
1616
RUN pip install -r /home/model-server/requirements.txt
1717

18-
COPY torch_serve/text/handler.py torch_serve/text/config.py torch_serve/text/vocab.json /home/model-server/
18+
ADD torch_serve/ /home/model-server/torch_serve/
1919

2020
COPY model/text/fastai_cls_weights.pth /home/model-server/
2121

2222
RUN torch-model-archiver \
2323
--model-name=fastai_model \
2424
--version=1.0 \
2525
--serialized-file=/home/model-server/fastai_cls_weights.pth \
26-
--extra-files=/home/model-server/config.py,/home/model-server/vocab.json \
27-
--handler=/home/model-server/handler.py \
26+
--extra-files=/home/model-server/torch_serve/text/config.py,/home/model-server/torch_serve/text/vocab.json \
27+
--handler=/home/model-server/torch_serve/text/handler.py \
2828
--export-path=/home/model-server/model-store/
2929

3030

torch_serve/text/handler.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -77,14 +77,12 @@ def _load_preprocessor(self):
7777
self.preprocessor = preprocessor
7878

7979
def preprocess(self, data):
80-
logger.debug(f"input data : {data}")
8180
text = data[0].get("data")
8281
if text is None:
8382
text = data[0].get("body")
84-
logger.info(f"text : {text} of type {type(text)}")
8583
if not isinstance(text, list):
8684
text = [text]
87-
logger.info(f"text : {text}")
85+
logger.info(f"text: {text}")
8886
text_preprocessed = [self.preprocessor(el) for el in text]
8987
max_len = find_max_list(text_preprocessed)
9088
text_preprocessed = [
@@ -132,8 +130,12 @@ def handle(data, context):
132130
if data is None:
133131
return None
134132

133+
logger.info(f"raw data: {data}")
135134
data = _service.preprocess(data)
135+
logger.info(f"preprocessed data: {data}")
136136
data = _service.inference(data)
137+
logger.info(f"infered data: {data}")
137138
data = _service.postprocess(data)
139+
logger.info(f"postprocessed data: {data}")
138140

139141
return data

0 commit comments

Comments
 (0)