PaddleX/latest/pipeline_deploy/serving #3524
Replies: 29 comments 114 replies
-
您好,paddlex --serve --pipeline image_classification这样启动服务之后,关闭服务的命令是什么呀? |
Beta Was this translation helpful? Give feedback.
-
paddlex --serve --pipeline {产线名称或产线配置文件路径} [{其他命令行选项}] 如果在部署多个产线, 是否要启动多个服务吗? |
Beta Was this translation helpful? Give feedback.
-
服务化部署支持实时数据的高性能推理? |
Beta Was this translation helpful? Give feedback.
-
我在测试‘高稳定性服务化部署’,使用通用OCR SDK,本地docker启动成功,client.py测试GRPCInferenceService通过,Metrics Service访问成功,但是HTTPService请求失败,一直报错400 Bad Request。想问一下HTTPService的接口调用文档有没有,请求参数是什么,怎么才能正确访问HTTPService |
Beta Was this translation helpful? Give feedback.
-
SDK的下载链接都失效了,麻烦修复下吧 |
Beta Was this translation helpful? Give feedback.
-
您好 |
Beta Was this translation helpful? Give feedback.
-
docker部署的,为什么λ localhost ~/PaddleX paddlex --serve --pipeline OCR |
Beta Was this translation helpful? Give feedback.
-
高稳定性服务化部署如何通过http方式调用 |
Beta Was this translation helpful? Give feedback.
-
您好,这个SDK又不能下载了 |
Beta Was this translation helpful? Give feedback.
-
您好,想请教下CUDA版本为12.4的话 怎么获取ccr-2vdh3abv-pub.cnc.bj.baidubce.com/paddlex/hps:paddlex3.1-gpu的版本 这个只支持CUDA版本为11.8的吧? |
Beta Was this translation helpful? Give feedback.
-
服务化部署如何开启高性能推理? |
Beta Was this translation helpful? Give feedback.
-
执行 1.1 paddlex --install serving 一直报错啊,os 版本 AlmaLinux9.6, 在linux 上你们强烈建议使用 docker安装paddlex ,这里怎么没有使用docekr 安装的paddlex 执行服务部署的教程呀, 请大佬指教, 谢谢! Using cached future-1.0.0-py3-none-any.whl (491 kB) [notice] A new release of pip is available: 25.0.1 -> 25.1.1 During handling of the above exception, another exception occurred: Traceback (most recent call last): |
Beta Was this translation helpful? Give feedback.
-
https://paddlepaddle.github.io/PaddleX/latest/pipeline_deploy/serving.html#23 调整后的 pipeline_config.yaml pipeline_name: OCR text_type: general use_doc_preprocessor: True SubPipelines: SubModules: |
Beta Was this translation helpful? Give feedback.
-
对于这种启动服务后还需下载的模型,因为网络原因无法下载怎么办,可以手动挂载吗 I0718 07:56:17.887946 7 grpc_server.cc:4117] Started GRPCInferenceService at 0.0.0.0:8001 The above exception was the direct cause of the following exception: Traceback (most recent call last): The above exception was the direct cause of the following exception: Traceback (most recent call last): |
Beta Was this translation helpful? Give feedback.
-
2.4.2 手动构造 HTTP 请求,构造的请求体格式不对,"data"中的json应该是String类型。 |
Beta Was this translation helpful? Give feedback.
-
请大佬有空了帮忙看看这个问题:
以上运行后,有两个问题。
|
Beta Was this translation helpful? Give feedback.
-
使用 |
Beta Was this translation helpful? Give feedback.
-
高稳定性服务化部署可以不通过docker的方式部署吗?能不能出个具体的教程 |
Beta Was this translation helpful? Give feedback.
-
你好,我想在一个triton中部署两个不同的OCR模型(通用和单独训练过的),我要问一下如何修改一下server.sh这个脚本?我看上面配置了pipe_line_config.yaml这个路径,但是两个模型,存在两个流水线文件,如何配置?还是需要启动两个docker? |
Beta Was this translation helpful? Give feedback.
-
我用paddleX微调了Cascade-FasterRCNN-ResNet50-FPN模型,单模型评估后AP:63,然后通过服务化部署了object_detection产线,模型路径用的是自定义路径。在C#客户端通过http调用这个端口,单张3072*2048的图片,耗时2890毫秒,我们用的平台是windows,不太想用WSL的方式安装高性能推理,想请问一下,有没有其他的方式可以提高推理的速度?能直接在推理电脑里面能直接加载是做好的,类似于以前PaddelX GUI版本的,可以用原生库推理的,后面新版本的PaddleX会支持吗? |
Beta Was this translation helpful? Give feedback.
-
现在docker pull ccr-2vdh3abv-pub.cnc.bj.baidubce.com/paddlex/hps:paddlex3.2-gpu这个3.2版本是支持cuda12的吗 |
Beta Was this translation helpful? Give feedback.
-
请教一下,现在通过服务化部署产线,在容器里面显示的Time taken 都在100ms左右,但我在C#通过http通信耗时在400ms左右,我想了解一下这个耗时会有哪些因素影响?怎么可以调整达到容器的推理的速度呢? C#客户端的:推理耗时:456.5396毫秒 |
Beta Was this translation helpful? Give feedback.
-
shape参数有什么讲究吗?哪里可以看详细? |
Beta Was this translation helpful? Give feedback.
-
请问目前是不支持arm64版本进行高性能部署使用吗? 对应的paddle信息: 启动命令:nohup paddlex --serve --pipeline OCR --port 38080 --use_hpip > /logs/ocr.log 2>&1 & 异常问题: |
Beta Was this translation helpful? Give feedback.
-
root@localhost:/work# paddlex --pipeline OCR \
Creating model: ('PP-LCNet_x1_0_doc_ori', None) 图片为什么使用官方的命令OCR识别图片也没有有看到文字提取是那个地方有问题?,但是如果不使用NPU是可以解析对应的图片的:https://img1.baidu.com/it/u=1855070411,442203363&fm=253&app=138&f=JPEG?w=800&h=1422 图片地址如下 |
Beta Was this translation helpful? Give feedback.
-
直接用命令对比就可以看出问题 请问是什么地方有问题? root@localhost:/work# paddlex --pipeline OCR \
Creating model: ('PP-LCNet_x1_0_doc_ori', None) C++ Traceback (most recent call last):No stack trace in paddle, may be caused by external reasons. Error Message Summary:FatalError: Aborted (core dumped)
Creating model: ('PP-LCNet_x1_0_doc_ori', None)
corrupted size vs. prev_size |
Beta Was this translation helpful? Give feedback.
-
经验证H20使用ccr-2vdh3abv-pub.cnc.bj.baidubce.com/paddlex/hps:paddlex3.2-gpu镜像(cuda11.8),会出现无法识别图片的情况(即:识别的图片是空的、返回的rec_texts也是空的),请问是否有支持cuda12.6的镜像可以使用? |
Beta Was this translation helpful? Give feedback.
-
目前 50 系显卡是不是依旧不支持高性能推理以及高稳定性部署? |
Beta Was this translation helpful? Give feedback.
-
你好,我正在v100服务器上使用高稳定性服务化部署OCR服务,实例配置情况为: 1 backend: "python"
2 max_batch_size: 16
3 input [
4 {
5 name: "input"
6 data_type: TYPE_STRING
7 dims: [ 1 ]
8 }
9 ]
10 output [
11 {
12 name: "output"
13 data_type: TYPE_STRING
14 dims: [ 1 ]
15 }
16 ]
17 instance_group [
18 {
19 count: 2
20 kind: KIND_GPU
21 gpus: [ 2, 3, 4 ]
22 }
23 ] 实例可以正常运行。但当我使用nvidia-smi实时监控gpu占用时,发现在异步并发调用接口(grpc方式)时,始终只能在一张卡上进行推理计算,现象是:在gpu 2上推理几秒钟,切到gpu 3上推理几秒钟……始终无法在多块gpu上并行计算,请问这个是什么原因,有无解决办法呢? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
PaddleX/latest/pipeline_deploy/serving
https://paddlepaddle.github.io/PaddleX/latest/pipeline_deploy/serving.html
Beta Was this translation helpful? Give feedback.
All reactions