diff --git a/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/resnet.py b/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/resnet.py index 41723bde18..96ca16cd80 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/resnet.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/resnet.py @@ -29,9 +29,9 @@ def __init__( """Load a image classification model exported by torchvision.ResNet. :param model_file: (str)Path of model file, e.g resnet/resnet50.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ # call super() to initialize the backend_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/yolov5cls.py b/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/yolov5cls.py index 89900ae67e..92074728ff 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/yolov5cls.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/classification/contrib/yolov5cls.py @@ -89,9 +89,9 @@ def __init__( """Load a YOLOv5Cls model exported by YOLOv5Cls. :param model_file: (str)Path of model file, e.g ./YOLOv5Cls.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(YOLOv5Cls, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/classification/ppcls/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/classification/ppcls/__init__.py index d3cbcb4c90..9ba3eca4fe 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/classification/ppcls/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/classification/ppcls/__init__.py @@ -88,10 +88,10 @@ def __init__( """Load a image classification model exported by PaddleClas. :param model_file: (str)Path of model file, e.g resnet50/inference.pdmodel - :param params_file: (str)Path of parameters file, e.g resnet50/inference.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g resnet50/inference.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str) Path of configuration file for deploy, e.g resnet50/inference_cls.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PaddleClasModel, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/classification/ppshitu/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/classification/ppshitu/__init__.py index 84ed73a1e2..2e6b88006a 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/classification/ppshitu/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/classification/ppshitu/__init__.py @@ -85,10 +85,10 @@ def __init__( """Load a image PPShiTuV2Recognizer model exported by PaddleClas. :param model_file: (str)Path of model file, e.g PPLCNet/inference.pdmodel - :param params_file: (str)Path of parameters file, e.g PPLCNet/inference.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g PPLCNet/inference.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str) Path of configuration file for deploy, e.g PPLCNet/inference_cls.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPShiTuV2Recognizer, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/fastestdet.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/fastestdet.py index 25e08a2105..4ac37de47e 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/fastestdet.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/fastestdet.py @@ -105,9 +105,9 @@ def __init__( """Load a FastestDet model exported by FastestDet. :param model_file: (str)Path of model file, e.g ./FastestDet.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(FastestDet, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/nanodet_plus.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/nanodet_plus.py index a4af542396..52302fd9e8 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/nanodet_plus.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/nanodet_plus.py @@ -29,9 +29,9 @@ def __init__( """Load a NanoDetPlus model exported by NanoDet. :param model_file: (str)Path of model file, e.g ./nanodet.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/rkyolo/rkyolov5.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/rkyolo/rkyolov5.py index 8d1c157587..27f99c9471 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/rkyolo/rkyolov5.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/rkyolo/rkyolov5.py @@ -148,9 +148,9 @@ def __init__(self, model_file, runtime_option=None, model_format=ModelFormat.RKN """Load a RKYOLOV5 model exported by RKYOLOV5. :param model_file: (str)Path of model file, e.g ./yolov5.rknn - :param params_file: (str)Path of parameters file, e.g , if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g , if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option @@ -207,7 +207,7 @@ def __init__(self, model_file, runtime_option=None, model_format=ModelFormat.RKN :param model_file: (str)Path of model file, e.g ./yolox.rknn :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option @@ -264,7 +264,7 @@ def __init__(self, model_file, runtime_option=None, model_format=ModelFormat.RKN :param model_file: (str)Path of model file, e.g ./yolov7.rknn :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/scaled_yolov4.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/scaled_yolov4.py index 8001ddccaf..1537e82730 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/scaled_yolov4.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/scaled_yolov4.py @@ -29,9 +29,9 @@ def __init__( """Load a ScaledYOLOv4 model exported by ScaledYOLOv4. :param model_file: (str)Path of model file, e.g ./scaled_yolov4.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolor.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolor.py index 680e70c991..2688a39f17 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolor.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolor.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOR model exported by YOLOR :param model_file: (str)Path of model file, e.g ./yolor.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5.py index 31faa9997a..3a7d67066d 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5.py @@ -174,9 +174,9 @@ def __init__( """Load a YOLOv5 model exported by YOLOv5. :param model_file: (str)Path of model file, e.g ./yolov5.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5lite.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5lite.py index 740703b561..6be4edb46e 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5lite.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5lite.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOv5Lite model exported by YOLOv5Lite. :param model_file: (str)Path of model file, e.g ./yolov5lite.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5seg.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5seg.py index 0ebc53ddd8..a4b7449b10 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5seg.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov5seg.py @@ -174,9 +174,9 @@ def __init__( """Load a YOLOv5Seg model exported by YOLOv5. :param model_file: (str)Path of model file, e.g ./yolov5s-seg.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(YOLOv5Seg, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov6.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov6.py index 8227f9990e..3fe104bc0b 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov6.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov6.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOv6 model exported by YOLOv6. :param model_file: (str)Path of model file, e.g ./yolov6.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7.py index 86c3eacd5f..543307d0ed 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7.py @@ -134,9 +134,9 @@ def __init__( """Load a YOLOv7 model exported by YOLOv7. :param model_file: (str)Path of model file, e.g ./yolov7.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_ort.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_ort.py index dffda530bf..ce128b890a 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_ort.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_ort.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOv7End2EndORT model exported by YOLOv7. :param model_file: (str)Path of model file, e.g ./yolov7end2end_ort.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_trt.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_trt.py index a61ad2b241..c215b5b021 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_trt.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov7end2end_trt.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOv7End2EndTRT model exported by YOLOv7. :param model_file: (str)Path of model file, e.g ./yolov7end2end_trt.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov8.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov8.py index 2f788d52a0..5a72640389 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov8.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolov8.py @@ -174,9 +174,9 @@ def __init__( """Load a YOLOv8 model exported by YOLOv8. :param model_file: (str)Path of model file, e.g ./yolov8s.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(YOLOv8, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolox.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolox.py index 67b170825c..d83524a920 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolox.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/contrib/yolox.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOX model exported by YOLOX. :param model_file: (str)Path of model file, e.g ./yolox.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/detection/ppdet/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/detection/ppdet/__init__.py index d19092afc7..9e73bb9818 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/detection/ppdet/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/detection/ppdet/__init__.py @@ -102,10 +102,10 @@ def __init__( """Load a PPYOLOE model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ppyoloe/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ppyoloe/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ppyoloe/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -175,9 +175,9 @@ def __init__( """Load a PPYOLO model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ppyolo/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ppyolo/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ppyolo/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -216,10 +216,10 @@ def __init__( """Load a YOLOX model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolox/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -258,10 +258,10 @@ def __init__( """Load a PicoDet model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g picodet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g picodet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g picodet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -297,10 +297,10 @@ def __init__( """Load a FasterRCNN model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g fasterrcnn/model.pdmodel - :param params_file: (str)Path of parameters file, e.g fasterrcnn/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g fasterrcnn/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -339,10 +339,10 @@ def __init__( """Load a YOLOv3 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolov3/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolov3/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolov3/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -381,10 +381,10 @@ def __init__( """Load a SOLOv2 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g solov2/model.pdmodel - :param params_file: (str)Path of parameters file, e.g solov2/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g solov2/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g solov2/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -423,10 +423,10 @@ def __init__( """Load a MaskRCNN model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g fasterrcnn/model.pdmodel - :param params_file: (str)Path of parameters file, e.g fasterrcnn/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g fasterrcnn/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -474,10 +474,10 @@ def __init__( """Load a SSD model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ssd/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ssd/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ssd/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -516,10 +516,10 @@ def __init__( """Load a YOLOv5 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolov5/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolov5/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolov5/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -545,10 +545,10 @@ def __init__( """Load a YOLOv6 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolov6/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolov6/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolov6/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -574,10 +574,10 @@ def __init__( """Load a YOLOv7 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolov7/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolov7/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolov7/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -603,10 +603,10 @@ def __init__( """Load a YOLOv8 model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g yolov8/model.pdmodel - :param params_file: (str)Path of parameters file, e.g yolov8/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolov8/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g yolov8/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -629,10 +629,10 @@ def __init__( """Load a RTMDet model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g rtmdet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g rtmdet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g rtmdet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -658,10 +658,10 @@ def __init__( """Load a CascadeRCNN model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g cascadercnn/model.pdmodel - :param params_file: (str)Path of parameters file, e.g cascadercnn/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g cascadercnn/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -687,10 +687,10 @@ def __init__( """Load a PSSDet model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g pssdet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g pssdet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g pssdet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -716,10 +716,10 @@ def __init__( """Load a RetinaNet model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g retinanet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g retinanet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g retinanet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -745,10 +745,10 @@ def __init__( """Load a PPYOLOESOD model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ppyoloesod/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ppyoloesod/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ppyoloesod/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -774,10 +774,10 @@ def __init__( """Load a FCOS model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g fcos/model.pdmodel - :param params_file: (str)Path of parameters file, e.g fcos/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g fcos/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -803,10 +803,10 @@ def __init__( """Load a TTFNet model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ttfnet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ttfnet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ttfnet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -832,10 +832,10 @@ def __init__( """Load a TOOD model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g tood/model.pdmodel - :param params_file: (str)Path of parameters file, e.g tood/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g tood/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -861,10 +861,10 @@ def __init__( """Load a GFL model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g gfl/model.pdmodel - :param params_file: (str)Path of parameters file, e.g gfl/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g gfl/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) @@ -890,10 +890,10 @@ def __init__( """Load a PaddleDetectionModel model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ppyoloe/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ppyoloe/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ppyoloe/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PaddleDetectionModel, self).__init__(runtime_option) @@ -963,10 +963,10 @@ def __init__( """Load a PPYOLOER model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g ppyoloe_r/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ppyoloe_r/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ppyoloe_r/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe_r/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPYOLOE, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/face_landmark_1000.py b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/face_landmark_1000.py index 73d116ac24..b564ff44d3 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/face_landmark_1000.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/face_landmark_1000.py @@ -29,9 +29,9 @@ def __init__( """Load a face alignment model exported by FaceLandmark1000. :param model_file: (str)Path of model file, e.g ./FaceLandmark1000.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ super(FaceLandmark1000, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pfld.py b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pfld.py index 2bd7f4cc91..e5aaad2714 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pfld.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pfld.py @@ -29,9 +29,9 @@ def __init__( """Load a face alignment model exported by PFLD. :param model_file: (str)Path of model file, e.g pfld/pfld-106-v3.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ super(PFLD, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pipnet.py b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pipnet.py index 523ee0dc25..506acd7b89 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pipnet.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facealign/contrib/pipnet.py @@ -29,9 +29,9 @@ def __init__( """Load a face alignment model exported by PIPNet. :param model_file: (str)Path of model file, e.g ./PIPNet.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ super(PIPNet, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/blazeface.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/blazeface.py index c46bf4a8eb..ca8cfd16f6 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/blazeface.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/blazeface.py @@ -101,9 +101,9 @@ def __init__( """Load a BlazeFace model exported by BlazeFace. :param model_file: (str)Path of model file, e.g ./Blazeface.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(BlazeFace, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/centerface.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/centerface.py index f296c73d2a..29c02acb2d 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/centerface.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/centerface.py @@ -105,9 +105,9 @@ def __init__( """Load a CenterFace model exported by CenterFace. :param model_file: (str)Path of model file, e.g ./CenterFace.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(CenterFace, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/retinaface.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/retinaface.py index 0dd8174835..743b176083 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/retinaface.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/retinaface.py @@ -29,9 +29,9 @@ def __init__( """Load a RetinaFace model exported by RetinaFace. :param model_file: (str)Path of model file, e.g ./retinaface.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/scrfd.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/scrfd.py index f409c3cafe..3177a252bb 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/scrfd.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/scrfd.py @@ -29,9 +29,9 @@ def __init__( """Load a SCRFD model exported by SCRFD. :param model_file: (str)Path of model file, e.g ./scrfd.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/ultraface.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/ultraface.py index e77240adbf..3c8814d9c9 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/ultraface.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/ultraface.py @@ -29,9 +29,9 @@ def __init__( """Load a UltraFace model exported by UltraFace. :param model_file: (str)Path of model file, e.g ./ultraface.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov5face.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov5face.py index 635be07e8e..a5dc4c6c02 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov5face.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov5face.py @@ -29,9 +29,9 @@ def __init__( """Load a YOLOv5Face model exported by YOLOv5Face. :param model_file: (str)Path of model file, e.g ./yolov5face.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov7face.py b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov7face.py index c3f4bd7854..ab2d9fb5cf 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov7face.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/facedet/contrib/yolov7face.py @@ -148,9 +148,9 @@ def __init__( """Load a YOLOv7Face model exported by YOLOv7Face. :param model_file: (str)Path of model file, e.g ./yolov7face.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(YOLOv7Face, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/adaface/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/adaface/__init__.py index 43343870a4..a832b71c66 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/adaface/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/adaface/__init__.py @@ -63,9 +63,9 @@ def __init__( """Load a AdaFace model exported by PaddleClas. :param model_file: (str)Path of model file, e.g adaface/model.pdmodel - :param params_file: (str)Path of parameters file, e.g adaface/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g adaface/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(AdaFace, self).__init__(runtime_option) self._model = C.vision.faceid.AdaFace( diff --git a/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/insightface/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/insightface/__init__.py index e992d934bd..be1fcb688b 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/insightface/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/faceid/contrib/insightface/__init__.py @@ -99,9 +99,9 @@ def __init__( """Load a InsightFaceRecognitionBase model exported by PaddleClas. :param model_file: (str)Path of model file, e.g InsightFaceRecognitionBase/model.pdmodel - :param params_file: (str)Path of parameters file, e.g InsightFaceRecognitionBase/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g InsightFaceRecognitionBase/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(InsightFaceRecognitionBase, self).__init__(runtime_option) self._model = C.vision.faceid.InsightFaceRecognitionBase( @@ -155,9 +155,9 @@ def __init__( ): """Load a ArcFace model exported by PaddleClas. :param model_file: (str)Path of model file, e.g ArcFace/model.pdmodel - :param params_file: (str)Path of parameters file, e.g ArcFace/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ArcFace/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(InsightFaceRecognitionBase, self).__init__(runtime_option) @@ -178,9 +178,9 @@ def __init__( ): """Load a CosFace model exported by PaddleClas. :param model_file: (str)Path of model file, e.g CosFace/model.pdmodel - :param params_file: (str)Path of parameters file, e.g CosFace/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g CosFace/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(InsightFaceRecognitionBase, self).__init__(runtime_option) @@ -201,9 +201,9 @@ def __init__( ): """Load a PartialFC model exported by PaddleClas. :param model_file: (str)Path of model file, e.g PartialFC/model.pdmodel - :param params_file: (str)Path of parameters file, e.g PartialFC/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g PartialFC/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(InsightFaceRecognitionBase, self).__init__(runtime_option) @@ -224,9 +224,9 @@ def __init__( ): """Load a VPL model exported by PaddleClas. :param model_file: (str)Path of model file, e.g VPL/model.pdmodel - :param params_file: (str)Path of parameters file, e.g VPL/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g VPL/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(InsightFaceRecognitionBase, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/generation/contrib/anemigan.py b/deploy/ultra-infer/python/ultra_infer/vision/generation/contrib/anemigan.py index 2ed5fa5bae..500e660b58 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/generation/contrib/anemigan.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/generation/contrib/anemigan.py @@ -57,9 +57,9 @@ def __init__( """Load a AnimeGAN model. :param model_file: (str)Path of model file, e.g ./model.pdmodel - :param params_file: (str)Path of parameters file, e.g ./model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g ./model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # call super constructor to initialize self._runtime_option super(AnimeGAN, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/headpose/contrib/fsanet.py b/deploy/ultra-infer/python/ultra_infer/vision/headpose/contrib/fsanet.py index d93a80ef62..5938a7cfc9 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/headpose/contrib/fsanet.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/headpose/contrib/fsanet.py @@ -29,9 +29,9 @@ def __init__( """Load a headpose model exported by FSANet. :param model_file: (str)Path of model file, e.g fsanet/fsanet-var.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ super(FSANet, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/keypointdetection/pptinypose/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/keypointdetection/pptinypose/__init__.py index 2a7d4a94ed..9d66aa07e1 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/keypointdetection/pptinypose/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/keypointdetection/pptinypose/__init__.py @@ -30,10 +30,10 @@ def __init__( """load a PPTinyPose model exported by PaddleDetection. :param model_file: (str)Path of model file, e.g pptinypose/model.pdmodel - :param params_file: (str)Path of parameters file, e.g pptinypose/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g pptinypose/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g pptinypose/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPTinyPose, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/modnet.py b/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/modnet.py index 4ffe6591d8..680b7f3418 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/modnet.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/modnet.py @@ -29,9 +29,9 @@ def __init__( """Load a MODNet model exported by MODNet. :param model_file: (str)Path of model file, e.g ./modnet.onnx - :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g yolox/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ # 调用基函数进行backend_option的初始化 # 初始化后的option保存在self._runtime_option diff --git a/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/rvm.py b/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/rvm.py index 4520ff55a6..23a77cdace 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/rvm.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/matting/contrib/rvm.py @@ -29,9 +29,9 @@ def __init__( """Load a video matting model exported by RobustVideoMatting. :param model_file: (str)Path of model file, e.g rvm/rvm_mobilenetv3_fp32.onnx - :param params_file: (str)Path of parameters file, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model, default is ONNX + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model, default is ONNX """ super(RobustVideoMatting, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/matting/ppmatting/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/matting/ppmatting/__init__.py index 270b9d11a8..75a4a9b3d6 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/matting/ppmatting/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/matting/ppmatting/__init__.py @@ -30,10 +30,10 @@ def __init__( """Load a PPMatting model exported by PaddleSeg. :param model_file: (str)Path of model file, e.g PPMatting-512/model.pdmodel - :param params_file: (str)Path of parameters file, e.g PPMatting-512/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g PPMatting-512/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str)Path of configuration file for deployment, e.g PPMatting-512/deploy.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPMatting, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/ocr/ppocr/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/ocr/ppocr/__init__.py index a308d2ea9d..a5c29f7851 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/ocr/ppocr/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/ocr/ppocr/__init__.py @@ -80,7 +80,7 @@ def run(self, runtime_results): return self._postprocessor.run(runtime_results) -class UVDocWarpper(UltraInferModel): +class UVDocWrapper(UltraInferModel): def __init__( self, model_file="", @@ -93,18 +93,18 @@ def __init__( :param model_file: (str)Path of model file, e.g ./ch_PP-OCRv3_rec_infer/model.pdmodel. :param params_file: (str)Path of parameter file, e.g ./ch_PP-OCRv3_rec_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ - super(UVDocWarpper, self).__init__(runtime_option) + super(UVDocWrapper, self).__init__(runtime_option) if len(model_file) == 0: - self._model = C.vision.ocr.UVDocWarpper() + self._model = C.vision.ocr.UVDocWrapper() self._runnable = False else: - self._model = C.vision.ocr.UVDocWarpper( + self._model = C.vision.ocr.UVDocWrapper( model_file, params_file, self._runtime_option, model_format ) - assert self.initialized, "UVDocWarpper initialize failed." + assert self.initialized, "UVDocWrapper initialize failed." self._runnable = True def clone(self): @@ -112,11 +112,11 @@ def clone(self): :return: a new OCR recognition model object """ - class UVDocWarpperClone(UVDocWarpper): + class UVDocWrapperClone(UVDocWrapper): def __init__(self, model): self._model = model - clone_model = UVDocWarpperClone(self._model.clone()) + clone_model = UVDocWrapperClone(self._model.clone()) return clone_model def predict(self, input_image): @@ -474,7 +474,7 @@ def __init__( :param model_file: (str)Path of model file, e.g ./ch_PP-OCRv3_det_infer/model.pdmodel. :param params_file: (str)Path of parameter file, e.g ./ch_PP-OCRv3_det_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(DBDetector, self).__init__(runtime_option) @@ -620,7 +620,7 @@ def __init__( :param model_file: (str)Path of model file, e.g ./ch_PP-OCRv3_det_infer/model.pdmodel. :param params_file: (str)Path of parameter file, e.g ./ch_PP-OCRv3_det_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(DBCURVEDetector, self).__init__(runtime_option) @@ -837,7 +837,7 @@ def __init__( :param model_file: (str)Path of model file, e.g ./ch_ppocr_mobile_v2.0_cls_infer/model.pdmodel. :param params_file: (str)Path of parameter file, e.g ./ch_ppocr_mobile_v2.0_cls_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(Classifier, self).__init__(runtime_option) @@ -1002,7 +1002,7 @@ def __init__( :param params_file: (str)Path of parameter file, e.g ./ch_PP-OCRv3_rec_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param label_path: (str)Path of label file used by OCR recognition model. e.g ./ppocr_keys_v1.txt :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(Recognizer, self).__init__(runtime_option) @@ -1127,7 +1127,7 @@ def __init__( :param params_file: (str)Path of parameter file, e.g ./ch_ppocr_mobile_v2.0_cls_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param table_char_dict_path: (str)Path of table_char_dict file, e.g ../ppocr/utils/dict/table_structure_dict_ch.txt :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(StructureV2Table, self).__init__(runtime_option) @@ -1232,7 +1232,7 @@ def __init__( :param model_file: (str)Path of model file, e.g ./picodet_lcnet_x1_0_fgd_layout_infer/model.pdmodel. :param params_file: (str)Path of parameter file, e.g ./picodet_lcnet_x1_0_fgd_layout_infer/model.pdiparams, if the model format is ONNX, this parameter will be ignored. :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(StructureV2Layout, self).__init__(runtime_option) @@ -1713,7 +1713,7 @@ def __init__( :param ser_dict_path: (str) class file path :param class_path: (str) class file path :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU. - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model. + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model. """ super(StructureV2SERViLayoutXLMModel, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/caddn.py b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/caddn.py index ddde3635b8..faf7330275 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/caddn.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/caddn.py @@ -61,7 +61,7 @@ def __init__( :param params_file: (str)Path of parameters file, e.g ./Caddn.pdiparams :param config_file: (str)Path of config file, e.g ./infer_cfg.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(Caddn, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/centerpoint.py b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/centerpoint.py index 275cb2b08c..fb1fe6338c 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/centerpoint.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/centerpoint.py @@ -47,7 +47,7 @@ def __init__( :param params_file: (str)Path of parameters file, e.g ./Centerpoint.pdiparams :param config_file: (str)Path of config file, e.g ./infer_cfg.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(Centerpoint, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/petr.py b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/petr.py index e071a6fde1..3d17b47e54 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/petr.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/petr.py @@ -61,7 +61,7 @@ def __init__( :param params_file: (str)Path of parameters file, e.g ./petr.pdiparams :param config_file: (str)Path of config file, e.g ./infer_cfg.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(Petr, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/smoke.py b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/smoke.py index 4003594008..c493e9f5ab 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/smoke.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/perception/paddle3d/smoke.py @@ -61,7 +61,7 @@ def __init__( :param params_file: (str)Path of parameters file, e.g ./smoke.pdiparams :param config_file: (str)Path of config file, e.g ./infer_cfg.yaml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(Smoke, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/segmentation/ppseg/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/segmentation/ppseg/__init__.py index 79aa09b2b8..c3cbb7112d 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/segmentation/ppseg/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/segmentation/ppseg/__init__.py @@ -40,10 +40,10 @@ def __init__( """Load a image segmentation model exported by PaddleSeg. :param model_file: (str)Path of model file, e.g unet/model.pdmodel - :param params_file: (str)Path of parameters file, e.g unet/model.pdiparams, if the model_fomat is ModelFormat.ONNX, this param will be ignored, can be set as empty string + :param params_file: (str)Path of parameters file, e.g unet/model.pdiparams, if the model_format is ModelFormat.ONNX, this param will be ignored, can be set as empty string :param config_file: (str) Path of configuration file for deploy, e.g unet/deploy.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PaddleSegModel, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/sr/ppsr/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/sr/ppsr/__init__.py index e98f89d3a5..b3a8399add 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/sr/ppsr/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/sr/ppsr/__init__.py @@ -30,7 +30,7 @@ def __init__( :param model_file: (str)Path of model file, e.g PPMSVSR/inference.pdmodel :param params_file: (str)Path of parameters file, e.g PPMSVSR/inference.pdiparams :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPMSVSR, self).__init__(runtime_option) @@ -65,7 +65,7 @@ def __init__( :param model_file: (str)Path of model file, e.g EDVR/inference.pdmodel :param params_file: (str)Path of parameters file, e.g EDVR/inference.pdiparams :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPMSVSR, self).__init__(runtime_option) @@ -100,7 +100,7 @@ def __init__( :param model_file: (str)Path of model file, e.g BasicVSR/inference.pdmodel :param params_file: (str)Path of parameters file, e.g BasicVSR/inference.pdiparams :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPMSVSR, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/python/ultra_infer/vision/tracking/pptracking/__init__.py b/deploy/ultra-infer/python/ultra_infer/vision/tracking/pptracking/__init__.py index c16555f60f..3a191ddb68 100644 --- a/deploy/ultra-infer/python/ultra_infer/vision/tracking/pptracking/__init__.py +++ b/deploy/ultra-infer/python/ultra_infer/vision/tracking/pptracking/__init__.py @@ -32,7 +32,7 @@ def __init__( :param params_file: (str)Path of parameters file, e.g ppyoloe/model.pdiparams :param config_file: (str)Path of configuration file for deployment, e.g ppyoloe/infer_cfg.yml :param runtime_option: (ultra_infer.RuntimeOption)RuntimeOption for inference this model, if it's None, will use the default backend on CPU - :param model_format: (ultra_infer.ModelForamt)Model format of the loaded model + :param model_format: (ultra_infer.ModelFormat)Model format of the loaded model """ super(PPTracking, self).__init__(runtime_option) diff --git a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/ocrmodel_pybind.cc b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/ocrmodel_pybind.cc index 6730f32084..2310ba3d8a 100644 --- a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/ocrmodel_pybind.cc +++ b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/ocrmodel_pybind.cc @@ -69,26 +69,26 @@ void BindPPOCRModel(pybind11::module &m) { return results; }); - pybind11::class_(m, - "UVDocWarpper") + pybind11::class_(m, + "UVDocWrapper") .def(pybind11::init()) .def(pybind11::init<>()) .def_property_readonly("preprocessor", - &vision::ocr::UVDocWarpper::GetPreprocessor) + &vision::ocr::UVDocWrapper::GetPreprocessor) .def_property_readonly("postprocessor", - &vision::ocr::UVDocWarpper::GetPostprocessor) + &vision::ocr::UVDocWrapper::GetPostprocessor) .def("clone", - [](vision::ocr::UVDocWarpper &self) { return self.Clone(); }) + [](vision::ocr::UVDocWrapper &self) { return self.Clone(); }) .def("predict", - [](vision::ocr::UVDocWarpper &self, pybind11::array &data) { + [](vision::ocr::UVDocWrapper &self, pybind11::array &data) { auto mat = PyArrayToCvMat(data); FDTensor res; self.Predict(mat, &res); res.StopSharing(); return res; }) - .def("batch_predict", [](vision::ocr::UVDocWarpper &self, + .def("batch_predict", [](vision::ocr::UVDocWrapper &self, std::vector &data) { std::vector images; for (size_t i = 0; i < data.size(); ++i) { diff --git a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.cc b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.cc similarity index 86% rename from deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.cc rename to deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.cc index c693cc4f50..3cbc42328f 100644 --- a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.cc +++ b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.cc @@ -20,8 +20,8 @@ namespace ultra_infer { namespace vision { namespace ocr { -UVDocWarpper::UVDocWarpper() {} -UVDocWarpper::UVDocWarpper(const std::string &model_file, +UVDocWrapper::UVDocWrapper() {} +UVDocWrapper::UVDocWrapper(const std::string &model_file, const std::string ¶ms_file, const RuntimeOption &custom_option, const ModelFormat &model_format) { @@ -46,7 +46,7 @@ UVDocWarpper::UVDocWarpper(const std::string &model_file, } // Init -bool UVDocWarpper::Initialize() { +bool UVDocWrapper::Initialize() { if (!InitRuntime()) { FDERROR << "Failed to initialize ultra_infer backend." << std::endl; return false; @@ -54,18 +54,18 @@ bool UVDocWarpper::Initialize() { return true; } -std::unique_ptr UVDocWarpper::Clone() const { - std::unique_ptr clone_model = - utils::make_unique(UVDocWarpper(*this)); +std::unique_ptr UVDocWrapper::Clone() const { + std::unique_ptr clone_model = + utils::make_unique(UVDocWrapper(*this)); clone_model->SetRuntime(clone_model->CloneRuntime()); return clone_model; } -bool UVDocWarpper::Predict(cv::Mat *im, FDTensor *result) { +bool UVDocWrapper::Predict(cv::Mat *im, FDTensor *result) { return Predict(*im, result); } -bool UVDocWarpper::Predict(const cv::Mat &img, FDTensor *result) { +bool UVDocWrapper::Predict(const cv::Mat &img, FDTensor *result) { std::vector results; if (!BatchPredict({img}, &results)) { return false; @@ -74,7 +74,7 @@ bool UVDocWarpper::Predict(const cv::Mat &img, FDTensor *result) { return true; } -bool UVDocWarpper::BatchPredict(const std::vector &images, +bool UVDocWrapper::BatchPredict(const std::vector &images, std::vector *results) { std::vector fd_images = WrapMat(images); if (!preprocessor_.Run(&fd_images, &reused_input_tensors_)) { diff --git a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.h b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.h similarity index 88% rename from deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.h rename to deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.h index babc273de2..ed00b87844 100644 --- a/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwarpper.h +++ b/deploy/ultra-infer/ultra_infer/vision/ocr/ppocr/uvdocwrapper.h @@ -27,12 +27,12 @@ namespace vision { */ namespace ocr { -/*! @brief UVDocWarpper object is used to load the detection model provided by +/*! @brief UVDocWrapper object is used to load the detection model provided by * PaddleOCR. */ -class ULTRAINFER_DECL UVDocWarpper : public UltraInferModel { +class ULTRAINFER_DECL UVDocWrapper : public UltraInferModel { public: - UVDocWarpper(); + UVDocWrapper(); /** \brief Set path of model file, and the configuration of runtime * * \param[in] model_file Path of model file, e.g @@ -43,17 +43,17 @@ class ULTRAINFER_DECL UVDocWarpper : public UltraInferModel { * backend defined in `valid_cpu_backends`. \param[in] model_format Model * format of the loaded model, default is Paddle format. */ - UVDocWarpper(const std::string &model_file, + UVDocWrapper(const std::string &model_file, const std::string ¶ms_file = "", const RuntimeOption &custom_option = RuntimeOption(), const ModelFormat &model_format = ModelFormat::PADDLE); - /** \brief Clone a new UVDocWarpper with less memory usage when multiple + /** \brief Clone a new UVDocWrapper with less memory usage when multiple * instances of the same model are created * - * \return new UVDocWarpper* type unique pointer + * \return new UVDocWrapper* type unique pointer */ - virtual std::unique_ptr Clone() const; + virtual std::unique_ptr Clone() const; /// Get model's name std::string ModelName() const { return "pp-uvdoc"; } @@ -87,10 +87,10 @@ class ULTRAINFER_DECL UVDocWarpper : public UltraInferModel { * OCR detection model result will be written to this structure. \return true * if the prediction is succeeded, otherwise false. */ - /// Get preprocessor reference of UVDocWarpperPreprocessor + /// Get preprocessor reference of UVDocWrapperPreprocessor virtual UVDocPreprocessor &GetPreprocessor() { return preprocessor_; } - // / Get postprocessor reference of UVDocWarpperPostprocessor + // / Get postprocessor reference of UVDocWrapperPostprocessor virtual UVDocPostprocessor &GetPostprocessor() { return postprocessor_; } private: diff --git a/paddlex/repo_apis/PaddleClas_api/cls/config.py b/paddlex/repo_apis/PaddleClas_api/cls/config.py index 63d6fa9989..494db89b66 100644 --- a/paddlex/repo_apis/PaddleClas_api/cls/config.py +++ b/paddlex/repo_apis/PaddleClas_api/cls/config.py @@ -250,16 +250,16 @@ def update_num_workers(self, num_workers: int): ] self.update(_cfg) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = [ - f"DataLoader.Train.loader.use_shared_memory={shared_memeory}", - f"DataLoader.Eval.loader.use_shared_memory={shared_memeory}", + f"DataLoader.Train.loader.use_shared_memory={shared_memory}", + f"DataLoader.Eval.loader.use_shared_memory={shared_memory}", ] self.update(_cfg) diff --git a/paddlex/repo_apis/PaddleClas_api/shitu_rec/config.py b/paddlex/repo_apis/PaddleClas_api/shitu_rec/config.py index 1c39202196..2731e2e054 100644 --- a/paddlex/repo_apis/PaddleClas_api/shitu_rec/config.py +++ b/paddlex/repo_apis/PaddleClas_api/shitu_rec/config.py @@ -104,17 +104,17 @@ def update_num_workers(self, num_workers: int): ] self.update(_cfg) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = [ - f"DataLoader.Train.loader.use_shared_memory={shared_memeory}", - f"DataLoader.Eval.Query.loader.use_shared_memory={shared_memeory}", - f"DataLoader.Eval.Gallery.loader.use_shared_memory={shared_memeory}", + f"DataLoader.Train.loader.use_shared_memory={shared_memory}", + f"DataLoader.Eval.Query.loader.use_shared_memory={shared_memory}", + f"DataLoader.Eval.Gallery.loader.use_shared_memory={shared_memory}", ] self.update(_cfg) diff --git a/paddlex/repo_apis/PaddleDetection_api/object_det/config.py b/paddlex/repo_apis/PaddleDetection_api/object_det/config.py index 3e37dbead4..1d0eb366e9 100644 --- a/paddlex/repo_apis/PaddleDetection_api/object_det/config.py +++ b/paddlex/repo_apis/PaddleDetection_api/object_det/config.py @@ -350,14 +350,14 @@ def update_print_mem_info(self, print_mem_info: bool): assert isinstance(print_mem_info, bool), "print_mem_info should be a bool" self.update({"print_mem_info": f"{print_mem_info}"}) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" - self.update({"print_mem_info": f"{shared_memeory}"}) + assert isinstance(shared_memory, bool), "shared_memory should be a bool" + self.update({"print_mem_info": f"{shared_memory}"}) def update_shuffle(self, shuffle: bool): """update shuffle setting of train and eval dataloader diff --git a/paddlex/repo_apis/PaddleOCR_api/formula_rec/config.py b/paddlex/repo_apis/PaddleOCR_api/formula_rec/config.py index 11d67d2393..62374730fd 100644 --- a/paddlex/repo_apis/PaddleOCR_api/formula_rec/config.py +++ b/paddlex/repo_apis/PaddleOCR_api/formula_rec/config.py @@ -327,16 +327,16 @@ def update_print_mem_info(self, print_mem_info: bool): assert isinstance(print_mem_info, bool), "print_mem_info should be a bool" self.update({"Global.print_mem_info": f"{print_mem_info}"}) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = { - "Train.loader.use_shared_memory": f"{shared_memeory}", - "Train.loader.use_shared_memory": f"{shared_memeory}", + "Train.loader.use_shared_memory": f"{shared_memory}", + "Train.loader.use_shared_memory": f"{shared_memory}", } self.update(_cfg) diff --git a/paddlex/repo_apis/PaddleOCR_api/text_rec/config.py b/paddlex/repo_apis/PaddleOCR_api/text_rec/config.py index 3c3fdb8c1a..ef811a9fe5 100644 --- a/paddlex/repo_apis/PaddleOCR_api/text_rec/config.py +++ b/paddlex/repo_apis/PaddleOCR_api/text_rec/config.py @@ -340,16 +340,16 @@ def update_print_mem_info(self, print_mem_info: bool): assert isinstance(print_mem_info, bool), "print_mem_info should be a bool" self.update({"Global.print_mem_info": f"{print_mem_info}"}) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = { - "Train.loader.use_shared_memory": f"{shared_memeory}", - "Train.loader.use_shared_memory": f"{shared_memeory}", + "Train.loader.use_shared_memory": f"{shared_memory}", + "Train.loader.use_shared_memory": f"{shared_memory}", } self.update(_cfg) diff --git a/paddlex/repo_apis/PaddleVideo_api/video_cls/config.py b/paddlex/repo_apis/PaddleVideo_api/video_cls/config.py index 14620942bc..e15f1b931f 100644 --- a/paddlex/repo_apis/PaddleVideo_api/video_cls/config.py +++ b/paddlex/repo_apis/PaddleVideo_api/video_cls/config.py @@ -221,16 +221,16 @@ def update_num_workers(self, num_workers: int): } self.update(_cfg) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = [ - f"DataLoader.Train.loader.use_shared_memory={shared_memeory}", - f"DataLoader.Eval.loader.use_shared_memory={shared_memeory}", + f"DataLoader.Train.loader.use_shared_memory={shared_memory}", + f"DataLoader.Eval.loader.use_shared_memory={shared_memory}", ] self.update(_cfg) diff --git a/paddlex/repo_apis/PaddleVideo_api/video_det/config.py b/paddlex/repo_apis/PaddleVideo_api/video_det/config.py index 561e1c134e..70289f3677 100644 --- a/paddlex/repo_apis/PaddleVideo_api/video_det/config.py +++ b/paddlex/repo_apis/PaddleVideo_api/video_det/config.py @@ -221,16 +221,16 @@ def update_num_workers(self, num_workers: int): } self.update(_cfg) - def update_shared_memory(self, shared_memeory: bool): + def update_shared_memory(self, shared_memory: bool): """update shared memory setting of train and eval dataloader Args: - shared_memeory (bool): whether or not to use shared memory + shared_memory (bool): whether or not to use shared memory """ - assert isinstance(shared_memeory, bool), "shared_memeory should be a bool" + assert isinstance(shared_memory, bool), "shared_memory should be a bool" _cfg = { - "DataLoader.Train.loader.use_shared_memory": shared_memeory, - "DataLoader.Eval.loader.use_shared_memory": shared_memeory, + "DataLoader.Train.loader.use_shared_memory": shared_memory, + "DataLoader.Eval.loader.use_shared_memory": shared_memory, } self.update(_cfg)