@@ -56,21 +56,10 @@ BasicBackend::BasicBackend(const ONNX_NAMESPACE::ModelProto& model_proto,
5656 exe_network_ = global_context_.ie_core .LoadNetwork (ie_cnn_network_, remote_context_, subgraph_context_.subgraph_name );
5757 LOGS_DEFAULT (INFO) << log_tag << " Loaded model to the plugin" ;
5858 } else {
59- #if defined(OPENVINO_2023_0)
60- if (subgraph_context.precision != InferenceEngine::Precision::FP16) {
61- const std::string model = model_proto.SerializeAsString ();
62- exe_network_ = global_context_.ie_core .LoadNetwork (model, hw_target, device_config, subgraph_context_.subgraph_name );
63- LOGS_DEFAULT (INFO) << log_tag << " Loaded model to the plugin" ;
64- } else {
65- ie_cnn_network_ = CreateOVModel (model_proto, global_context_, subgraph_context_, const_outputs_map_);
66- exe_network_ = global_context_.ie_core .LoadNetwork (ie_cnn_network_, hw_target, device_config, subgraph_context_.subgraph_name );
67- LOGS_DEFAULT (INFO) << log_tag << " Loaded model to the plugin" ;
68- }
69- #else
7059 ie_cnn_network_ = CreateOVModel (model_proto, global_context_, subgraph_context_, const_outputs_map_);
7160 exe_network_ = global_context_.ie_core .LoadNetwork (ie_cnn_network_, hw_target, device_config, subgraph_context_.subgraph_name );
7261 LOGS_DEFAULT (INFO) << log_tag << " Loaded model to the plugin" ;
73- # endif
62+ }
7463#else
7564#if defined(OPENVINO_2023_0)
7665 if (subgraph_context.precision != InferenceEngine::Precision::FP16 && global_context_.enable_dynamic_shapes == false ) {
0 commit comments