diff --git a/modules/llama_cpp_plugin/include/plugin.hpp b/modules/llama_cpp_plugin/include/plugin.hpp index cab026b3f..802b929b6 100644 --- a/modules/llama_cpp_plugin/include/plugin.hpp +++ b/modules/llama_cpp_plugin/include/plugin.hpp @@ -36,10 +36,16 @@ class LlamaCppPlugin : public IPlugin { virtual std::shared_ptr import_model(std::istream& model, const ov::SoPtr& context, const ov::AnyMap& properties) const override; - virtual ov::SupportedOpsMap query_model(const std::shared_ptr& model, const ov::AnyMap& properties) const override; + + virtual std::shared_ptr import_model(const ov::Tensor& model, const ov::AnyMap& properties) const override; + + virtual std::shared_ptr import_model(const ov::Tensor& model, + const ov::SoPtr& context, + const ov::AnyMap& properties) const override; + private: size_t m_num_threads = 0; }; diff --git a/modules/llama_cpp_plugin/src/plugin.cpp b/modules/llama_cpp_plugin/src/plugin.cpp index 737c326ac..d633de744 100644 --- a/modules/llama_cpp_plugin/src/plugin.cpp +++ b/modules/llama_cpp_plugin/src/plugin.cpp @@ -33,6 +33,18 @@ std::shared_ptr LlamaCppPlugin::compile_model(const std::sha OPENVINO_THROW_NOT_IMPLEMENTED("Currently only direct GGUF file loading is " "supported for the LLAMA_CPP* plugins"); } + +std::shared_ptr LlamaCppPlugin::import_model(const ov::Tensor& model, const ov::AnyMap& properties) const { + OPENVINO_THROW("This method may not be used with LLAMA_CPP* plugins"); +} + +std::shared_ptr LlamaCppPlugin::import_model(const ov::Tensor& model, + const ov::SoPtr& context, + const ov::AnyMap& properties) const { + OPENVINO_THROW("This method may not be used with LLAMA_CPP* plugins"); +} + + std::shared_ptr LlamaCppPlugin::compile_model(const std::string& fname, const ov::AnyMap& properties) const { size_t num_threads = 0;