From 7cdac86e65ee314c39b93113ced08bc5ebd17335 Mon Sep 17 00:00:00 2001 From: Shekar-77 Date: Sun, 12 Apr 2026 13:48:37 +0530 Subject: [PATCH 1/2] Add model download and update text decoder function --- .../blip-visual-language-processing.ipynb | 31 +++++++++++++------ 1 file changed, 22 insertions(+), 9 deletions(-) diff --git a/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb b/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb index 12aa99790d3..b330ce8d1d0 100644 --- a/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb +++ b/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb @@ -605,6 +605,23 @@ "text_decoder.forward = partial(text_decoder_forward, ov_text_decoder_with_past=ov_text_decoder_with_past)" ] }, + { + "cell_type": "code", + "execution_count": null, + "id": "01ba0c67-cc02-4781-bb3a-472146362a7a", + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "if not Path(\"./blip_model.py\").exists():\n", + " download_file(url=\"https://raw.githubusercontent.com/openvinotoolkit/openvino_notebooks/latest/notebooks/blip-visual-language-processing/blip_model.py\")\n", + "from blip_model import OVBlipModel\n", + "\n", + "ov_model = OVBlipModel(model.config, model.decoder_start_token_id, ov_vision_model, ov_text_encoder, text_decoder)\n", + "out = ov_model.generate_answer(**inputs, max_length=20)" + ] + }, { "attachments": {}, "cell_type": "markdown", @@ -618,18 +635,14 @@ { "cell_type": "code", "execution_count": null, - "id": "01ba0c67-cc02-4781-bb3a-472146362a7a", - "metadata": { - "tags": [] - }, + "id": "fd5bfc87", + "metadata": {}, "outputs": [], "source": [ - "if not Path(\"./blip_model.py\").exists():\n", - " download_file(url=\"https://raw.githubusercontent.com/openvinotoolkit/openvino_notebooks/latest/notebooks/blip-visual-language-processing/blip_model.py\")\n", - "from blip_model import OVBlipModel\n", + "from functools import partial\n", + "from blip_model import text_decoder_forward\n", "\n", - "ov_model = OVBlipModel(model.config, model.decoder_start_token_id, ov_vision_model, ov_text_encoder, text_decoder)\n", - "out = ov_model.generate_answer(**inputs, max_length=20)" + "text_decoder.forward = partial(text_decoder_forward, ov_text_decoder_with_past=ov_text_decoder_with_past)" ] }, { From d5032e08bd297a7bbe32ac691bf51991bd3fb9d7 Mon Sep 17 00:00:00 2001 From: Shekar-77 Date: Sun, 12 Apr 2026 13:54:13 +0530 Subject: [PATCH 2/2] Refactor code cells for model import and usage --- .../blip-visual-language-processing.ipynb | 43 ++++++++++++------- 1 file changed, 28 insertions(+), 15 deletions(-) diff --git a/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb b/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb index b330ce8d1d0..4042fbea76f 100644 --- a/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb +++ b/notebooks/blip-visual-language-processing/blip-visual-language-processing.ipynb @@ -606,21 +606,34 @@ ] }, { - "cell_type": "code", - "execution_count": null, - "id": "01ba0c67-cc02-4781-bb3a-472146362a7a", - "metadata": { - "tags": [] - }, - "outputs": [], - "source": [ - "if not Path(\"./blip_model.py\").exists():\n", - " download_file(url=\"https://raw.githubusercontent.com/openvinotoolkit/openvino_notebooks/latest/notebooks/blip-visual-language-processing/blip_model.py\")\n", - "from blip_model import OVBlipModel\n", - "\n", - "ov_model = OVBlipModel(model.config, model.decoder_start_token_id, ov_vision_model, ov_text_encoder, text_decoder)\n", - "out = ov_model.generate_answer(**inputs, max_length=20)" - ] + "cell_type": "code", + "execution_count": null, + "id": "7a3b4c12-d8e9-4f01-a2b3-c4d5e6f7a8b9", + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "from pathlib import Path\n", + "\n", + "if not Path(\"./blip_model.py\").exists():\n", + " download_file(url=\"https://raw.githubusercontent.com/openvinotoolkit/openvino_notebooks/latest/notebooks/blip-visual-language-processing/blip_model.py\")" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "01ba0c67-cc02-4781-bb3a-472146362a7a", + "metadata": { + "tags": [] + }, + "outputs": [], + "source": [ + "from blip_model import OVBlipModel\n", + "\n", + "ov_model = OVBlipModel(model.config, model.decoder_start_token_id, ov_vision_model, ov_text_encoder, text_decoder)\n", + "out = ov_model.generate_answer(**inputs, max_length=20)" + ] }, { "attachments": {},