Skip to content

Commit 77fac3a

Browse files
efazalopenshift-merge-bot[bot]
authored andcommitted
Updated example notebook to use Natural questions dataset to train RAG model, embedding and generator model, added compute metrics to observe model training, updated inference code. Removed geenration questions.
Signed-off-by: Esa Fazal <[email protected]>
1 parent 189a93f commit 77fac3a

File tree

3 files changed

+690
-2359
lines changed

3 files changed

+690
-2359
lines changed

examples/kfto-sft-feast-rag/README.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,9 @@ The core idea is to enhance a generator model (like BART) by providing it with r
1010

1111
Before you begin, ensure you have the following setup:
1212

13-
* An OpenShift cluster with OpenShift AI (RHOAI) 2.17+ installed:
13+
* An OpenShift cluster with OpenShift AI (RHOAI) 2.20+ installed:
1414
* The `dashboard`, `trainingoperator` and `workbenches` components enabled
15+
* Workbench with medium size container, 1 NVIDIA GPU accelerator, and cluster storage of 200GB.
1516
* Sufficient worker nodes for your configuration(s) with NVIDIA GPUs (Ampere-based or newer recommended)
1617
* A dynamic storage provisioner supporting RWX PVC provisioning
1718
* A standalone Milvus deployment. See example [here](https://github.com/rh-aiservices-bu/llm-on-openshift/tree/main/vector-databases/milvus#deployment).
@@ -27,7 +28,8 @@ You must run this notebook from within an OpenShift AI Workbench. Follow these s
2728
* Once the project is created, click on _Create a workbench_.
2829
* Then create a workbench with a preferred name and with the following settings:
2930
* Select the `PyTorch` (or the `ROCm-PyTorch`) workbench image with the recommended version.
30-
* Select the `Medium` as the deployment container size and add an accelerator.
31+
* Select the `Medium` as the deployment container size.
32+
* Add an accelerator (GPU).
3133
* Create a storage that'll be shared between the workbench and the fine-tuning runs.
3234
Make sure it uses a storage class with RWX capability and give it enough size according to the size of the model you want to fine-tune.
3335
> [!NOTE]

0 commit comments

Comments
 (0)