You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/kfto-dreambooth/README.md
+35-1Lines changed: 35 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,6 +5,8 @@ The finetuning is performed on OpenShift environment using Kubeflow Training ope
5
5
6
6
This example is based on HuggingFace DreamBooth Hackathon example - https://huggingface.co/learn/diffusion-course/en/hackathon/dreambooth
7
7
8
+
> [!TIP]
9
+
> **Multi-Team Resource Management**: For enterprise scenarios with multiple teams sharing GPU resources, see the [**Kueue Multi-Team Resource Management Workshop**](../../workshops/kueue/README.md). It demonstrates how to use this LLM fine-tuning example with Kueue for fair resource allocation, borrowing policies, and workload scheduling across teams.
8
10
9
11
## Requirements
10
12
@@ -45,4 +47,36 @@ This example is based on HuggingFace DreamBooth Hackathon example - https://hugg
45
47
* From the workbench, clone this repository, i.e., `https://github.com/opendatahub-io/distributed-workloads.git`
46
48
* Navigate to the `distributed-workloads/examples/kfto-dreambooth` directory and open the `dreambooth` notebook
47
49
48
-
You can now proceed with the instructions from the notebook. Enjoy!
50
+
You can now proceed with the instructions from the notebook. Enjoy!
51
+
52
+
> [!IMPORTANT]
53
+
> **Kueue Integration (RHOAI 2.21+):**
54
+
> * If using RHOAI 2.21+, the example supports Kueue integration for workload management:
55
+
> * When using Kueue:
56
+
> * Follow the [Configure Kueue (Optional)](#configure-kueue-optional) section to set up required resources
57
+
> * Add the local-queue name label to your job configuration to enforce workload management
58
+
> * You can skip Kueue usage by:
59
+
> * Disabling the existing `kueue-validating-admission-policy-binding`
60
+
> * Omitting the local-queue-name label in your job configuration
61
+
>
62
+
> **Note:** Kueue Enablement via Validating Admission Policy was introduced in RHOAI-2.21. You can skip this section if using an earlier RHOAI release version.
63
+
64
+
### Configure Kueue (Optional)
65
+
66
+
> [!NOTE]
67
+
> This section is only required if you plan to use Kueue for workload management (RHOAI 2.21+) or Kueue is not already configured in your cluster.
68
+
> The Kueue resource YAML files referenced below are located in the [Kueue workshop directory](../../workshops/kueue), specifically in `workshops/kueue/resources/`. You can use these files as templates for your own setup or copy them into your project as needed.
69
+
70
+
* Update the `nodeLabels` in the `workshops/kueue/resources/resource_flavor.yaml` file to match your AI worker nodes
" # labels={\"kueue.x-k8s.io/queue-name\": \"<LOCAL_QUEUE_NAME>\"}, # Optional: Add local queue name and uncomment these lines if using Kueue for resource management\n",
Copy file name to clipboardExpand all lines: examples/kfto-feast/README.md
+41Lines changed: 41 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -35,6 +35,9 @@ By integrating Feast into the fine-tuning pipeline, we ensure that the training
35
35
36
36
---
37
37
38
+
> [!TIP]
39
+
> **Multi-Team Resource Management**: For enterprise scenarios with multiple teams sharing GPU resources, see the [**Kueue Multi-Team Resource Management Workshop**](../../workshops/kueue/README.md). It demonstrates how to use this LLM fine-tuning example with Kueue for fair resource allocation, borrowing policies, and workload scheduling across teams.
40
+
38
41
## Requirements
39
42
40
43
* An OpenShift cluster with OpenShift AI (RHOAI) 2.17+ installed:
@@ -109,4 +112,42 @@ By following this notebook, you'll gain hands-on experience in setting up a **fe
109
112
110
113
You can now proceed with the instructions from the notebook. Enjoy!
111
114
115
+
> [!IMPORTANT]
116
+
> **Hugging Face Token Requirements:**
117
+
> * You will need a Hugging Face token if using gated models:
118
+
> * The examples use gated Llama models that require a token (e.g., https://huggingface.co/meta-llama/Llama-3.1-8B)
119
+
> * Set the `HF_TOKEN` environment variable in your job configuration
120
+
> * Note: You can skip the token if switching to non-gated models
121
+
>
122
+
> **Kueue Integration (RHOAI 2.21+):**
123
+
> * If using RHOAI 2.21+, the example supports Kueue integration for workload management:
124
+
> * When using Kueue:
125
+
> * Follow the [Configure Kueue (Optional)](#configure-kueue-optional) section to set up required resources
126
+
> * Add the local-queue name label to your job configuration to enforce workload management
127
+
> * You can skip Kueue usage by:
128
+
> * Disabling the existing `kueue-validating-admission-policy-binding`
129
+
> * Omitting the local-queue-name label in your job configuration
130
+
>
131
+
> **Note:** Kueue Enablement via Validating Admission Policy was introduced in RHOAI-2.21. You can skip this section if using an earlier RHOAI release version.
132
+
133
+
### Configure Kueue (Optional)
134
+
135
+
> [!NOTE]
136
+
> This section is only required if you plan to use Kueue for workload management (RHOAI 2.21+) or Kueue is not already configured in your cluster.
137
+
> The Kueue resource YAML files referenced below are located in the [Kueue workshop directory](../../workshops/kueue), specifically in `workshops/kueue/resources/`. You can use these files as templates for your own setup or copy them into your project as needed.
138
+
139
+
* Update the `nodeLabels` in the `workshops/kueue/resources/resource_flavor.yaml` file to match your AI worker nodes
Copy file name to clipboardExpand all lines: examples/kfto-feast/kfto_feast.ipynb
+1Lines changed: 1 addition & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -1328,6 +1328,7 @@
1328
1328
"\"USE_LORA\": \"true\", # Whether to apply LoRA adapters in the standard (full‑precision) mode.\n",
1329
1329
"\"USE_QLORA\":\"false\", # Whether to apply QLoRA, which loads the model in 4‑bit quantized mode and then applies LoRA adapters.\n",
1330
1330
" }, \n",
1331
+
" # labels={\"kueue.x-k8s.io/queue-name\": \"<LOCAL_QUEUE_NAME>\"}, # Optional: Add local queue name and uncomment these lines if using Kueue for resource management\n",
0 commit comments