Skip to content

Commit db90191

Browse files
committed
deleted unsupported notebooks, added @notebooks, changed generalml_p38_cpu_x86_64_v1 to generalml_p311_cpu_x86_64_v1
1 parent 2a9c2eb commit db90191

26 files changed

+649
-3429
lines changed

notebook_examples/README.md

Lines changed: 246 additions & 251 deletions
Large diffs are not rendered by default.

notebook_examples/aqua-batch-inferencing.ipynb

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,23 @@
11
{
22
"cells": [
3+
{
4+
"cell_type": "raw",
5+
"id": "fac705fa",
6+
"metadata": {
7+
"vscode": {
8+
"languageId": "raw"
9+
}
10+
},
11+
"source": [
12+
"@notebook{aqua-batch-inferencing.ipynb,\n",
13+
" title: AI Quick Action - Batch inferencing,\n",
14+
" summary: Perform batch inferencing on LLMs using AI Quick Actions.,\n",
15+
" developed_on: generalml_p311_cpu_x86_64_v1,\n",
16+
" keywords: quick action, batch, inferencing, llm\n",
17+
" license: Universal Permissive License v 1.0\n",
18+
"}"
19+
]
20+
},
321
{
422
"cell_type": "markdown",
523
"id": "f9990553-e6f8-4293-b89e-946f3e4c17e3",

notebook_examples/aqua-deploy-llm-byoc.ipynb

Lines changed: 19 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,31 @@
11
{
22
"cells": [
3+
{
4+
"cell_type": "raw",
5+
"id": "748139a6",
6+
"metadata": {
7+
"vscode": {
8+
"languageId": "raw"
9+
}
10+
},
11+
"source": [
12+
"@notebook{aqua-deploy-llm-byoc.ipynb,\n",
13+
" title: Deploy LLM Models using BYOC,\n",
14+
" summary: Deploy and perform inferencing using AI Quick Action models.,\n",
15+
" developed_on: generalml_p311_cpu_x86_64_v1,\n",
16+
" keywords: byoc, llm, quick action, deploy\n",
17+
" license: Universal Permissive License v 1.0\n",
18+
"}"
19+
]
20+
},
321
{
422
"cell_type": "markdown",
523
"id": "443c3053",
624
"metadata": {},
725
"source": [
826
"# Deploy LLM Models using BYOC\n",
927
"\n",
10-
"This guide demonstrates how to deploy and perform inference using AI Quick Action registered models with Oracle Data Science Service Managed Containers (SMC) powered by vLLM. In this example, we will use a model downloaded from Hugging Face specifically, [Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) from Meta. \n"
28+
"This guide demonstrates how to deploy and perform inferencing using AI Quick Action registered models with Oracle Data Science Service Managed Containers (SMC) powered by vLLM. In this example, we will use a model downloaded from Hugging Face specifically, [Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) from Meta. \n"
1129
]
1230
},
1331
{

0 commit comments

Comments
 (0)