Skip to content

Commit 587ace3

Browse files
authored
Merge pull request #496 from elizjo/main
Notebook support for generalml_p311_cpu_x86_64_v1 conda pack
2 parents 58a5560 + db90191 commit 587ace3

26 files changed

+649
-3429
lines changed

notebook_examples/README.md

Lines changed: 246 additions & 251 deletions
Large diffs are not rendered by default.

notebook_examples/aqua-batch-inferencing.ipynb

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,23 @@
11
{
22
"cells": [
3+
{
4+
"cell_type": "raw",
5+
"id": "fac705fa",
6+
"metadata": {
7+
"vscode": {
8+
"languageId": "raw"
9+
}
10+
},
11+
"source": [
12+
"@notebook{aqua-batch-inferencing.ipynb,\n",
13+
" title: AI Quick Action - Batch inferencing,\n",
14+
" summary: Perform batch inferencing on LLMs using AI Quick Actions.,\n",
15+
" developed_on: generalml_p311_cpu_x86_64_v1,\n",
16+
" keywords: quick action, batch, inferencing, llm\n",
17+
" license: Universal Permissive License v 1.0\n",
18+
"}"
19+
]
20+
},
321
{
422
"cell_type": "markdown",
523
"id": "f9990553-e6f8-4293-b89e-946f3e4c17e3",

notebook_examples/aqua-deploy-llm-byoc.ipynb

Lines changed: 19 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,31 @@
11
{
22
"cells": [
3+
{
4+
"cell_type": "raw",
5+
"id": "748139a6",
6+
"metadata": {
7+
"vscode": {
8+
"languageId": "raw"
9+
}
10+
},
11+
"source": [
12+
"@notebook{aqua-deploy-llm-byoc.ipynb,\n",
13+
" title: Deploy LLM Models using BYOC,\n",
14+
" summary: Deploy and perform inferencing using AI Quick Action models.,\n",
15+
" developed_on: generalml_p311_cpu_x86_64_v1,\n",
16+
" keywords: byoc, llm, quick action, deploy\n",
17+
" license: Universal Permissive License v 1.0\n",
18+
"}"
19+
]
20+
},
321
{
422
"cell_type": "markdown",
523
"id": "443c3053",
624
"metadata": {},
725
"source": [
826
"# Deploy LLM Models using BYOC\n",
927
"\n",
10-
"This guide demonstrates how to deploy and perform inference using AI Quick Action registered models with Oracle Data Science Service Managed Containers (SMC) powered by vLLM. In this example, we will use a model downloaded from Hugging Face specifically, [Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) from Meta. \n"
28+
"This guide demonstrates how to deploy and perform inferencing using AI Quick Action registered models with Oracle Data Science Service Managed Containers (SMC) powered by vLLM. In this example, we will use a model downloaded from Hugging Face specifically, [Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) from Meta. \n"
1129
]
1230
},
1331
{

0 commit comments

Comments
 (0)