You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: modules/developer-lightspeed/con-llm-requirements.adoc
+3-4Lines changed: 3 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,11 +5,10 @@
5
5
6
6
{ls-short} follows a _Bring Your Own Model_ approach. This model means that to function, {ls-short} requires access to a large language model (LLM) which you must provide. An LLM is a type of generative AI that interprets natural language and generates human-like text or audio responses. When an LLM is used as a virtual assistant, the LLM can interpret questions and provide answers in a conversational manner.
7
7
8
-
LLMs are usually provided by a service or server. Since {ls-short} does not provide an LLM for you, you must configure your preferred LLM provider during installation.
9
-
You can use {ls-short} with a number of LLM providers that offer the OpenAI API interface including the following LLMS:
8
+
LLMs are usually provided by a service or server. Since {ls-short} does not provide an LLM for you, you must configure your preferred LLM provider during installation. You can configure the underlying Llama Stack server to integrate with a number of LLM `providers`` that offer compatibility with the OpenAI API including the following LLMs:
10
9
11
10
* OpenAI (cloud-based inference service)
12
-
* Red Hat OpenShift AI (enterprise model builder & inference server)
13
-
* Red Hat Enterprise Linux AI (enterprise inference server)
11
+
* {rhoai-brand-name} (enterprise model builder & inference server)
You must install and configure the {ls-short}, the {lcs-name}, and Llama Stack containers manually.
6
+
{ls-short} operates on a _Bring Your Own Model_ approach, meaning you must provide and configure access to your preferred Large Language Model (LLM) provider for the service to function. The {lcs-name} ({lcs-short}) acts as an intermediary and service layer for interfacing with LLM providers, relying on the Llama Stack service for core operations. {lcs-short} handles LLM provider setup, authentication, feedback, and Retrieval Augmented Generation (RAG). The {lcs-short} is designed to run as a sidecar container within the {product-very-short} Pod.
7
+
8
+
[NOTE]
9
+
====
10
+
If you have already installed the previous {ls-short} (Developer Preview) with Road-Core Service (RCS), remove the previous {ls-short} configurations and settings and reinstall.
11
+
This step is necessary as {ls-short} originally required the use of the {rcs-name} as a sidecar container for interfacing with LLM providers. The updated architecture removes and replaces RCS with the new Lightspeed Core Service (LCS) and Llama Stack, which require a new configuration for volumes, containers, and secrets.
12
+
====
7
13
8
14
.Prerequisites
9
15
* You are logged into your {ocp-short} account.
@@ -13,6 +19,8 @@ You must install and configure the {ls-short}, the {lcs-name}, and Llama Stack c
13
19
14
20
.Procedure
15
21
22
+
You must install and configure the {ls-short}, the {lcs-name}, and Llama Stack containers manually.
23
+
16
24
. Create the {lcs-short} ConfigMap (`lightspeed-stack.yaml`).
17
25
+
18
26
[IMPORTANT]
@@ -55,11 +63,9 @@ mcp_servers:
55
63
provider_id: model-context-protocol
56
64
url: https://<RHDH_HOST>/api/mcp-actions/v1
57
65
----
58
-
+
59
66
where:
60
67
61
-
62
-
(Optional) `mcp_servers`:: Set this configuration if you want to integrate MCP. See {model-context-protocol-link}[{model-context-protocol-title}].
68
+
`mcp_servers`:: Optional: Set this configuration if you want to integrate MCP. See {model-context-protocol-link}[{model-context-protocol-title}].
63
69
`mcp_servers:name`:: This value must match the entry in the {ls-short} app config file for MCP servers.
64
70
`model-context-protocol`:: This is the tool runtime provider defined and configured in the llama-stack `run.yaml` configuration for use in {lcs-short}.
... Add the Llama Stack Secret file as shown in the following example:
489
+
... Add the {ls-short} and Llama Stack Secret file as shown in the following example:
457
490
+
458
491
[source,yaml]
459
492
----
460
493
envFrom:
494
+
- secretRef:
495
+
name: lightspeed-secrets
461
496
- secretRef:
462
497
name: llama-stack-secrets
463
498
----
499
+
... In your `deployment.patch.spec.template.spec.containers.env` section, set the {lcs-short} and Llama Stack environment variables as shown in the following example:
500
+
+
501
+
[source,yaml]
502
+
----
503
+
- name: PROJECT
504
+
value: rhdh
505
+
- name: LLAMA_CONFIG_FILE
506
+
value: /app-root/config/llama-stack/run.yaml
507
+
----
464
508
+
465
509
[NOTE]
466
510
====
467
-
* Your {product-very-short} container is typically already present in your Kubernetes custom resource (CR). You are adding the two additional container definitions for {lcs-short} and Llama Stack as the {lcs-short} sidecar.
511
+
* Your {product-very-short} container is typically already present in your CR. You are adding the two additional container definitions for {lcs-short} and Llama Stack as the {lcs-short} sidecar.
468
512
====
469
513
... Click *Save*. The Pods are automatically restarted.
470
514
+
@@ -515,10 +559,6 @@ spec:
515
559
subPath: lightspeed-stack.yaml
516
560
- mountPath: /tmp/data/feedback
517
561
name: shared-storage
518
-
# OPTIONAL: `llama-stack-config mount path configuration is only needed if you want to override the custom configuration
0 commit comments