You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/class1/class1.rst
+12-9Lines changed: 12 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@ Class 1: The fundamental of Generative Artificial Intelligent (AI)
5
5
6
6
AI Primer
7
7
---------
8
-
As GenAI Practitioner, you need to minimally understand what it means with the following term and able to articulate the concept introduced. The following not an exhasive list. The purpose if to get you started. For details, please refer to various contents. Content below were extracted from collection of resources from Internet.
8
+
As GenAI Practitioner, you need to minimally understand what it means with the following term and able to articulate the concept introduced. The following not an exhaustive list. The purpose is to get you started. For details, please refer to various contents. Content below were extracted from collection of resources from Internet.
9
9
10
10
11
11
What is AI and how different with GenAI?
@@ -27,12 +27,13 @@ In summary, LLMs are a specialized application of machine learning, focusing on
27
27
What is SLM?
28
28
~~~~~~~~~~~~
29
29
SLM (Small Language Model) is a type of artificial intelligence model designed for natural language processing that has fewer parameters (between million to few billion) and computational requirements compared to large language models.
30
-
Parameter Scale Examples
31
30
32
-
Small language model: 100 million to 1 billion parameters
33
-
Medium model: 1-10 billion parameters
34
-
Large models like GPT-3: 175 billion parameters
35
-
Very large models: 500 billion to 1 trillion parameters
31
+
**Parameter Scale Examples**
32
+
33
+
- Small language model: 100 million to 1 billion parameters
34
+
- Medium model: 1-10 billion parameters
35
+
- Large models like GPT-3: 175 billion parameters
36
+
- Very large models: 500 billion to 1 trillion parameters
36
37
37
38
38
39
What is ML?
@@ -43,13 +44,15 @@ What hallucination means in AI?
43
44
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
44
45
Hallucination in AI is when an AI model generates information that is false, inaccurate, or completely made up, even though it might sound convincing. It's like the AI "imagining" things that aren't real or aren't supported by its training data.
45
46
46
-
For instance, if you ask an AI about a person named "Olivia Smith" it might confidently generate a detailed biography about a specific Oliver Smith, complete with birth date and achievements, even though it's not referring to any real person – it's just combining patterns it learned during training
47
+
For instance, if you ask an AI about a person named "Olivia Smith" it might confidently generate a detailed biography about a specific Olivia Smith, complete with birth date and achievements, even though it's not referring to any real person – it's just combining patterns it learned during training
47
48
48
49
49
50
What "token" means in context in AI?
50
51
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
51
-
A token is the basic unit of text that an AI model processes and understands. A token can be a word, part of a word, or even a punctuation mark. Tokenization is the process of breaking down text into these individual tokens
52
+
A token is the basic unit of text that an AI model processes and understands. A token can be a word, part of a word, or even a punctuation mark. Tokenization is the process of breaking down text into these individual tokens.
53
+
52
54
Typically, 1 token is approximately:
55
+
53
56
- 4 characters of English text
54
57
- About 3/4 of a word
55
58
- Varies slightly between different AI models
@@ -94,7 +97,7 @@ What is "temperature" in AI?
94
97
~~~~~~~~~~~~~~~~~~~~~~~~~~~~
95
98
Temperature controlled “the creativity of the response. It is a hyperparameter that controls the randomness or creativity of the model's output during text generation.
96
99
97
-
Low temperature (close to 0) makes the output more deterministic and focused. Model selects the most probable words. Responses become more percise and consistent. Lesss creative and more conservative outputs.
100
+
Low temperature (close to 0) makes the output more deterministic and focused. Model selects the most probable words. Responses become more percise and consistent. Less creative and more conservative outputs.
98
101
99
102
High temperature (closer to 1 or above) increases randomness and creativity. Model is more likely to choose less probable words. Outputs become more diverse and unpredictable. Can generate more unique and imaginative responses.
After login to linux Jumphost, change directory to **webapps**. Jumphost server was installed with utilities called 'direnv' - https://direnv.net/. Its a tools that will load the environment file (kubeconfig) when you switch to that directory. Its an efficient tools to switch K8S context from one cluster to the other just by changing directory.
9
+
.. image:: ./_static/class2-0.png
10
+
11
+
After login/putty to linux Jumphost (from Windows10 Jumphost), change directory to **webapps**. Jumphost server was installed with utilities called 'direnv' - https://direnv.net/. Its a tools that will load the environment file (kubeconfig) when you switch to that directory. Its an efficient tools to switch K8S context from one cluster to the other just by changing directory.
10
12
11
13
.. NOTE::
12
14
Refer to **Prerequsite** section to find the password for the Windows10 and Linux Jumphost.
@@ -96,6 +98,8 @@ values-plus-nap.yaml ::
96
98
2 - Deploy Arcadia Financial Modern Apps
97
99
----------------------------------------
98
100
101
+
.. image:: ./_static/class2-3-0.png
102
+
99
103
Deploy Arcadia Financial application on Kubernetes. Arcadia Trading consist of multiple microservices.
Copy file name to clipboardExpand all lines: docs/class3/class3.rst
+15-6Lines changed: 15 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -43,6 +43,8 @@ With the growing popularity of Generative AI, your organization has decided to u
43
43
2 - Deploy Nginx Ingress Controller for AIGW K8S
44
44
------------------------------------------------
45
45
46
+
.. image:: ./_static/class3-1-1.png
47
+
46
48
.. code-block:: bash
47
49
48
50
cd~/ai-gateway/nginx-ingress
@@ -67,6 +69,8 @@ With the growing popularity of Generative AI, your organization has decided to u
67
69
3 - Deploy Open-WebUI with Ollama Service
68
70
-----------------------------------------
69
71
72
+
.. image:: ./_static/class3-4-0.png
73
+
70
74
**Open-Webui** is a self-hosted WebUI that allow user to interact with AI models. It allow user to download respective language model that use by Ollama.
71
75
72
76
**Ollama** is a open-source tools that allow user to run large language model. User can expose model access via Ollama inference API.
@@ -199,7 +203,7 @@ Test interacting with LLM model. Feel free to test with different language model
199
203
.. attention::
200
204
Please do notes that GenAI is hallucinating and providing wrong information - about F5 Inc headquarters. Please ignore as smaller model (smaller parameter, less intelligent) tend to hallucinate more compare to a larger model. Large models with more parameters are more capable and intelligent than smaller models, but require expensive machines with multiple GPUs to run.
201
205
202
-
Its also depends on dataset use for the training - "Garbage In, Garbage Out".
206
+
It also depends on dataset use for the training - "Garbage In, Garbage Out".
203
207
204
208
205
209
5 - Deploy LLM model service (Ollama)
@@ -238,6 +242,9 @@ ollama-ingress-http.yaml ::
238
242
239
243
6 - Deploy LLM orchestrator service (Flowise AI)
240
244
------------------------------------------------
245
+
246
+
.. image:: ./_static/class3-13-0.png
247
+
241
248
Deploy LLM Orchstrator to facilitate AI component communication. Flowise AI - an open source low-code tool for developer to build customized LLM orchstration flow and AI agent is used. (https://flowiseai.com/). Flowise complements LangChain by offering a visual interface.
242
249
243
250
.. code-block:: bash
@@ -271,7 +278,7 @@ Deploy LLM Orchstrator to facilitate AI component communication. Flowise AI - an
271
278
272
279
.. image:: ./_static/class3-13.png
273
280
274
-
Flowise is installed with the following custom values. You can login with the password shown in the value file.
281
+
Flowise is installed with the following custom values. Plese take notes of the password as you may need it for the next section.
275
282
276
283
values.yaml ::
277
284
@@ -327,15 +334,15 @@ Confirm that you can login and access to LLM orchestrator (flowise)
327
334
328
335
.. image:: ./_static/class3-15.png
329
336
330
-
Import arcadia RAG chatflow into flowise. Select **Add New**, settings icons and **Load Chatflow**
337
+
Import arcadia RAG chatflow into flowise. Select **Add New**, click **Settings icons** and **Load Chatflow**
331
338
332
339
.. image:: ./_static/class3-16.png
333
340
334
341
A copy of the chatflow located on the jumphost **Documents** directory. Select the chatflow json file.
335
342
336
343
.. image:: ./_static/class3-17.png
337
344
338
-
Save the chatflow with a name as shown.
345
+
Save the chatflow (arcadia-rag)
339
346
340
347
.. image:: ./_static/class3-18.png
341
348
@@ -354,6 +361,8 @@ Save the chatflow with a name as shown.
354
361
7 - Deploy Vector Database (Qdrant)
355
362
-----------------------------------
356
363
364
+
.. image:: ./_static/class3-20-0.png
365
+
357
366
**Qdrant** is a vector similarity search engine and vector database. It provides a production-ready service with a convenient API to store, search, and manage vectors points.
358
367
359
368
|
@@ -426,7 +435,7 @@ Load the imported "arcadia-rag" chatflow.
426
435
.. image:: ./_static/class3-24.png
427
436
428
437
429
-
Here is the full RAG pipeline implmented in a low-code platform.
438
+
Here is the full RAG pipeline implemented in a low-code platform.
430
439
431
440
.. image:: ./_static/class3-25.png
432
441
@@ -437,7 +446,7 @@ Here are some of the node/chain used.
0 commit comments