You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/learning-paths/servers-and-cloud-computing/multiarch_ollama_on_gke/_index.md
+12-14Lines changed: 12 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,20 +1,19 @@
1
1
---
2
-
title: Run ollama on both arm64 and amd64 nodes, using the same multi-architecture container image on GKE.
2
+
title: Use GKE to run Ollama on arm64 and amd64 nodes using a multi-architecture container image
3
3
4
4
minutes_to_complete: 30
5
5
6
-
who_is_this_for: This learning path will show you how easy it is to migrate from homogenous amd64 k8s clusters, to a hybrid (arm64 and amd64) cluster with multi-architectural container images on GKE. Demonstrated with the ollama application, you'll see for yourself the price/performance advantages of running on arm64. Although tutorial will be GKE-specific with ollama, the provided YAML can act as a template for deployment on any on any multi-architectural application and cloud.
6
+
who_is_this_for: This topic explains how migrate a homogenous amd64 k8s cluster to a hybrid (arm64 and amd64) cluster using a multi-architecture container image on GKE. Ollama is the application used to demonstrate the migration.
7
7
8
8
9
9
learning_objectives:
10
10
- Spin up a GKE cluster with amd64 and arm64 nodes.
11
-
- Apply ollama amd64-based and arm64-based Deployments and Services using the same container image.
12
-
- Ping, pull models, and make inferences to experience each architectures' performance first-hand.
13
-
- Experiment further on your own by researching which existing, and future workloads could benefit most from single, or multi-architectural clusters.
11
+
- Apply Ollama amd64-based and arm64-based deployments and services using the same container image.
12
+
- Ping, pull models, and make inferences to experience the performance of each architecture.
14
13
15
14
prerequisites:
16
15
- A [Google Cloud account](https://console.cloud.google.com/).
17
-
- A computer with [Google Cloud CLI](/install-guides/gcloud) and [kubectl](/install-guides/kubectl/) installed.
16
+
- A local computer with the [Google Cloud CLI](/install-guides/gcloud) and [kubectl](/install-guides/kubectl/) installed.
18
17
- The [GKE Cloud Plugin](https://cloud.google.com/kubernetes-engine/docs/how-to/cluster-access-for-kubectl#gcloud)
19
18
20
19
@@ -26,35 +25,34 @@ skilllevels: Introductory
26
25
27
26
subjects: Containers and Virtualization
28
27
cloud_service_providers: Google Cloud
29
-
30
28
31
29
armips:
32
30
- Neoverse
33
31
34
32
operatingsystems:
35
33
- Linux
36
-
- MacOs
34
+
- macOS
37
35
38
36
tools_software_languages:
39
-
- LLM
40
-
- ollama
41
-
- GenAI
37
+
- LLM
38
+
- Ollama
39
+
- GenAI
42
40
43
41
further_reading:
44
42
- resource:
45
-
title: ollama - Get up and running with large language models
43
+
title: Ollama - Get up and running with large language models
0 commit comments