You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/en/docs/concepts/workloads/autoscaling.md
+18-17Lines changed: 18 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,38 +11,39 @@ weight: 40
11
11
In Kubernetes, you can _scale_ a workload depending on the current demand of resources.
12
12
This allows your cluster to react to changes in resource demand more elastically and efficiently.
13
13
14
+
When you scale a workload, you can either increase or decrease the number of replicas managed by
15
+
the workload, or adjust the resources available to the replicas in-place.
16
+
17
+
The first approach is referred to as _horizontal scaling_, while the second is referred to as
18
+
_vertical scaling_.
19
+
14
20
There are manual and automatic ways to scale your workloads, depending on your use case.
15
21
16
22
<!-- body -->
17
23
18
24
## Scaling workloads manually
19
25
20
-
Kubernetes supports _manual scaling_ of workloads, either by changing the number of
21
-
{{< glossary_tooltip text="replicas" term_id="replica">}} defined for an object that manages a set of
22
-
{{< glossary_tooltip text="Pods" term_id="pod" >}} (for example a {{< glossary_tooltip text="Deployment" term_id="deployment" >}}),
23
-
or by adjusting the resource requests and limits of the replicas managed by the workload
24
-
(for example CPU or memory):
26
+
Kubernetes supports _manual scaling_ of workloads. Horizontal scaling can be done
27
+
using the `kubectl` CLI.
28
+
For vertical scaling, you need to _patch_ the resource definition of your workload.
29
+
30
+
See below for examples of both strategies.
25
31
26
-
-[Running multiple instances of your app](/docs/tutorials/kubernetes-basics/scale/scale-intro/)
27
-
-[Resizing CPU and memory resources assigned to containers](/docs/tasks/configure-pod-container/resize-container-resources)
32
+
-**Horizontal scaling**: [Running multiple instances of your app](/docs/tutorials/kubernetes-basics/scale/scale-intro/)
33
+
-**Vertical scaling**: [Resizing CPU and memory resources assigned to containers](/docs/tasks/configure-pod-container/resize-container-resources)
28
34
29
35
{{< note >}}
30
-
Resizing a workload in-place **without** restarting the Pods or its Containers requires Kubernetes version 1.27 or later.
36
+
Resizing a workload in-place **without** restarting the {{< glossary_tooltip text="Pods" term_id="pod" >}}
37
+
or its {{< glossary_tooltip text="Containers" term_id="container" >}} requires Kubernetes version 1.27 or later.
31
38
{{< /note >}}
32
39
33
40
## Scaling workloads automatically
34
41
35
42
Kubernetes also supports _automatic scaling_ of workloads, which is the focus of this page.
36
43
37
-
The concept of _Autoscaling_ in Kubernetes refers to the ability to automatically update the workloads of your cluster. This can be either an object that manages a set of Pods (for example a
38
-
{{< glossary_tooltip text="Deployment" term_id="deployment" >}} or Pods or PodTemplates themselves.
39
-
40
-
Depending on _what_ is being scaled, there are also different options for _how_ to scale:
41
-
42
-
- scale the number of available instances (such as Replicas in a Deployment)
43
-
- scale the available resources on existing instances themselves (such as CPU or memory of a {{< glossary_tooltip text="Container" term_id="container" >}} in a Pod)
44
-
45
-
The first option is referred to as _horizontal scaling_, while the second is referred to as _vertical scaling_.
44
+
The concept of _Autoscaling_ in Kubernetes refers to the ability to automatically update an
0 commit comments