You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/api-management/how-to-self-hosted-gateway-on-kubernetes-in-production.md
+4-5Lines changed: 4 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,25 +38,24 @@ While we provide [guidance on the minimum number of replicas](#number-of-replica
38
38
39
39
There are two ways to autoscale the self-hosted gateway horizontally:
40
40
41
-
- Autoscale based on resource usage (CPU & Memory)
41
+
- Autoscale based on resource usage (CPU and memory)
42
42
- Autoscale based on the number of requests per second
43
43
44
44
This is possible through native Kubernetes functionality, or by using [Kubernetes Event-driven Autoscaling (KEDA)](https://keda.sh/). KEDA is a CNCF Incubation project that strives to make application autoscaling simple.
45
45
46
46
> [!NOTE]
47
47
> KEDA is an open-source technology that is not supported by Azure support and needs to be operated by customers.
48
48
49
-
### Resource-based Autoscaling
49
+
### Resource-based autoscaling
50
50
51
-
Kubernetes allows you to autoscale the self-hosted gateway based on resource usage by using a [Horizontal Pod Autoscaler](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/). It allows you to [define CPU & memory thresholds](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/#support-for-resource-metrics), and the number of replicas to scale out or in.
51
+
Kubernetes allows you to autoscale the self-hosted gateway based on resource usage by using a [Horizontal Pod Autoscaler](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/). It allows you to [define CPU and memory thresholds](https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/#support-for-resource-metrics), and the number of replicas to scale out or in.
52
52
53
53
An alternative is to use Kubernetes Event-driven Autoscaling (KEDA) allowing you to scale workloads based on a [variety of scalers](https://keda.sh/docs/latest/scalers/), including CPU and memory.
54
-
We recommend using KEDA to scale the self-hosted gateway horizontally, if you are already using KEDA to scale other workloads as a unified app autoscaler. If that is not the case, then we strongly suggest to rely on the native Kubernetes functionality through Horizontal Pod Autoscaler.
55
54
56
55
> [!TIP]
57
56
> If you are already using KEDA to scale other workloads, we recommend using KEDA as a unified app autoscaler. If that is not the case, then we strongly suggest to rely on the native Kubernetes functionality through Horizontal Pod Autoscaler.
58
57
59
-
### Traffic-based Autoscaling
58
+
### Traffic-based autoscaling
60
59
61
60
Kubernetes does not provide an out-of-the-box mechanism for traffic-based autoscaling.
0 commit comments