Skip to content

Commit 3c1418c

Browse files
committed
Addressing review comments
Signed-off-by: ndixita <ndixita@google.com>
1 parent 90b6b90 commit 3c1418c

File tree

1 file changed

+15
-10
lines changed
  • keps/sig-node/5419-pod-level-resources-in-place-resize

1 file changed

+15
-10
lines changed

keps/sig-node/5419-pod-level-resources-in-place-resize/README.md

Lines changed: 15 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -229,11 +229,10 @@ type PodStatus struct {
229229

230230
##### Resize Restart Policy
231231

232-
Pod-level resize policy is not supported in the alpha stage of Pod-level resource
233-
feature. While a pod-level resize policy might be beneficial for VM-based runtimes
234-
like Kata Containers (potentially allowing the hypervisor to restart the entire VM
235-
on resize), this is a topic for future consideration. We plan to engage with the
236-
Kata community to discuss this further and will re-evaluate the need for a pod-level
232+
Pod-level resize policy is not supported in this KEP. While a pod-level resize policy might
233+
be beneficial for VM-based runtimes like Kata Containers (potentially allowing the hypervisor
234+
to restart the entire VM on resize), this is a topic for future consideration as a separate KEP.
235+
We plan to engage with the Kata community to discuss this further and will re-evaluate the need for a pod-level
237236
policy in subsequent development stages.
238237

239238
The absence of a pod-level resize policy means that container restarts are
@@ -477,10 +476,15 @@ necessary to implement this enhancement.
477476

478477
e2e tests provide good test coverage of interactions between the new pod-level
479478
resource feature and existing Kubernetes components i.e. API server, kubelet, cgroup
480-
enforcement. We may replicate and/or move some of the E2E tests functionality into integration tests before Beta using data from any issues we uncover that are not covered by planned and implemented tests.
479+
enforcement. We may replicate and/or move some of the E2E tests functionality into
480+
integration tests before GA using data from any issues we uncover that are not
481+
covered by planned and implemented tests.
481482

482483
#### e2e tests
483484

485+
* - [Pod Level Resources Resize](https://github.com/kubernetes/kubernetes/blob/master/test/e2e/common/node/pod_level_resources_resize.go): [SIG Node](https://testgrid.k8s.io/sig-node-presubmits#pr-kubelet-e2e-podlevelresources-resize), [triage search](https://storage.googleapis.com/k8s-triage/index.html?ci=0&pr=1&sig=node&job=pr-kubelet-e2e-podlevelresources-resize)
486+
487+
484488
Following scenarios need to be covered:
485489

486490
* Cgroup settings when pod-level resources are set.
@@ -511,9 +515,6 @@ Support the basic functionality for kubelet to translate pod-level requests/limi
511515
* Actual pod resource data may be cached in memory, which will be refreshed after
512516
each successful pod resize or for every cache-miss.
513517
* Coverage for upgrade->downgrade->upgrade scenarios.
514-
* Consensus with Kata community to on pod-level resize policy. If it is required to
515-
support pod-level resize policy, we will have to do another alpha launch of the
516-
feature as it would involve an API change.
517518
* Extend instrumentation from
518519
[KEP#1287](https://github.com/kubernetes/enhancements/blob/ef7e11d088086afd84d26c9249a4ca480df2d05a/keps/sig-node/1287-in-place-update-pod-resources/README.md)
519520
for Pod-level resource resize.
@@ -691,6 +692,9 @@ Testing plan:
691692
* Restart API server with feature enabled
692693
* Verify original test pod is still running
693694

695+
Initial manual verification was completed following the Alpha release ([Results](https://docs.google.com/document/d/19dKnTxH34YjSzrQCMqmpNp9iJk4YfWXXf5ytqNvV4c0/edit?usp=sharing)).
696+
Comprehensive automated testing and E2E coverage are slated for implementation prior to GA graduation."
697+
694698
###### Is the rollout accompanied by any deprecations and/or removals of features, APIs, fields of API types, flags, etc.?
695699

696700
No
@@ -828,7 +832,7 @@ Focusing mostly on:
828832
- One new PATCH PodStatus API call in response to Pod resize request.
829833
- No additional overhead unless Pod resize is invoked.
830834
- estimated throughput
831-
- Proportional to the number of resize requests ssued by users or controllers (e.g., VPA). For a typical cluster this is expected to be < 1% of total Pod update traffic.
835+
- Proportional to the number of resize requests issued by users or controllers (e.g., VPA). For a typical cluster this is expected to be < 1% of total Pod update traffic.
832836
- originating component(s) (e.g. Kubelet, Feature-X-controller)
833837
- Kubelet
834838
focusing mostly on:
@@ -883,6 +887,7 @@ Think about adding additional work or introducing new steps in between
883887

884888
[existing SLIs/SLOs]: https://git.k8s.io/community/sig-scalability/slos/slos.md#kubernetes-slisslos
885889
-->
890+
No. The computational overhead introduced in API server, scheduler and kubelet due to additional validation for pod level resource specifications should be negligible. Thorough testing and monitoring will ensure the SLIs/SLOs are not impacted.
886891

887892
###### Will enabling / using this feature result in non-negligible increase of resource usage (CPU, RAM, disk, IO, ...) in any components?
888893
No

0 commit comments

Comments
 (0)