[Bug]: KubernetesClientException in ZooKeeper StrimziPodSet #8533
Replies: 7 comments 10 replies
-
I think you would need to share the full log. This looks like a problem with your Kubernetes environment - e.g. due to some misbehaving admission controllers etc. But hard to say without the full log. |
Beta Was this translation helpful? Give feedback.
-
Possibly related to:
Regarding potential issues with the k8s environment - we have 10+ other applications deployed and haven't seen any issues with k8s itself thus far. |
Beta Was this translation helpful? Give feedback.
-
Logs from
|
Beta Was this translation helpful? Give feedback.
-
@scholzj thanks for your response - do you have any tips on how to debug further and where to look?
|
Beta Was this translation helpful? Give feedback.
-
Found it. We have verified this timeout is directly caused by the Vertical Pod Autoscaler application's With the With the $ kubectl get mutatingwebhookconfiguration.admissionregistration.k8s.io/vpa-webhook-config -o yaml
apiVersion: admissionregistration.k8s.io/v1
kind: MutatingWebhookConfiguration
metadata:
creationTimestamp: "2023-05-22T18:08:33Z"
generation: 1
name: vpa-webhook-config
resourceVersion: "6131859"
uid: 77208c43-5c2e-426d-b94a-eaa935b1a34b
webhooks:
- admissionReviewVersions:
- v1
clientConfig:
caBundle: LS0tLS1CRUdJTiBDRVJU...
service:
name: vpa-webhook
namespace: kube-system
port: 443
failurePolicy: Ignore
matchPolicy: Equivalent
name: vpa.k8s.io
namespaceSelector: {}
objectSelector: {}
reinvocationPolicy: Never
rules:
- apiGroups:
- ""
apiVersions:
- v1
operations:
- CREATE
resources:
- pods
scope: '*'
- apiGroups:
- autoscaling.k8s.io
apiVersions:
- '*'
operations:
- CREATE
- UPDATE
resources:
- verticalpodautoscalers
scope: '*'
sideEffects: None
timeoutSeconds: 30 |
Beta Was this translation helpful? Give feedback.
-
I have opened kubernetes/autoscaler#5792 with the VPA app in question here. |
Beta Was this translation helpful? Give feedback.
-
Setting the following for the strimzi helm chart seems to be working (Pods now coming up): logLevel: DEBUG
operationTimeoutMs: 600000
extraEnvs:
- name: KUBERNETES_REQUEST_TIMEOUT
value: '60000' plus revert VPA to its standard Frustratingly I cannot get VPA to actually log what it is mad about, at any |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Bug Description
Bug Description
On the initial installation of Kafka using strimzi-kafka-operator, we are seeing an
HttpTimeoutException: request timed out
in thedeployment/strimzi-kafka-operator
logs and aReason=KubernetesClientException
on the ZooKeeperStrimziPodSet
.Steps to reproduce
kafka-system
namespace with followingvalues.yaml
:elastic
namespace:Expected behavior
Kafka cluster comes up cleanly, or strimzi logs give more of an indication of what specifically is failing.
Strimzi version
0.35.0
Kubernetes version
1.26
Installation method
Helm
Infrastructure
Amazon EKS
Configuration files and logs
Additional context
We have all TCP traffic allowed between EKS worker nodes.
Beta Was this translation helpful? Give feedback.
All reactions