You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Unfortunately, I really struggle in creating even the most simple setups and I'm a little bit frustrated :( ..
I'm trying to move from my Docker root server (128GB RAM), which ran smoothly about ~4 years to kubernetes, because
There are two heavy services (~each using ~50GB mysql DB RAM)
I'm using the infrastructure for CI/CD - which creates a lot of spikes
because of increasing usage of heavy services (since the Twitter-chaos ;-) ), the whole root server recently slowed down and I REALLY need to change something because of I/O issues
So I went to this project some month ago (maybe some know I added calico at first place) and stopped it because of lack of time.
Now I'm back and try to setup my cluser as a new home.
My services:
private usage (~6-7 users)
Nextcloud with lot of space (=> longhorn support)
Email & Groupware setup (=> egress service for smtp otgoing)
Gitea instance
Keycloak / SSO
small public services
two wordpress instances
personal Friendica instance
large public services
Two Friendica setups (mysql, php-fpm, nginx, own asset service, cronjob service) with ~6-700 users
CI/CD for open source projects (integration tests with mysql, redis, ... instances)
After after successful terraform apply, the whole setup starts crashing ..
log of one controller:
Jul 26 20:21:52 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:52.589Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487, vote: 6862d5ef6fbce322] cast MsgPreVote for 6862d5ef6fbce322 [logterm: 14, index: 26487] at term 14"}Jul 26 20:21:53 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:53.380Z","logger":"etcd-client","caller":"[email protected]/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0009c8540/127.0.0.1:2379","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"}Jul 26 20:21:53 k3s-control-plane-hel1-rkf k3s[9174]: time="2023-07-26T20:21:53Z" level=error msg="Failed to check local etcd status for learner management: context deadline exceeded"Jul 26 20:21:53 k3s-control-plane-hel1-rkf k3s[9174]: time="2023-07-26T20:21:53Z" level=info msg="Waiting to retrieve kube-proxy configuration; server is not ready: https://127.0.0.1:6443/v1-k3s/readyz: 500 Internal Server Error"Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.217Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 is starting a new election at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 became pre-candidate at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 received MsgPreVoteResp from b8721dfb5f0a2306 at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487] sent MsgPreVote request to 37c27cde23b06db4 at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487] sent MsgPreVote request to 6862d5ef6fbce322 at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487] sent MsgPreVote request to 6c89d2f8e23fc045 at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.218Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487] sent MsgPreVote request to 8a38914f73defeab at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.244Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 received MsgPreVoteResp from 6862d5ef6fbce322 at term 14"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:21:56.244Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 has received 2 MsgPreVoteResp votes and 0 vote rejections"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_SNAPSHOT","remote-peer-id":"8a38914f73defeab","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_RAFT_MESSAGE","remote-peer-id":"37c27cde23b06db4","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_SNAPSHOT","remote-peer-id":"37c27cde23b06db4","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_SNAPSHOT","remote-peer-id":"6c89d2f8e23fc045","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_RAFT_MESSAGE","remote-peer-id":"8a38914f73defeab","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:56 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:21:56.296Z","caller":"rafthttp/probing_status.go:68","msg":"prober detected unhealthy status","round-tripper-name":"ROUND_TRIPPER_RAFT_MESSAGE","remote-peer-id":"6c89d2f8e23fc045","rtt":"0s","error":"net/http: TLS handshake timeout"}Jul 26 20:21:58 k3s-control-plane-hel1-rkf k3s[9174]: time="2023-07-26T20:21:58Z" level=info msg="Waiting to retrieve kube-proxy configuration; server is not ready: https://127.0.0.1:6443/v1-k3s/readyz: 500 Internal Server Error"Jul 26 20:22:00 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:22:00.088Z","logger":"raft","caller":"etcdserver/zap_raft.go:77","msg":"b8721dfb5f0a2306 [logterm: 14, index: 26487, vote: 6862d5ef6fbce322] cast MsgPreVote for 6862d5ef6fbce322 [logterm: 14, index: 26487] at term 14"}Jul 26 20:22:01 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"warn","ts":"2023-07-26T20:22:01.002Z","logger":"etcd-client","caller":"[email protected]/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"etcd-endpoints://0xc0009c8540/127.0.0.1:2379","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = latest balancer error: last connection error: connection error: desc = \"transport: authentication handshake failed: context deadline exceeded\""}Jul 26 20:22:01 k3s-control-plane-hel1-rkf k3s[9174]: {"level":"info","ts":"2023-07-26T20:22:01.003Z","logger":"etcd-client","caller":"[email protected]/client.go:210","msg":"Auto sync endpoints failed.","error":"context deadline exceeded"}
Other control-plane nodes suddenly stops being available per SSH or per kubectl and showing OOM errors, like:
[ 1670.92471][ T6426] out of memory: Killed process 4046 (kured) total-vm:751132kb, anon-rss: 7040kB, file-rss:128kB, shmem-rss:0kB, UID:0 pgtables:176kB oom_score_adj:1000
Do I really need more resources just for the naked cluster without any application on it .. This seems like it's not properly working...
I would REALLY love to use this setup, the whole feature-catalog you show is exactly what I think, what would be necessary to move to the next level with my environment :)
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all,
Unfortunately, I really struggle in creating even the most simple setups and I'm a little bit frustrated :( ..
I'm trying to move from my Docker root server (128GB RAM), which ran smoothly about ~4 years to kubernetes, because
So I went to this project some month ago (maybe some know I added calico at first place) and stopped it because of lack of time.
Now I'm back and try to setup my cluser as a new home.
My services:
private usage (~6-7 users)
small public services
large public services
My current
kube.tf
:After after successful terraform apply, the whole setup starts crashing ..
log of one controller:
Other control-plane nodes suddenly stops being available per SSH or per kubectl and showing OOM errors, like:
[ 1670.92471][ T6426] out of memory: Killed process 4046 (kured) total-vm:751132kb, anon-rss: 7040kB, file-rss:128kB, shmem-rss:0kB, UID:0 pgtables:176kB oom_score_adj:1000
Do I really need more resources just for the naked cluster without any application on it .. This seems like it's not properly working...
I would REALLY love to use this setup, the whole feature-catalog you show is exactly what I think, what would be necessary to move to the next level with my environment :)
HELP!!! :-)
Beta Was this translation helpful? Give feedback.
All reactions