You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+55Lines changed: 55 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -134,6 +134,61 @@ All we need to is to run these commands:
134
134
135
135
Now all containers should be up and running.
136
136
137
+
### Upgrading to 1.15+
138
+
139
+
In this version the terraform providers were upgraded to newer versions and this introduces
140
+
role name changes and a lot of other things. This means that after the upgrade, you can expect
141
+
issues with certain kube-system pods in a crashloop.
142
+
143
+
The reason this happens is that the role names have changed that infra creates. They're using a
144
+
prefix and a suffix now.
145
+
146
+
AWS authenticates the service accounts for certain kube-system pods like aws-loadbalancer-controller,
147
+
but after this change that role mapping breaks.
148
+
149
+
There are ways to fix that manually:
150
+
* Apply the application again after applying the infra. This should fix the role names for two pods.
151
+
* Go to the service account of the aws-load-balancer-controller pod.
152
+
* The service account has a forward mapping to the role ARN they need to assume on the cloud in the annotations
153
+
* Update that annotation.
154
+
155
+
Example:
156
+
157
+
```yaml
158
+
apiVersion: v1
159
+
automountServiceAccountToken: true
160
+
kind: ServiceAccount
161
+
metadata:
162
+
annotations:
163
+
eks.amazonaws.com/role-arn: arn:aws:iam::1234567889:role/datafold-lb-controller-2025082013431968900000001 <-- This role ARN should correspond to the role.
Check kubernetes for any failing pods in the kube-system namespace, possibly these need updating in the same
172
+
way if the pods continue in the crashloop backoff phase.
173
+
174
+
* In the newest version of Amazon Linux 3, Datadog cannot determine the local hostname, which it needs for tagging. Updating to the most recent datadog operator solves this issue:
0 commit comments