@@ -89,7 +89,7 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
89
89
e2elog .Logf ("unable to dump pods: %v" , err )
90
90
}
91
91
err = clearDaemonSetNodeLabels (f .ClientSet )
92
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
92
+ framework . ExpectNoError (err )
93
93
})
94
94
95
95
f = framework .NewDefaultFramework ("daemonsets" )
@@ -106,12 +106,12 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
106
106
c = f .ClientSet
107
107
108
108
updatedNS , err := updateNamespaceAnnotations (c , ns )
109
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
109
+ framework . ExpectNoError (err )
110
110
111
111
ns = updatedNS .Name
112
112
113
113
err = clearDaemonSetNodeLabels (c )
114
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
114
+ framework . ExpectNoError (err )
115
115
})
116
116
117
117
/*
@@ -124,19 +124,19 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
124
124
125
125
ginkgo .By (fmt .Sprintf ("Creating simple DaemonSet %q" , dsName ))
126
126
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (newDaemonSet (dsName , image , label ))
127
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
127
+ framework . ExpectNoError (err )
128
128
129
129
ginkgo .By ("Check that daemon pods launch on every node of the cluster." )
130
130
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
131
131
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
132
132
err = checkDaemonStatus (f , dsName )
133
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
133
+ framework . ExpectNoError (err )
134
134
135
135
ginkgo .By ("Stop a daemon pod, check that the daemon pod is revived." )
136
136
podList := listDaemonPods (c , ns , label )
137
137
pod := podList .Items [0 ]
138
138
err = c .CoreV1 ().Pods (ns ).Delete (pod .Name , nil )
139
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
139
+ framework . ExpectNoError (err )
140
140
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
141
141
framework .ExpectNoError (err , "error waiting for daemon pod to revive" )
142
142
})
@@ -153,7 +153,7 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
153
153
ds := newDaemonSet (dsName , image , complexLabel )
154
154
ds .Spec .Template .Spec .NodeSelector = nodeSelector
155
155
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (ds )
156
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
156
+ framework . ExpectNoError (err )
157
157
158
158
ginkgo .By ("Initially, daemon pods should not be running on any nodes." )
159
159
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))
@@ -169,14 +169,14 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
169
169
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkDaemonPodOnNodes (f , ds , []string {newNode .Name }))
170
170
framework .ExpectNoError (err , "error waiting for daemon pods to be running on new nodes" )
171
171
err = checkDaemonStatus (f , dsName )
172
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
172
+ framework . ExpectNoError (err )
173
173
174
174
ginkgo .By ("Update the node label to green, and wait for daemons to be unscheduled" )
175
175
nodeSelector [daemonsetColorLabel ] = "green"
176
176
greenNode , err := setDaemonSetNodeLabels (c , nodeList .Items [0 ].Name , nodeSelector )
177
177
framework .ExpectNoError (err , "error removing labels on node" )
178
- gomega . Expect ( wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))).
179
- NotTo ( gomega . HaveOccurred () , "error waiting for daemon pod to not be running on nodes" )
178
+ err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))
179
+ framework . ExpectNoError ( err , "error waiting for daemon pod to not be running on nodes" )
180
180
181
181
ginkgo .By ("Update DaemonSet node selector to green, and change its update strategy to RollingUpdate" )
182
182
patch := fmt .Sprintf (`{"spec":{"template":{"spec":{"nodeSelector":{"%s":"%s"}}},"updateStrategy":{"type":"RollingUpdate"}}}` ,
@@ -188,7 +188,7 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
188
188
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkDaemonPodOnNodes (f , ds , []string {greenNode .Name }))
189
189
framework .ExpectNoError (err , "error waiting for daemon pods to be running on new nodes" )
190
190
err = checkDaemonStatus (f , dsName )
191
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
191
+ framework . ExpectNoError (err )
192
192
})
193
193
194
194
// We defer adding this test to conformance pending the disposition of moving DaemonSet scheduling logic to the
@@ -216,7 +216,7 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
216
216
},
217
217
}
218
218
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (ds )
219
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
219
+ framework . ExpectNoError (err )
220
220
221
221
ginkgo .By ("Initially, daemon pods should not be running on any nodes." )
222
222
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))
@@ -232,13 +232,13 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
232
232
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkDaemonPodOnNodes (f , ds , []string {newNode .Name }))
233
233
framework .ExpectNoError (err , "error waiting for daemon pods to be running on new nodes" )
234
234
err = checkDaemonStatus (f , dsName )
235
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
235
+ framework . ExpectNoError (err )
236
236
237
237
ginkgo .By ("Remove the node label and wait for daemons to be unscheduled" )
238
238
_ , err = setDaemonSetNodeLabels (c , nodeList .Items [0 ].Name , map [string ]string {})
239
239
framework .ExpectNoError (err , "error removing labels on node" )
240
- gomega . Expect ( wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))).
241
- NotTo ( gomega . HaveOccurred () , "error waiting for daemon pod to not be running on nodes" )
240
+ err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnNoNodes (f , ds ))
241
+ framework . ExpectNoError ( err , "error waiting for daemon pod to not be running on nodes" )
242
242
})
243
243
244
244
/*
@@ -250,13 +250,13 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
250
250
251
251
ginkgo .By (fmt .Sprintf ("Creating a simple DaemonSet %q" , dsName ))
252
252
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (newDaemonSet (dsName , image , label ))
253
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
253
+ framework . ExpectNoError (err )
254
254
255
255
ginkgo .By ("Check that daemon pods launch on every node of the cluster." )
256
256
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
257
257
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
258
258
err = checkDaemonStatus (f , dsName )
259
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
259
+ framework . ExpectNoError (err )
260
260
261
261
ginkgo .By ("Set a daemon pod's phase to 'Failed', check that the daemon pod is revived." )
262
262
podList := listDaemonPods (c , ns , label )
@@ -282,15 +282,15 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
282
282
ds := newDaemonSet (dsName , image , label )
283
283
ds .Spec .UpdateStrategy = apps.DaemonSetUpdateStrategy {Type : apps .OnDeleteDaemonSetStrategyType }
284
284
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (ds )
285
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
285
+ framework . ExpectNoError (err )
286
286
287
287
ginkgo .By ("Check that daemon pods launch on every node of the cluster." )
288
288
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
289
289
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
290
290
291
291
// Check history and labels
292
292
ds , err = c .AppsV1 ().DaemonSets (ns ).Get (ds .Name , metav1.GetOptions {})
293
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
293
+ framework . ExpectNoError (err )
294
294
waitForHistoryCreated (c , ns , label , 1 )
295
295
first := curHistory (listDaemonHistories (c , ns , label ), ds )
296
296
firstHash := first .Labels [apps .DefaultDaemonSetUniqueLabelKey ]
@@ -300,19 +300,19 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
300
300
ginkgo .By ("Update daemon pods image." )
301
301
patch := getDaemonSetImagePatch (ds .Spec .Template .Spec .Containers [0 ].Name , RedisImage )
302
302
ds , err = c .AppsV1 ().DaemonSets (ns ).Patch (dsName , types .StrategicMergePatchType , []byte (patch ))
303
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
303
+ framework . ExpectNoError (err )
304
304
305
305
ginkgo .By ("Check that daemon pods images aren't updated." )
306
306
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkDaemonPodsImageAndAvailability (c , ds , image , 0 ))
307
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
307
+ framework . ExpectNoError (err )
308
308
309
309
ginkgo .By ("Check that daemon pods are still running on every node of the cluster." )
310
310
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
311
311
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
312
312
313
313
// Check history and labels
314
314
ds , err = c .AppsV1 ().DaemonSets (ns ).Get (ds .Name , metav1.GetOptions {})
315
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
315
+ framework . ExpectNoError (err )
316
316
waitForHistoryCreated (c , ns , label , 2 )
317
317
cur := curHistory (listDaemonHistories (c , ns , label ), ds )
318
318
gomega .Expect (cur .Revision ).To (gomega .Equal (int64 (2 )))
@@ -331,15 +331,15 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
331
331
ds := newDaemonSet (dsName , image , label )
332
332
ds .Spec .UpdateStrategy = apps.DaemonSetUpdateStrategy {Type : apps .RollingUpdateDaemonSetStrategyType }
333
333
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (ds )
334
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
334
+ framework . ExpectNoError (err )
335
335
336
336
ginkgo .By ("Check that daemon pods launch on every node of the cluster." )
337
337
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
338
338
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
339
339
340
340
// Check history and labels
341
341
ds , err = c .AppsV1 ().DaemonSets (ns ).Get (ds .Name , metav1.GetOptions {})
342
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
342
+ framework . ExpectNoError (err )
343
343
waitForHistoryCreated (c , ns , label , 1 )
344
344
cur := curHistory (listDaemonHistories (c , ns , label ), ds )
345
345
hash := cur .Labels [apps .DefaultDaemonSetUniqueLabelKey ]
@@ -349,26 +349,26 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
349
349
ginkgo .By ("Update daemon pods image." )
350
350
patch := getDaemonSetImagePatch (ds .Spec .Template .Spec .Containers [0 ].Name , RedisImage )
351
351
ds , err = c .AppsV1 ().DaemonSets (ns ).Patch (dsName , types .StrategicMergePatchType , []byte (patch ))
352
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
352
+ framework . ExpectNoError (err )
353
353
354
354
// Time to complete the rolling upgrade is proportional to the number of nodes in the cluster.
355
355
// Get the number of nodes, and set the timeout appropriately.
356
356
nodes , err := c .CoreV1 ().Nodes ().List (metav1.ListOptions {})
357
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
357
+ framework . ExpectNoError (err )
358
358
nodeCount := len (nodes .Items )
359
359
retryTimeout := dsRetryTimeout + time .Duration (nodeCount * 30 )* time .Second
360
360
361
361
ginkgo .By ("Check that daemon pods images are updated." )
362
362
err = wait .PollImmediate (dsRetryPeriod , retryTimeout , checkDaemonPodsImageAndAvailability (c , ds , RedisImage , 1 ))
363
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
363
+ framework . ExpectNoError (err )
364
364
365
365
ginkgo .By ("Check that daemon pods are still running on every node of the cluster." )
366
366
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
367
367
framework .ExpectNoError (err , "error waiting for daemon pod to start" )
368
368
369
369
// Check history and labels
370
370
ds , err = c .AppsV1 ().DaemonSets (ns ).Get (ds .Name , metav1.GetOptions {})
371
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
371
+ framework . ExpectNoError (err )
372
372
waitForHistoryCreated (c , ns , label , 2 )
373
373
cur = curHistory (listDaemonHistories (c , ns , label ), ds )
374
374
hash = cur .Labels [apps .DefaultDaemonSetUniqueLabelKey ]
@@ -389,7 +389,7 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
389
389
ds := newDaemonSet (dsName , image , label )
390
390
ds .Spec .UpdateStrategy = apps.DaemonSetUpdateStrategy {Type : apps .RollingUpdateDaemonSetStrategyType }
391
391
ds , err := c .AppsV1 ().DaemonSets (ns ).Create (ds )
392
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
392
+ framework . ExpectNoError (err )
393
393
394
394
e2elog .Logf ("Check that daemon pods launch on every node of the cluster" )
395
395
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkRunningOnAllNodes (f , ds ))
@@ -401,11 +401,11 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
401
401
newDS , err := framework .UpdateDaemonSetWithRetries (c , ns , ds .Name , func (update * apps.DaemonSet ) {
402
402
update .Spec .Template .Spec .Containers [0 ].Image = newImage
403
403
})
404
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
404
+ framework . ExpectNoError (err )
405
405
406
406
// Make sure we're in the middle of a rollout
407
407
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkAtLeastOneNewPod (c , ns , label , newImage ))
408
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
408
+ framework . ExpectNoError (err )
409
409
410
410
pods := listDaemonPods (c , ns , label )
411
411
var existingPods , newPods []* v1.Pod
@@ -433,11 +433,11 @@ var _ = SIGDescribe("Daemon set [Serial]", func() {
433
433
rollbackDS , err := framework .UpdateDaemonSetWithRetries (c , ns , ds .Name , func (update * apps.DaemonSet ) {
434
434
update .Spec .Template .Spec .Containers [0 ].Image = image
435
435
})
436
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
436
+ framework . ExpectNoError (err )
437
437
438
438
e2elog .Logf ("Make sure DaemonSet rollback is complete" )
439
439
err = wait .PollImmediate (dsRetryPeriod , dsRetryTimeout , checkDaemonPodsImageAndAvailability (c , rollbackDS , image , 1 ))
440
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
440
+ framework . ExpectNoError (err )
441
441
442
442
// After rollback is done, compare current pods with previous old pods during rollout, to make sure they're not restarted
443
443
pods = listDaemonPods (c , ns , label )
@@ -487,7 +487,7 @@ func listDaemonPods(c clientset.Interface, ns string, label map[string]string) *
487
487
selector := labels .Set (label ).AsSelector ()
488
488
options := metav1.ListOptions {LabelSelector : selector .String ()}
489
489
podList , err := c .CoreV1 ().Pods (ns ).List (options )
490
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
490
+ framework . ExpectNoError (err )
491
491
gomega .Expect (len (podList .Items )).To (gomega .BeNumerically (">" , 0 ))
492
492
return podList
493
493
}
@@ -748,7 +748,7 @@ func listDaemonHistories(c clientset.Interface, ns string, label map[string]stri
748
748
selector := labels .Set (label ).AsSelector ()
749
749
options := metav1.ListOptions {LabelSelector : selector .String ()}
750
750
historyList , err := c .AppsV1 ().ControllerRevisions (ns ).List (options )
751
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
751
+ framework . ExpectNoError (err )
752
752
gomega .Expect (len (historyList .Items )).To (gomega .BeNumerically (">" , 0 ))
753
753
return historyList
754
754
}
@@ -761,7 +761,7 @@ func curHistory(historyList *apps.ControllerRevisionList, ds *apps.DaemonSet) *a
761
761
// Every history should have the hash label
762
762
gomega .Expect (len (history .Labels [apps .DefaultDaemonSetUniqueLabelKey ])).To (gomega .BeNumerically (">" , 0 ))
763
763
match , err := daemon .Match (ds , history )
764
- gomega . Expect (err ). NotTo ( gomega . HaveOccurred () )
764
+ framework . ExpectNoError (err )
765
765
if match {
766
766
curHistory = history
767
767
foundCurHistories ++
0 commit comments