@@ -31,6 +31,7 @@ import (
31
31
var (
32
32
depl1Replicas int32 = 200
33
33
depl2Replicas int32 = 5
34
+ depl3Replicas int32 = 10
34
35
35
36
depl1MaxUnavailable = intstr .FromInt (10 )
36
37
depl2MaxUnavailable = intstr .FromString ("25%" )
@@ -73,6 +74,8 @@ func TestDeploymentStore(t *testing.T) {
73
74
# TYPE kube_deployment_spec_strategy_rollingupdate_max_surge gauge
74
75
# HELP kube_deployment_labels [STABLE] Kubernetes labels converted to Prometheus labels.
75
76
# TYPE kube_deployment_labels gauge
77
+ # HELP kube_deployment_deletion_timestamp Unix deletion timestamp.
78
+ # TYPE kube_deployment_deletion_timestamp gauge
76
79
`
77
80
cases := []generateMetricsTestCase {
78
81
{
@@ -191,8 +194,30 @@ func TestDeploymentStore(t *testing.T) {
191
194
kube_deployment_status_condition{deployment="depl2",namespace="ns2",condition="ReplicaFailure",status="unknown"} 0
192
195
` ,
193
196
},
197
+ {
198
+ Obj : & v1.Deployment {
199
+ ObjectMeta : metav1.ObjectMeta {
200
+ Name : "deployment-terminating" ,
201
+ Namespace : "ns3" ,
202
+ CreationTimestamp : metav1.Time {Time : time .Unix (1600000000 , 0 )},
203
+ DeletionTimestamp : & metav1.Time {Time : time .Unix (1800000000 , 0 )},
204
+ Labels : map [string ]string {
205
+ "app" : "example3" ,
206
+ },
207
+ Generation : 22 ,
208
+ },
209
+ Spec : v1.DeploymentSpec {
210
+ Paused : true ,
211
+ Replicas : & depl3Replicas ,
212
+ },
213
+ },
214
+ Want : `
215
+ # HELP kube_deployment_deletion_timestamp Unix deletion timestamp.
216
+ # TYPE kube_deployment_deletion_timestamp gauge
217
+ kube_deployment_deletion_timestamp{deployment="deployment-terminating",namespace="ns3"} 1.8e+09` ,
218
+ MetricNames : []string {"kube_deployment_deletion_timestamp" },
219
+ },
194
220
}
195
-
196
221
for i , c := range cases {
197
222
c .Func = generator .ComposeMetricGenFuncs (deploymentMetricFamilies (c .AllowAnnotationsList , nil ))
198
223
c .Headers = generator .ExtractMetricFamilyHeaders (deploymentMetricFamilies (c .AllowAnnotationsList , nil ))
0 commit comments