@@ -277,8 +277,8 @@ func TestFullScrapeCycle(t *testing.T) {
277
277
# TYPE kube_pod_status_scheduled_time gauge
278
278
# TYPE kube_pod_status_unschedulable gauge
279
279
kube_pod_annotations{namespace="default",pod="pod0",uid="abc-0"} 1
280
- kube_pod_container_info{namespace="default",pod="pod0",uid="abc-0",container="container2 ",image_spec="k8s.gcr.io/hyperkube2_spec",image="k8s.gcr.io/hyperkube2",image_id="docker://sha256:bbb",container_id="docker://cd456"} 1
281
- kube_pod_container_info{namespace="default",pod="pod0",uid="abc-0",container="container3 ",image_spec="k8s.gcr.io/hyperkube3_spec",image="k8s.gcr.io/hyperkube3",image_id="docker://sha256:ccc",container_id="docker://ef789"} 1
280
+ kube_pod_container_info{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 ",image_spec="k8s.gcr.io/hyperkube2_spec",image="k8s.gcr.io/hyperkube2",image_id="docker://sha256:bbb",container_id="docker://cd456"} 1
281
+ kube_pod_container_info{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 ",image_spec="k8s.gcr.io/hyperkube3_spec",image="k8s.gcr.io/hyperkube3",image_id="docker://sha256:ccc",container_id="docker://ef789"} 1
282
282
kube_pod_container_resource_limits{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1",node="node1",resource="cpu",unit="core"} 0.2
283
283
kube_pod_container_resource_limits{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1",node="node1",resource="ephemeral_storage",unit="byte"} 3e+08
284
284
kube_pod_container_resource_limits{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1",node="node1",resource="memory",unit="byte"} 1e+08
@@ -293,18 +293,18 @@ kube_pod_container_resource_requests{namespace="default",pod="pod0",uid="abc-0",
293
293
kube_pod_container_resource_requests{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1",node="node1",resource="storage",unit="byte"} 4e+08
294
294
kube_pod_container_resource_requests{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2",node="node1",resource="cpu",unit="core"} 0.3
295
295
kube_pod_container_resource_requests{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2",node="node1",resource="memory",unit="byte"} 2e+08
296
- kube_pod_container_status_last_terminated_reason{namespace="default",pod="pod0",uid="abc-0",container="container2 ",reason="OOMKilled"} 1
297
- kube_pod_container_status_ready{namespace="default",pod="pod0",uid="abc-0",container="container2 "} 0
298
- kube_pod_container_status_ready{namespace="default",pod="pod0",uid="abc-0",container="container3 "} 0
299
- kube_pod_container_status_restarts_total{namespace="default",pod="pod0",uid="abc-0",container="container2 "} 0
300
- kube_pod_container_status_restarts_total{namespace="default",pod="pod0",uid="abc-0",container="container3 "} 0
301
- kube_pod_container_status_running{namespace="default",pod="pod0",uid="abc-0",container="container2 "} 0
302
- kube_pod_container_status_running{namespace="default",pod="pod0",uid="abc-0",container="container3 "} 0
303
- kube_pod_container_status_terminated{namespace="default",pod="pod0",uid="abc-0",container="container2 "} 0
304
- kube_pod_container_status_terminated{namespace="default",pod="pod0",uid="abc-0",container="container3 "} 0
305
- kube_pod_container_status_waiting_reason{namespace="default",pod="pod0",uid="abc-0",container="container2 ",reason="CrashLoopBackOff"} 1
306
- kube_pod_container_status_waiting{namespace="default",pod="pod0",uid="abc-0",container="container2 "} 1
307
- kube_pod_container_status_waiting{namespace="default",pod="pod0",uid="abc-0",container="container3 "} 0
296
+ kube_pod_container_status_last_terminated_reason{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 ",reason="OOMKilled"} 1
297
+ kube_pod_container_status_ready{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 "} 0
298
+ kube_pod_container_status_ready{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 "} 0
299
+ kube_pod_container_status_restarts_total{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 "} 0
300
+ kube_pod_container_status_restarts_total{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 "} 0
301
+ kube_pod_container_status_running{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 "} 0
302
+ kube_pod_container_status_running{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 "} 0
303
+ kube_pod_container_status_terminated{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 "} 0
304
+ kube_pod_container_status_terminated{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 "} 0
305
+ kube_pod_container_status_waiting_reason{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 ",reason="CrashLoopBackOff"} 1
306
+ kube_pod_container_status_waiting{namespace="default",pod="pod0",uid="abc-0",container="pod1_con1 "} 1
307
+ kube_pod_container_status_waiting{namespace="default",pod="pod0",uid="abc-0",container="pod1_con2 "} 0
308
308
kube_pod_created{namespace="default",pod="pod0",uid="abc-0"} 1.5e+09
309
309
kube_pod_info{namespace="default",pod="pod0",uid="abc-0",host_ip="1.1.1.1",pod_ip="1.2.3.4",node="node1",created_by_kind="<none>",created_by_name="<none>",priority_class="",host_network="false"} 1
310
310
kube_pod_labels{namespace="default",pod="pod0",uid="abc-0"} 1
@@ -779,7 +779,7 @@ func pod(client *fake.Clientset, index int) error {
779
779
Phase : v1 .PodRunning ,
780
780
ContainerStatuses : []v1.ContainerStatus {
781
781
{
782
- Name : "container2 " ,
782
+ Name : "pod1_con1 " ,
783
783
Image : "k8s.gcr.io/hyperkube2" ,
784
784
ImageID : "docker://sha256:bbb" ,
785
785
ContainerID : "docker://cd456" ,
@@ -795,7 +795,7 @@ func pod(client *fake.Clientset, index int) error {
795
795
},
796
796
},
797
797
{
798
- Name : "container3 " ,
798
+ Name : "pod1_con2 " ,
799
799
Image : "k8s.gcr.io/hyperkube3" ,
800
800
ImageID : "docker://sha256:ccc" ,
801
801
ContainerID : "docker://ef789" ,
0 commit comments