@@ -137,7 +137,7 @@ const (
137
137
housekeepingPeriod = time .Second * 2
138
138
139
139
// Period for performing eviction monitoring.
140
- // TODO ensure this is in sync with internal cadvisor housekeeping.
140
+ // ensure this is kept in sync with internal cadvisor housekeeping.
141
141
evictionMonitoringPeriod = time .Second * 10
142
142
143
143
// The path in containers' filesystems where the hosts file is mounted.
@@ -442,9 +442,7 @@ func NewMainKubelet(kubeCfg *kubeletconfiginternal.KubeletConfiguration,
442
442
}
443
443
nodeLister := corelisters .NewNodeLister (nodeIndexer )
444
444
445
- // TODO: get the real node object of ourself,
446
- // and use the real node name and UID.
447
- // TODO: what is namespace for node?
445
+ // construct a node reference used for events
448
446
nodeRef := & v1.ObjectReference {
449
447
Kind : "Node" ,
450
448
Name : string (nodeName ),
@@ -958,8 +956,6 @@ type Kubelet struct {
958
956
streamingRuntime kubecontainer.StreamingRuntime
959
957
960
958
// Container runtime service (needed by container runtime Start()).
961
- // TODO(CD): try to make this available without holding a reference in this
962
- // struct. For example, by adding a getter to generic runtime.
963
959
runtimeService internalapi.RuntimeService
964
960
965
961
// reasonCache caches the failure reason of the last creation of all containers, which is
@@ -1079,7 +1075,6 @@ type Kubelet struct {
1079
1075
// maintains Node.Spec.Unschedulable value from previous run of tryUpdateNodeStatus()
1080
1076
lastNodeUnschedulable bool
1081
1077
1082
- // TODO: think about moving this to be centralized in PodWorkers in follow-on.
1083
1078
// the list of handlers to call during pod admission.
1084
1079
admitHandlers lifecycle.PodAdmitHandlers
1085
1080
@@ -1337,7 +1332,6 @@ func (kl *Kubelet) initializeModules() error {
1337
1332
func (kl * Kubelet ) initializeRuntimeDependentModules () {
1338
1333
if err := kl .cadvisor .Start (); err != nil {
1339
1334
// Fail kubelet and rely on the babysitter to retry starting kubelet.
1340
- // TODO(random-liu): Add backoff logic in the babysitter
1341
1335
klog .Fatalf ("Failed to start cAdvisor %v" , err )
1342
1336
}
1343
1337
@@ -1744,7 +1738,6 @@ func (kl *Kubelet) deletePod(pod *v1.Pod) error {
1744
1738
podPair := kubecontainer.PodPair {APIPod : pod , RunningPod : & runningPod }
1745
1739
1746
1740
kl .podKiller .KillPod (& podPair )
1747
- // TODO: delete the mirror pod here?
1748
1741
1749
1742
// We leave the volume/directory cleanup to the periodic cleanup routine.
1750
1743
return nil
@@ -2070,8 +2063,6 @@ func (kl *Kubelet) HandlePodUpdates(pods []*v1.Pod) {
2070
2063
kl .handleMirrorPod (pod , start )
2071
2064
continue
2072
2065
}
2073
- // TODO: Evaluate if we need to validate and reject updates.
2074
-
2075
2066
mirrorPod , _ := kl .podManager .GetMirrorPodByPod (pod )
2076
2067
kl .dispatchWork (pod , kubetypes .SyncPodUpdate , mirrorPod , start )
2077
2068
}
@@ -2160,7 +2151,6 @@ func (kl *Kubelet) updateRuntimeUp() {
2160
2151
return
2161
2152
}
2162
2153
// Periodically log the whole runtime status for debugging.
2163
- // TODO(random-liu): Consider to send node event when optional
2164
2154
// condition is unmet.
2165
2155
klog .V (4 ).Infof ("Container runtime status: %v" , s )
2166
2156
networkReady := s .GetRuntimeCondition (kubecontainer .NetworkReady )
0 commit comments