@@ -22,17 +22,14 @@ import (
22
22
"errors"
23
23
"math"
24
24
"strconv"
25
- "time"
26
25
27
26
"go.uber.org/zap"
28
27
v1 "k8s.io/api/core/v1"
29
28
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
30
29
"k8s.io/apimachinery/pkg/labels"
31
30
"k8s.io/apimachinery/pkg/types"
32
31
"k8s.io/apimachinery/pkg/util/sets"
33
- "k8s.io/apimachinery/pkg/util/wait"
34
32
corev1 "k8s.io/client-go/listers/core/v1"
35
-
36
33
"knative.dev/pkg/logging"
37
34
38
35
"knative.dev/eventing/pkg/scheduler"
@@ -42,7 +39,7 @@ type StateAccessor interface {
42
39
// State returns the current state (snapshot) about placed vpods
43
40
// Take into account reserved vreplicas and update `reserved` to reflect
44
41
// the current state.
45
- State (reserved map [types.NamespacedName ]map [string ]int32 ) (* State , error )
42
+ State (ctx context. Context , reserved map [types.NamespacedName ]map [string ]int32 ) (* State , error )
46
43
}
47
44
48
45
// state provides information about the current scheduling of all vpods
@@ -152,8 +149,6 @@ func (s *State) IsSchedulablePod(ordinal int32) bool {
152
149
153
150
// stateBuilder reconstruct the state from scratch, by listing vpods
154
151
type stateBuilder struct {
155
- ctx context.Context
156
- logger * zap.SugaredLogger
157
152
vpodLister scheduler.VPodLister
158
153
capacity int32
159
154
schedulerPolicy scheduler.SchedulerPolicyType
@@ -166,11 +161,9 @@ type stateBuilder struct {
166
161
}
167
162
168
163
// NewStateBuilder returns a StateAccessor recreating the state from scratch each time it is requested
169
- func NewStateBuilder (ctx context. Context , namespace , sfsname string , lister scheduler.VPodLister , podCapacity int32 , schedulerPolicy scheduler.SchedulerPolicyType , schedPolicy * scheduler. SchedulerPolicy , deschedPolicy * scheduler.SchedulerPolicy , podlister corev1.PodNamespaceLister , nodeLister corev1.NodeLister , statefulSetCache * scheduler.ScaleCache ) StateAccessor {
164
+ func NewStateBuilder (sfsname string , lister scheduler.VPodLister , podCapacity int32 , schedulerPolicy scheduler.SchedulerPolicyType , schedPolicy , deschedPolicy * scheduler.SchedulerPolicy , podlister corev1.PodNamespaceLister , nodeLister corev1.NodeLister , statefulSetCache * scheduler.ScaleCache ) StateAccessor {
170
165
171
166
return & stateBuilder {
172
- ctx : ctx ,
173
- logger : logging .FromContext (ctx ),
174
167
vpodLister : lister ,
175
168
capacity : podCapacity ,
176
169
schedulerPolicy : schedulerPolicy ,
@@ -183,15 +176,18 @@ func NewStateBuilder(ctx context.Context, namespace, sfsname string, lister sche
183
176
}
184
177
}
185
178
186
- func (s * stateBuilder ) State (reserved map [types.NamespacedName ]map [string ]int32 ) (* State , error ) {
179
+ func (s * stateBuilder ) State (ctx context. Context , reserved map [types.NamespacedName ]map [string ]int32 ) (* State , error ) {
187
180
vpods , err := s .vpodLister ()
188
181
if err != nil {
189
182
return nil , err
190
183
}
191
184
192
- scale , err := s .statefulSetCache .GetScale (s .ctx , s .statefulSetName , metav1.GetOptions {})
185
+ logger := logging .FromContext (ctx ).With ("subcomponent" , "statebuilder" )
186
+ ctx = logging .WithLogger (ctx , logger )
187
+
188
+ scale , err := s .statefulSetCache .GetScale (ctx , s .statefulSetName , metav1.GetOptions {})
193
189
if err != nil {
194
- s . logger .Infow ("failed to get statefulset" , zap .Error (err ))
190
+ logger .Infow ("failed to get statefulset" , zap .Error (err ))
195
191
return nil , err
196
192
}
197
193
@@ -235,36 +231,35 @@ func (s *stateBuilder) State(reserved map[types.NamespacedName]map[string]int32)
235
231
}
236
232
237
233
for podId := int32 (0 ); podId < scale .Spec .Replicas && s .podLister != nil ; podId ++ {
238
- var pod * v1.Pod
239
- wait .PollUntilContextTimeout (context .Background (), 50 * time .Millisecond , 5 * time .Second , true , func (ctx context.Context ) (bool , error ) {
240
- pod , err = s .podLister .Get (PodNameFromOrdinal (s .statefulSetName , podId ))
241
- return err == nil , nil
242
- })
243
-
244
- if pod != nil {
245
- if isPodUnschedulable (pod ) {
246
- // Pod is marked for eviction - CANNOT SCHEDULE VREPS on this pod.
247
- continue
248
- }
249
-
250
- node , err := s .nodeLister .Get (pod .Spec .NodeName )
251
- if err != nil {
252
- return nil , err
253
- }
234
+ pod , err := s .podLister .Get (PodNameFromOrdinal (s .statefulSetName , podId ))
235
+ if err != nil {
236
+ logger .Warnw ("Failed to get pod" , zap .Int32 ("ordinal" , podId ), zap .Error (err ))
237
+ continue
238
+ }
239
+ if isPodUnschedulable (pod ) {
240
+ // Pod is marked for eviction - CANNOT SCHEDULE VREPS on this pod.
241
+ logger .Debugw ("Pod is unschedulable" , zap .Any ("pod" , pod ))
242
+ continue
243
+ }
254
244
255
- if isNodeUnschedulable ( node ) {
256
- // Node is marked as Unschedulable - CANNOT SCHEDULE VREPS on a pod running on this node.
257
- continue
258
- }
245
+ node , err := s . nodeLister . Get ( pod . Spec . NodeName )
246
+ if err != nil {
247
+ return nil , err
248
+ }
259
249
260
- // Pod has no annotation or not annotated as unschedulable and
261
- // not on an unschedulable node, so add to feasible
262
- schedulablePods .Insert (podId )
250
+ if isNodeUnschedulable (node ) {
251
+ // Node is marked as Unschedulable - CANNOT SCHEDULE VREPS on a pod running on this node.
252
+ logger .Debugw ("Pod is on an unschedulable node" , zap .Any ("pod" , node ))
253
+ continue
263
254
}
255
+
256
+ // Pod has no annotation or not annotated as unschedulable and
257
+ // not on an unschedulable node, so add to feasible
258
+ schedulablePods .Insert (podId )
264
259
}
265
260
266
261
for _ , p := range schedulablePods .List () {
267
- free , last = s .updateFreeCapacity (free , last , PodNameFromOrdinal (s .statefulSetName , p ), 0 )
262
+ free , last = s .updateFreeCapacity (logger , free , last , PodNameFromOrdinal (s .statefulSetName , p ), 0 )
268
263
}
269
264
270
265
// Getting current state from existing placements for all vpods
@@ -286,15 +281,14 @@ func (s *stateBuilder) State(reserved map[types.NamespacedName]map[string]int32)
286
281
// Account for reserved vreplicas
287
282
vreplicas = withReserved (vpod .GetKey (), podName , vreplicas , reserved )
288
283
289
- free , last = s .updateFreeCapacity (free , last , podName , vreplicas )
284
+ free , last = s .updateFreeCapacity (logger , free , last , podName , vreplicas )
290
285
291
286
withPlacement [vpod .GetKey ()][podName ] = true
292
287
293
- var pod * v1.Pod
294
- wait .PollUntilContextTimeout (context .Background (), 50 * time .Millisecond , 5 * time .Second , true , func (ctx context.Context ) (bool , error ) {
295
- pod , err = s .podLister .Get (podName )
296
- return err == nil , nil
297
- })
288
+ pod , err := s .podLister .Get (podName )
289
+ if err != nil {
290
+ logger .Warnw ("Failed to get pod" , zap .String ("podName" , podName ), zap .Error (err ))
291
+ }
298
292
299
293
if pod != nil && schedulablePods .Has (OrdinalFromPodName (pod .GetName ())) {
300
294
nodeName := pod .Spec .NodeName //node name for this pod
@@ -315,11 +309,10 @@ func (s *stateBuilder) State(reserved map[types.NamespacedName]map[string]int32)
315
309
continue
316
310
}
317
311
318
- var pod * v1.Pod
319
- wait .PollUntilContextTimeout (context .Background (), 50 * time .Millisecond , 5 * time .Second , true , func (ctx context.Context ) (bool , error ) {
320
- pod , err = s .podLister .Get (podName )
321
- return err == nil , nil
322
- })
312
+ pod , err := s .podLister .Get (podName )
313
+ if err != nil {
314
+ logger .Warnw ("Failed to get pod" , zap .String ("podName" , podName ), zap .Error (err ))
315
+ }
323
316
324
317
if pod != nil && schedulablePods .Has (OrdinalFromPodName (pod .GetName ())) {
325
318
nodeName := pod .Spec .NodeName //node name for this pod
@@ -330,15 +323,15 @@ func (s *stateBuilder) State(reserved map[types.NamespacedName]map[string]int32)
330
323
}
331
324
}
332
325
333
- free , last = s .updateFreeCapacity (free , last , podName , rvreplicas )
326
+ free , last = s .updateFreeCapacity (logger , free , last , podName , rvreplicas )
334
327
}
335
328
}
336
329
337
330
state := & State {FreeCap : free , SchedulablePods : schedulablePods .List (), LastOrdinal : last , Capacity : s .capacity , Replicas : scale .Spec .Replicas , NumZones : int32 (len (zoneMap )), NumNodes : int32 (len (nodeToZoneMap )),
338
331
SchedulerPolicy : s .schedulerPolicy , SchedPolicy : s .schedPolicy , DeschedPolicy : s .deschedPolicy , NodeToZoneMap : nodeToZoneMap , StatefulSetName : s .statefulSetName , PodLister : s .podLister ,
339
332
PodSpread : podSpread , NodeSpread : nodeSpread , ZoneSpread : zoneSpread , Pending : pending , ExpectedVReplicaByVPod : expectedVReplicasByVPod }
340
333
341
- s . logger .Infow ("cluster state info" , zap .Any ("state" , state ), zap .Any ("reserved" , toJSONable (reserved )))
334
+ logger .Infow ("cluster state info" , zap .Any ("state" , state ), zap .Any ("reserved" , toJSONable (reserved )))
342
335
343
336
return state , nil
344
337
}
@@ -350,7 +343,7 @@ func pendingFromVPod(vpod scheduler.VPod) int32 {
350
343
return int32 (math .Max (float64 (0 ), float64 (expected - scheduled )))
351
344
}
352
345
353
- func (s * stateBuilder ) updateFreeCapacity (free []int32 , last int32 , podName string , vreplicas int32 ) ([]int32 , int32 ) {
346
+ func (s * stateBuilder ) updateFreeCapacity (logger * zap. SugaredLogger , free []int32 , last int32 , podName string , vreplicas int32 ) ([]int32 , int32 ) {
354
347
ordinal := OrdinalFromPodName (podName )
355
348
free = grow (free , ordinal , s .capacity )
356
349
@@ -359,7 +352,7 @@ func (s *stateBuilder) updateFreeCapacity(free []int32, last int32, podName stri
359
352
// Assert the pod is not overcommitted
360
353
if free [ordinal ] < 0 {
361
354
// This should not happen anymore. Log as an error but do not interrupt the current scheduling.
362
- s . logger .Warnw ("pod is overcommitted" , zap .String ("podName" , podName ), zap .Int32 ("free" , free [ordinal ]))
355
+ logger .Warnw ("pod is overcommitted" , zap .String ("podName" , podName ), zap .Int32 ("free" , free [ordinal ]))
363
356
}
364
357
365
358
if ordinal > last {
0 commit comments