mirror of https://github.com/k3s-io/k3s.git
Merge pull request #71362 from aleksandra-malinowska/autoscaling-test-fix-26
Extend timeout before breaking node in autoscaling testpull/58/head
commit
610f48f424
|
@ -881,6 +881,19 @@ var _ = SIGDescribe("Cluster size autoscaling [Slow]", func() {
|
|||
clusterSize = manuallyIncreaseClusterSize(f, originalSizes)
|
||||
}
|
||||
|
||||
// If new nodes are disconnected too soon, they'll be considered not started
|
||||
// instead of unready, and cluster won't be considered unhealthy.
|
||||
//
|
||||
// More precisely, Cluster Autoscaler compares last transition time of
|
||||
// several readiness conditions to node create time. If it's within
|
||||
// 2 minutes, it'll assume node is just starting and not unhealthy.
|
||||
//
|
||||
// Nodes become ready in less than 1 minute after being created,
|
||||
// so waiting extra 2 minutes before breaking them (which triggers
|
||||
// readiness condition transition) should be sufficient, while
|
||||
// making no assumptions about minimal node startup time.
|
||||
time.Sleep(2 * time.Minute)
|
||||
|
||||
By("Block network connectivity to some nodes to simulate unhealthy cluster")
|
||||
nodesToBreakCount := int(math.Ceil(math.Max(float64(unhealthyClusterThreshold), 0.5*float64(clusterSize))))
|
||||
nodes, err := f.ClientSet.CoreV1().Nodes().List(metav1.ListOptions{FieldSelector: fields.Set{
|
||||
|
|
Loading…
Reference in New Issue