kubernetes/pkg/controller/tainteviction/taint_eviction_test.go

var timeForControllerToProgressForSanityCheck

func getPodsAssignedToNode(ctx context.Context, c *fake.Clientset) GetPodsByNodeNameFunc {}

func createNoExecuteTaint(index int) corev1.Taint {}

func addToleration(pod *corev1.Pod, index int, duration int64) *corev1.Pod {}

func addTaintsToNode(node *corev1.Node, key, value string, indices []int) *corev1.Node {}

var alwaysReady

func setupNewController(ctx context.Context, fakeClientSet *fake.Clientset) (*Controller, cache.Indexer, cache.Indexer) {}

type timestampedPod

type durationSlice

func (a durationSlice) Len() int           {}

func (a durationSlice) Swap(i, j int)      {}

func (a durationSlice) Less(i, j int) bool {}

func TestFilterNoExecuteTaints(t *testing.T) {}

func TestCreatePod(t *testing.T) {}

func TestDeletePod(t *testing.T) {}

func TestUpdatePod(t *testing.T) {}

func TestCreateNode(t *testing.T) {}

func TestDeleteNode(t *testing.T) {}

func TestUpdateNode(t *testing.T) {}

func TestUpdateNodeWithMultipleTaints(t *testing.T) {}

func TestUpdateNodeWithMultiplePods(t *testing.T) {}

func TestGetMinTolerationTime(t *testing.T) {}

// TestEventualConsistency verifies if getPodsAssignedToNode returns incomplete data
// (e.g. due to watch latency), it will reconcile the remaining pods eventually.
// This scenario is partially covered by TestUpdatePods, but given this is an important
// property of TaintManager, it's better to have explicit test for this.
func TestEventualConsistency(t *testing.T) {}

func verifyPodActions(t *testing.T, description string, fakeClientset *fake.Clientset, expectPatch, expectDelete bool) {}

// TestPodDeletionEvent Verify that the output events are as expected
func TestPodDeletionEvent(t *testing.T) {}