kubernetes/test/e2e/scheduling/predicates.go

const maxNumberOfPods

const defaultTimeout

var localStorageVersion

var workerNodes

type pausePodConfig

var _

func patchPod(cs clientset.Interface, old, new *v1.Pod) (*v1.Pod, error) {}

// printAllPodsOnNode outputs status of all kubelet pods into log.
func printAllPodsOnNode(ctx context.Context, c clientset.Interface, nodeName string) {}

func initPausePod(f *framework.Framework, conf pausePodConfig) *v1.Pod {}

func createPausePod(ctx context.Context, f *framework.Framework, conf pausePodConfig) *v1.Pod {}

func runPausePod(ctx context.Context, f *framework.Framework, conf pausePodConfig) *v1.Pod {}

func runPodAndGetNodeName(ctx context.Context, f *framework.Framework, conf pausePodConfig) string {}

func getRequestedCPU(pod v1.Pod) int64 {}

func getRequestedStorageEphemeralStorage(pod v1.Pod) int64 {}

// removeTaintFromNodeAction returns a closure that removes the given taint
// from the given node upon invocation.
func removeTaintFromNodeAction(cs clientset.Interface, nodeName string, testTaint v1.Taint) Action {}

// createPausePodAction returns a closure that creates a pause pod upon invocation.
func createPausePodAction(f *framework.Framework, conf pausePodConfig) Action {}

// WaitForSchedulerAfterAction performs the provided action and then waits for
// scheduler to act on the given pod.
func WaitForSchedulerAfterAction(ctx context.Context, f *framework.Framework, action Action, ns, podName string, expectSuccess bool) {}

// TODO: upgrade calls in PodAffinity tests when we're able to run them
func verifyResult(ctx context.Context, c clientset.Interface, expectedScheduled int, expectedNotScheduled int, ns string) {}

// GetNodeThatCanRunPod trying to launch a pod without a label to get a node which can launch it
func GetNodeThatCanRunPod(ctx context.Context, f *framework.Framework) string {}

// Get2NodesThatCanRunPod return a 2-node slice where can run pod.
func Get2NodesThatCanRunPod(ctx context.Context, f *framework.Framework) []string {}

func getNodeThatCanRunPodWithoutToleration(ctx context.Context, f *framework.Framework) string {}

// CreateHostPortPods creates RC with host port 4321
func CreateHostPortPods(ctx context.Context, f *framework.Framework, id string, replicas int, expectRunning bool) {}

// CreateNodeSelectorPods creates RC with host port 4321 and defines node selector
func CreateNodeSelectorPods(ctx context.Context, f *framework.Framework, id string, replicas int, nodeSelector map[string]string, expectRunning bool) error {}

// create pod which using hostport on the specified node according to the nodeSelector
// it starts an http server on the exposed port
func createHostPortPodOnNode(ctx context.Context, f *framework.Framework, podName, ns, hostIP string, port int32, protocol v1.Protocol, nodeSelector map[string]string, expectScheduled bool) {}

// GetPodsScheduled returns a number of currently scheduled and not scheduled Pods on worker nodes.
func GetPodsScheduled(workerNodes sets.Set[string], pods *v1.PodList) (scheduledPods, notScheduledPods []v1.Pod) {}

// getNodeHostIP returns the first internal IP on the node matching the main Cluster IP family
func getNodeHostIP(ctx context.Context, f *framework.Framework, nodeName string) string {}