mirror of
https://github.com/k3s-io/k3s.git
synced 2024-06-07 19:41:36 +00:00
807 lines
33 KiB
Go
807 lines
33 KiB
Go
/*
|
|
Copyright 2014 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package scheduler
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"math/rand"
|
|
"os"
|
|
"time"
|
|
|
|
"k8s.io/klog"
|
|
|
|
v1 "k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
|
"k8s.io/client-go/informers"
|
|
coreinformers "k8s.io/client-go/informers/core/v1"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
"k8s.io/client-go/tools/cache"
|
|
"k8s.io/client-go/tools/events"
|
|
podutil "k8s.io/kubernetes/pkg/api/v1/pod"
|
|
kubefeatures "k8s.io/kubernetes/pkg/features"
|
|
schedulerapi "k8s.io/kubernetes/pkg/scheduler/apis/config"
|
|
"k8s.io/kubernetes/pkg/scheduler/apis/config/scheme"
|
|
"k8s.io/kubernetes/pkg/scheduler/core"
|
|
frameworkplugins "k8s.io/kubernetes/pkg/scheduler/framework/plugins"
|
|
framework "k8s.io/kubernetes/pkg/scheduler/framework/v1alpha1"
|
|
internalcache "k8s.io/kubernetes/pkg/scheduler/internal/cache"
|
|
internalqueue "k8s.io/kubernetes/pkg/scheduler/internal/queue"
|
|
"k8s.io/kubernetes/pkg/scheduler/metrics"
|
|
nodeinfosnapshot "k8s.io/kubernetes/pkg/scheduler/nodeinfo/snapshot"
|
|
"k8s.io/kubernetes/pkg/scheduler/volumebinder"
|
|
)
|
|
|
|
const (
|
|
// BindTimeoutSeconds defines the default bind timeout
|
|
BindTimeoutSeconds = 100
|
|
// SchedulerError is the reason recorded for events when an error occurs during scheduling a pod.
|
|
SchedulerError = "SchedulerError"
|
|
// Percentage of framework metrics to be sampled.
|
|
frameworkMetricsSamplePercent = 10
|
|
)
|
|
|
|
// podConditionUpdater updates the condition of a pod based on the passed
|
|
// PodCondition
|
|
// TODO (ahmad-diaa): Remove type and replace it with scheduler methods
|
|
type podConditionUpdater interface {
|
|
update(pod *v1.Pod, podCondition *v1.PodCondition) error
|
|
}
|
|
|
|
// PodPreemptor has methods needed to delete a pod and to update 'NominatedPod'
|
|
// field of the preemptor pod.
|
|
// TODO (ahmad-diaa): Remove type and replace it with scheduler methods
|
|
type podPreemptor interface {
|
|
getUpdatedPod(pod *v1.Pod) (*v1.Pod, error)
|
|
deletePod(pod *v1.Pod) error
|
|
setNominatedNodeName(pod *v1.Pod, nominatedNode string) error
|
|
removeNominatedNodeName(pod *v1.Pod) error
|
|
}
|
|
|
|
// Scheduler watches for new unscheduled pods. It attempts to find
|
|
// nodes that they fit on and writes bindings back to the api server.
|
|
type Scheduler struct {
|
|
// It is expected that changes made via SchedulerCache will be observed
|
|
// by NodeLister and Algorithm.
|
|
SchedulerCache internalcache.Cache
|
|
|
|
Algorithm core.ScheduleAlgorithm
|
|
GetBinder func(pod *v1.Pod) Binder
|
|
// PodConditionUpdater is used only in case of scheduling errors. If we succeed
|
|
// with scheduling, PodScheduled condition will be updated in apiserver in /bind
|
|
// handler so that binding and setting PodCondition it is atomic.
|
|
podConditionUpdater podConditionUpdater
|
|
// PodPreemptor is used to evict pods and update 'NominatedNode' field of
|
|
// the preemptor pod.
|
|
podPreemptor podPreemptor
|
|
// Framework runs scheduler plugins at configured extension points.
|
|
Framework framework.Framework
|
|
|
|
// NextPod should be a function that blocks until the next pod
|
|
// is available. We don't use a channel for this, because scheduling
|
|
// a pod may take some amount of time and we don't want pods to get
|
|
// stale while they sit in a channel.
|
|
NextPod func() *framework.PodInfo
|
|
|
|
// Error is called if there is an error. It is passed the pod in
|
|
// question, and the error
|
|
Error func(*framework.PodInfo, error)
|
|
|
|
// Recorder is the EventRecorder to use
|
|
Recorder events.EventRecorder
|
|
|
|
// Close this to shut down the scheduler.
|
|
StopEverything <-chan struct{}
|
|
|
|
// VolumeBinder handles PVC/PV binding for the pod.
|
|
VolumeBinder *volumebinder.VolumeBinder
|
|
|
|
// Disable pod preemption or not.
|
|
DisablePreemption bool
|
|
|
|
// SchedulingQueue holds pods to be scheduled
|
|
SchedulingQueue internalqueue.SchedulingQueue
|
|
|
|
scheduledPodsHasSynced func() bool
|
|
|
|
// The final configuration of the framework.
|
|
Plugins schedulerapi.Plugins
|
|
PluginConfig []schedulerapi.PluginConfig
|
|
}
|
|
|
|
// Cache returns the cache in scheduler for test to check the data in scheduler.
|
|
func (sched *Scheduler) Cache() internalcache.Cache {
|
|
return sched.SchedulerCache
|
|
}
|
|
|
|
type schedulerOptions struct {
|
|
schedulerName string
|
|
schedulerAlgorithmSource schedulerapi.SchedulerAlgorithmSource
|
|
hardPodAffinitySymmetricWeight int32
|
|
disablePreemption bool
|
|
percentageOfNodesToScore int32
|
|
bindTimeoutSeconds int64
|
|
podInitialBackoffSeconds int64
|
|
podMaxBackoffSeconds int64
|
|
// Default registry contains all in-tree plugins
|
|
frameworkDefaultRegistry framework.Registry
|
|
// This registry contains out of tree plugins to be merged with default registry.
|
|
frameworkOutOfTreeRegistry framework.Registry
|
|
frameworkConfigProducerRegistry *frameworkplugins.ConfigProducerRegistry
|
|
frameworkPlugins *schedulerapi.Plugins
|
|
frameworkPluginConfig []schedulerapi.PluginConfig
|
|
}
|
|
|
|
// Option configures a Scheduler
|
|
type Option func(*schedulerOptions)
|
|
|
|
// WithName sets schedulerName for Scheduler, the default schedulerName is default-scheduler
|
|
func WithName(schedulerName string) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.schedulerName = schedulerName
|
|
}
|
|
}
|
|
|
|
// WithAlgorithmSource sets schedulerAlgorithmSource for Scheduler, the default is a source with DefaultProvider.
|
|
func WithAlgorithmSource(source schedulerapi.SchedulerAlgorithmSource) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.schedulerAlgorithmSource = source
|
|
}
|
|
}
|
|
|
|
// WithHardPodAffinitySymmetricWeight sets hardPodAffinitySymmetricWeight for Scheduler, the default value is 1
|
|
func WithHardPodAffinitySymmetricWeight(hardPodAffinitySymmetricWeight int32) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.hardPodAffinitySymmetricWeight = hardPodAffinitySymmetricWeight
|
|
}
|
|
}
|
|
|
|
// WithPreemptionDisabled sets disablePreemption for Scheduler, the default value is false
|
|
func WithPreemptionDisabled(disablePreemption bool) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.disablePreemption = disablePreemption
|
|
}
|
|
}
|
|
|
|
// WithPercentageOfNodesToScore sets percentageOfNodesToScore for Scheduler, the default value is 50
|
|
func WithPercentageOfNodesToScore(percentageOfNodesToScore int32) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.percentageOfNodesToScore = percentageOfNodesToScore
|
|
}
|
|
}
|
|
|
|
// WithBindTimeoutSeconds sets bindTimeoutSeconds for Scheduler, the default value is 100
|
|
func WithBindTimeoutSeconds(bindTimeoutSeconds int64) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.bindTimeoutSeconds = bindTimeoutSeconds
|
|
}
|
|
}
|
|
|
|
// WithFrameworkDefaultRegistry sets the framework's default registry. This is only used in integration tests.
|
|
func WithFrameworkDefaultRegistry(registry framework.Registry) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.frameworkDefaultRegistry = registry
|
|
}
|
|
}
|
|
|
|
// WithFrameworkOutOfTreeRegistry sets the registry for out-of-tree plugins. Those plugins
|
|
// will be appended to the default registry.
|
|
func WithFrameworkOutOfTreeRegistry(registry framework.Registry) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.frameworkOutOfTreeRegistry = registry
|
|
}
|
|
}
|
|
|
|
// WithFrameworkConfigProducerRegistry sets the framework plugin producer registry.
|
|
func WithFrameworkConfigProducerRegistry(registry *frameworkplugins.ConfigProducerRegistry) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.frameworkConfigProducerRegistry = registry
|
|
}
|
|
}
|
|
|
|
// WithFrameworkPlugins sets the plugins that the framework should be configured with.
|
|
func WithFrameworkPlugins(plugins *schedulerapi.Plugins) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.frameworkPlugins = plugins
|
|
}
|
|
}
|
|
|
|
// WithFrameworkPluginConfig sets the PluginConfig slice that the framework should be configured with.
|
|
func WithFrameworkPluginConfig(pluginConfig []schedulerapi.PluginConfig) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.frameworkPluginConfig = pluginConfig
|
|
}
|
|
}
|
|
|
|
// WithPodInitialBackoffSeconds sets podInitialBackoffSeconds for Scheduler, the default value is 1
|
|
func WithPodInitialBackoffSeconds(podInitialBackoffSeconds int64) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.podInitialBackoffSeconds = podInitialBackoffSeconds
|
|
}
|
|
}
|
|
|
|
// WithPodMaxBackoffSeconds sets podMaxBackoffSeconds for Scheduler, the default value is 10
|
|
func WithPodMaxBackoffSeconds(podMaxBackoffSeconds int64) Option {
|
|
return func(o *schedulerOptions) {
|
|
o.podMaxBackoffSeconds = podMaxBackoffSeconds
|
|
}
|
|
}
|
|
|
|
var defaultSchedulerOptions = schedulerOptions{
|
|
schedulerName: v1.DefaultSchedulerName,
|
|
schedulerAlgorithmSource: schedulerapi.SchedulerAlgorithmSource{
|
|
Provider: defaultAlgorithmSourceProviderName(),
|
|
},
|
|
hardPodAffinitySymmetricWeight: v1.DefaultHardPodAffinitySymmetricWeight,
|
|
disablePreemption: false,
|
|
percentageOfNodesToScore: schedulerapi.DefaultPercentageOfNodesToScore,
|
|
bindTimeoutSeconds: BindTimeoutSeconds,
|
|
podInitialBackoffSeconds: int64(internalqueue.DefaultPodInitialBackoffDuration.Seconds()),
|
|
podMaxBackoffSeconds: int64(internalqueue.DefaultPodMaxBackoffDuration.Seconds()),
|
|
frameworkConfigProducerRegistry: frameworkplugins.NewDefaultConfigProducerRegistry(),
|
|
// The plugins and pluginConfig options are currently nil because we currently don't have
|
|
// "default" plugins. All plugins that we run through the framework currently come from two
|
|
// sources: 1) specified in component config, in which case those two options should be
|
|
// set using their corresponding With* functions, 2) predicate/priority-mapped plugins, which
|
|
// pluginConfigProducerRegistry contains a mapping for and produces their configurations.
|
|
// TODO(ahg-g) Once predicates and priorities are migrated to natively run as plugins, the
|
|
// below two parameters will be populated accordingly.
|
|
frameworkPlugins: nil,
|
|
frameworkPluginConfig: nil,
|
|
}
|
|
|
|
// New returns a Scheduler
|
|
func New(client clientset.Interface,
|
|
informerFactory informers.SharedInformerFactory,
|
|
podInformer coreinformers.PodInformer,
|
|
recorder events.EventRecorder,
|
|
stopCh <-chan struct{},
|
|
opts ...Option) (*Scheduler, error) {
|
|
|
|
stopEverything := stopCh
|
|
if stopEverything == nil {
|
|
stopEverything = wait.NeverStop
|
|
}
|
|
|
|
options := defaultSchedulerOptions
|
|
for _, opt := range opts {
|
|
opt(&options)
|
|
}
|
|
|
|
schedulerCache := internalcache.New(30*time.Second, stopEverything)
|
|
volumeBinder := volumebinder.NewVolumeBinder(
|
|
client,
|
|
informerFactory.Core().V1().Nodes(),
|
|
informerFactory.Storage().V1().CSINodes(),
|
|
informerFactory.Core().V1().PersistentVolumeClaims(),
|
|
informerFactory.Core().V1().PersistentVolumes(),
|
|
informerFactory.Storage().V1().StorageClasses(),
|
|
time.Duration(options.bindTimeoutSeconds)*time.Second,
|
|
)
|
|
|
|
registry := options.frameworkDefaultRegistry
|
|
if registry == nil {
|
|
registry = frameworkplugins.NewDefaultRegistry(&frameworkplugins.RegistryArgs{
|
|
VolumeBinder: volumeBinder,
|
|
})
|
|
}
|
|
registry.Merge(options.frameworkOutOfTreeRegistry)
|
|
|
|
snapshot := nodeinfosnapshot.NewEmptySnapshot()
|
|
|
|
configurator := &Configurator{
|
|
client: client,
|
|
informerFactory: informerFactory,
|
|
podInformer: podInformer,
|
|
volumeBinder: volumeBinder,
|
|
schedulerCache: schedulerCache,
|
|
StopEverything: stopEverything,
|
|
hardPodAffinitySymmetricWeight: options.hardPodAffinitySymmetricWeight,
|
|
disablePreemption: options.disablePreemption,
|
|
percentageOfNodesToScore: options.percentageOfNodesToScore,
|
|
bindTimeoutSeconds: options.bindTimeoutSeconds,
|
|
podInitialBackoffSeconds: options.podInitialBackoffSeconds,
|
|
podMaxBackoffSeconds: options.podMaxBackoffSeconds,
|
|
enableNonPreempting: utilfeature.DefaultFeatureGate.Enabled(kubefeatures.NonPreemptingPriority),
|
|
registry: registry,
|
|
plugins: options.frameworkPlugins,
|
|
pluginConfig: options.frameworkPluginConfig,
|
|
pluginConfigProducerRegistry: options.frameworkConfigProducerRegistry,
|
|
nodeInfoSnapshot: snapshot,
|
|
algorithmFactoryArgs: AlgorithmFactoryArgs{
|
|
SharedLister: snapshot,
|
|
InformerFactory: informerFactory,
|
|
VolumeBinder: volumeBinder,
|
|
HardPodAffinitySymmetricWeight: options.hardPodAffinitySymmetricWeight,
|
|
},
|
|
configProducerArgs: &frameworkplugins.ConfigProducerArgs{},
|
|
}
|
|
|
|
var sched *Scheduler
|
|
source := options.schedulerAlgorithmSource
|
|
switch {
|
|
case source.Provider != nil:
|
|
// Create the config from a named algorithm provider.
|
|
sc, err := configurator.CreateFromProvider(*source.Provider)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("couldn't create scheduler using provider %q: %v", *source.Provider, err)
|
|
}
|
|
sched = sc
|
|
case source.Policy != nil:
|
|
// Create the config from a user specified policy source.
|
|
policy := &schedulerapi.Policy{}
|
|
switch {
|
|
case source.Policy.File != nil:
|
|
if err := initPolicyFromFile(source.Policy.File.Path, policy); err != nil {
|
|
return nil, err
|
|
}
|
|
case source.Policy.ConfigMap != nil:
|
|
if err := initPolicyFromConfigMap(client, source.Policy.ConfigMap, policy); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
sc, err := configurator.CreateFromConfig(*policy)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("couldn't create scheduler from policy: %v", err)
|
|
}
|
|
sched = sc
|
|
default:
|
|
return nil, fmt.Errorf("unsupported algorithm source: %v", source)
|
|
}
|
|
metrics.Register()
|
|
// Additional tweaks to the config produced by the configurator.
|
|
sched.Recorder = recorder
|
|
sched.DisablePreemption = options.disablePreemption
|
|
sched.StopEverything = stopEverything
|
|
sched.podConditionUpdater = &podConditionUpdaterImpl{client}
|
|
sched.podPreemptor = &podPreemptorImpl{client}
|
|
sched.scheduledPodsHasSynced = podInformer.Informer().HasSynced
|
|
|
|
AddAllEventHandlers(sched, options.schedulerName, informerFactory, podInformer)
|
|
return sched, nil
|
|
}
|
|
|
|
// initPolicyFromFile initialize policy from file
|
|
func initPolicyFromFile(policyFile string, policy *schedulerapi.Policy) error {
|
|
// Use a policy serialized in a file.
|
|
_, err := os.Stat(policyFile)
|
|
if err != nil {
|
|
return fmt.Errorf("missing policy config file %s", policyFile)
|
|
}
|
|
data, err := ioutil.ReadFile(policyFile)
|
|
if err != nil {
|
|
return fmt.Errorf("couldn't read policy config: %v", err)
|
|
}
|
|
err = runtime.DecodeInto(scheme.Codecs.UniversalDecoder(), []byte(data), policy)
|
|
if err != nil {
|
|
return fmt.Errorf("invalid policy: %v", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// initPolicyFromConfigMap initialize policy from configMap
|
|
func initPolicyFromConfigMap(client clientset.Interface, policyRef *schedulerapi.SchedulerPolicyConfigMapSource, policy *schedulerapi.Policy) error {
|
|
// Use a policy serialized in a config map value.
|
|
policyConfigMap, err := client.CoreV1().ConfigMaps(policyRef.Namespace).Get(policyRef.Name, metav1.GetOptions{})
|
|
if err != nil {
|
|
return fmt.Errorf("couldn't get policy config map %s/%s: %v", policyRef.Namespace, policyRef.Name, err)
|
|
}
|
|
data, found := policyConfigMap.Data[schedulerapi.SchedulerPolicyConfigMapKey]
|
|
if !found {
|
|
return fmt.Errorf("missing policy config map value at key %q", schedulerapi.SchedulerPolicyConfigMapKey)
|
|
}
|
|
err = runtime.DecodeInto(scheme.Codecs.UniversalDecoder(), []byte(data), policy)
|
|
if err != nil {
|
|
return fmt.Errorf("invalid policy: %v", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Run begins watching and scheduling. It waits for cache to be synced, then starts scheduling and blocked until the context is done.
|
|
func (sched *Scheduler) Run(ctx context.Context) {
|
|
if !cache.WaitForCacheSync(ctx.Done(), sched.scheduledPodsHasSynced) {
|
|
return
|
|
}
|
|
|
|
wait.UntilWithContext(ctx, sched.scheduleOne, 0)
|
|
}
|
|
|
|
// recordFailedSchedulingEvent records an event for the pod that indicates the
|
|
// pod has failed to schedule.
|
|
// NOTE: This function modifies "pod". "pod" should be copied before being passed.
|
|
func (sched *Scheduler) recordSchedulingFailure(podInfo *framework.PodInfo, err error, reason string, message string) {
|
|
sched.Error(podInfo, err)
|
|
pod := podInfo.Pod
|
|
sched.Recorder.Eventf(pod, nil, v1.EventTypeWarning, "FailedScheduling", "Scheduling", message)
|
|
if err := sched.podConditionUpdater.update(pod, &v1.PodCondition{
|
|
Type: v1.PodScheduled,
|
|
Status: v1.ConditionFalse,
|
|
Reason: reason,
|
|
Message: err.Error(),
|
|
}); err != nil {
|
|
klog.Errorf("Error updating the condition of the pod %s/%s: %v", pod.Namespace, pod.Name, err)
|
|
}
|
|
}
|
|
|
|
// preempt tries to create room for a pod that has failed to schedule, by preempting lower priority pods if possible.
|
|
// If it succeeds, it adds the name of the node where preemption has happened to the pod spec.
|
|
// It returns the node name and an error if any.
|
|
func (sched *Scheduler) preempt(ctx context.Context, state *framework.CycleState, fwk framework.Framework, preemptor *v1.Pod, scheduleErr error) (string, error) {
|
|
preemptor, err := sched.podPreemptor.getUpdatedPod(preemptor)
|
|
if err != nil {
|
|
klog.Errorf("Error getting the updated preemptor pod object: %v", err)
|
|
return "", err
|
|
}
|
|
|
|
node, victims, nominatedPodsToClear, err := sched.Algorithm.Preempt(ctx, state, preemptor, scheduleErr)
|
|
if err != nil {
|
|
klog.Errorf("Error preempting victims to make room for %v/%v: %v", preemptor.Namespace, preemptor.Name, err)
|
|
return "", err
|
|
}
|
|
var nodeName = ""
|
|
if node != nil {
|
|
nodeName = node.Name
|
|
// Update the scheduling queue with the nominated pod information. Without
|
|
// this, there would be a race condition between the next scheduling cycle
|
|
// and the time the scheduler receives a Pod Update for the nominated pod.
|
|
sched.SchedulingQueue.UpdateNominatedPodForNode(preemptor, nodeName)
|
|
|
|
// Make a call to update nominated node name of the pod on the API server.
|
|
err = sched.podPreemptor.setNominatedNodeName(preemptor, nodeName)
|
|
if err != nil {
|
|
klog.Errorf("Error in preemption process. Cannot set 'NominatedPod' on pod %v/%v: %v", preemptor.Namespace, preemptor.Name, err)
|
|
sched.SchedulingQueue.DeleteNominatedPodIfExists(preemptor)
|
|
return "", err
|
|
}
|
|
|
|
for _, victim := range victims {
|
|
if err := sched.podPreemptor.deletePod(victim); err != nil {
|
|
klog.Errorf("Error preempting pod %v/%v: %v", victim.Namespace, victim.Name, err)
|
|
return "", err
|
|
}
|
|
// If the victim is a WaitingPod, send a reject message to the PermitPlugin
|
|
if waitingPod := fwk.GetWaitingPod(victim.UID); waitingPod != nil {
|
|
waitingPod.Reject("preempted")
|
|
}
|
|
sched.Recorder.Eventf(victim, preemptor, v1.EventTypeNormal, "Preempted", "Preempting", "Preempted by %v/%v on node %v", preemptor.Namespace, preemptor.Name, nodeName)
|
|
|
|
}
|
|
metrics.PreemptionVictims.Observe(float64(len(victims)))
|
|
}
|
|
// Clearing nominated pods should happen outside of "if node != nil". Node could
|
|
// be nil when a pod with nominated node name is eligible to preempt again,
|
|
// but preemption logic does not find any node for it. In that case Preempt()
|
|
// function of generic_scheduler.go returns the pod itself for removal of
|
|
// the 'NominatedPod' field.
|
|
for _, p := range nominatedPodsToClear {
|
|
rErr := sched.podPreemptor.removeNominatedNodeName(p)
|
|
if rErr != nil {
|
|
klog.Errorf("Cannot remove 'NominatedPod' field of pod: %v", rErr)
|
|
// We do not return as this error is not critical.
|
|
}
|
|
}
|
|
return nodeName, err
|
|
}
|
|
|
|
// bindVolumes will make the API update with the assumed bindings and wait until
|
|
// the PV controller has completely finished the binding operation.
|
|
//
|
|
// If binding errors, times out or gets undone, then an error will be returned to
|
|
// retry scheduling.
|
|
func (sched *Scheduler) bindVolumes(assumed *v1.Pod) error {
|
|
klog.V(5).Infof("Trying to bind volumes for pod \"%v/%v\"", assumed.Namespace, assumed.Name)
|
|
err := sched.VolumeBinder.Binder.BindPodVolumes(assumed)
|
|
if err != nil {
|
|
klog.V(1).Infof("Failed to bind volumes for pod \"%v/%v\": %v", assumed.Namespace, assumed.Name, err)
|
|
|
|
// Unassume the Pod and retry scheduling
|
|
if forgetErr := sched.SchedulerCache.ForgetPod(assumed); forgetErr != nil {
|
|
klog.Errorf("scheduler cache ForgetPod failed: %v", forgetErr)
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
klog.V(5).Infof("Success binding volumes for pod \"%v/%v\"", assumed.Namespace, assumed.Name)
|
|
return nil
|
|
}
|
|
|
|
// assume signals to the cache that a pod is already in the cache, so that binding can be asynchronous.
|
|
// assume modifies `assumed`.
|
|
func (sched *Scheduler) assume(assumed *v1.Pod, host string) error {
|
|
// Optimistically assume that the binding will succeed and send it to apiserver
|
|
// in the background.
|
|
// If the binding fails, scheduler will release resources allocated to assumed pod
|
|
// immediately.
|
|
assumed.Spec.NodeName = host
|
|
|
|
if err := sched.SchedulerCache.AssumePod(assumed); err != nil {
|
|
klog.Errorf("scheduler cache AssumePod failed: %v", err)
|
|
return err
|
|
}
|
|
// if "assumed" is a nominated pod, we should remove it from internal cache
|
|
if sched.SchedulingQueue != nil {
|
|
sched.SchedulingQueue.DeleteNominatedPodIfExists(assumed)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// bind binds a pod to a given node defined in a binding object. We expect this to run asynchronously, so we
|
|
// handle binding metrics internally.
|
|
func (sched *Scheduler) bind(ctx context.Context, assumed *v1.Pod, targetNode string, state *framework.CycleState) error {
|
|
bindingStart := time.Now()
|
|
bindStatus := sched.Framework.RunBindPlugins(ctx, state, assumed, targetNode)
|
|
var err error
|
|
if !bindStatus.IsSuccess() {
|
|
if bindStatus.Code() == framework.Skip {
|
|
// All bind plugins chose to skip binding of this pod, call original binding function.
|
|
// If binding succeeds then PodScheduled condition will be updated in apiserver so that
|
|
// it's atomic with setting host.
|
|
err = sched.GetBinder(assumed).Bind(&v1.Binding{
|
|
ObjectMeta: metav1.ObjectMeta{Namespace: assumed.Namespace, Name: assumed.Name, UID: assumed.UID},
|
|
Target: v1.ObjectReference{
|
|
Kind: "Node",
|
|
Name: targetNode,
|
|
},
|
|
})
|
|
} else {
|
|
err = fmt.Errorf("Bind failure, code: %d: %v", bindStatus.Code(), bindStatus.Message())
|
|
}
|
|
}
|
|
if finErr := sched.SchedulerCache.FinishBinding(assumed); finErr != nil {
|
|
klog.Errorf("scheduler cache FinishBinding failed: %v", finErr)
|
|
}
|
|
if err != nil {
|
|
klog.V(1).Infof("Failed to bind pod: %v/%v", assumed.Namespace, assumed.Name)
|
|
if err := sched.SchedulerCache.ForgetPod(assumed); err != nil {
|
|
klog.Errorf("scheduler cache ForgetPod failed: %v", err)
|
|
}
|
|
return err
|
|
}
|
|
|
|
metrics.BindingLatency.Observe(metrics.SinceInSeconds(bindingStart))
|
|
metrics.DeprecatedBindingLatency.Observe(metrics.SinceInMicroseconds(bindingStart))
|
|
metrics.SchedulingLatency.WithLabelValues(metrics.Binding).Observe(metrics.SinceInSeconds(bindingStart))
|
|
metrics.DeprecatedSchedulingLatency.WithLabelValues(metrics.Binding).Observe(metrics.SinceInSeconds(bindingStart))
|
|
sched.Recorder.Eventf(assumed, nil, v1.EventTypeNormal, "Scheduled", "Binding", "Successfully assigned %v/%v to %v", assumed.Namespace, assumed.Name, targetNode)
|
|
return nil
|
|
}
|
|
|
|
// scheduleOne does the entire scheduling workflow for a single pod. It is serialized on the scheduling algorithm's host fitting.
|
|
func (sched *Scheduler) scheduleOne(ctx context.Context) {
|
|
fwk := sched.Framework
|
|
|
|
podInfo := sched.NextPod()
|
|
// pod could be nil when schedulerQueue is closed
|
|
if podInfo == nil || podInfo.Pod == nil {
|
|
return
|
|
}
|
|
pod := podInfo.Pod
|
|
if pod.DeletionTimestamp != nil {
|
|
sched.Recorder.Eventf(pod, nil, v1.EventTypeWarning, "FailedScheduling", "Scheduling", "skip schedule deleting pod: %v/%v", pod.Namespace, pod.Name)
|
|
klog.V(3).Infof("Skip schedule deleting pod: %v/%v", pod.Namespace, pod.Name)
|
|
return
|
|
}
|
|
|
|
klog.V(3).Infof("Attempting to schedule pod: %v/%v", pod.Namespace, pod.Name)
|
|
|
|
// Synchronously attempt to find a fit for the pod.
|
|
start := time.Now()
|
|
state := framework.NewCycleState()
|
|
state.SetRecordFrameworkMetrics(rand.Intn(100) < frameworkMetricsSamplePercent)
|
|
schedulingCycleCtx, cancel := context.WithCancel(ctx)
|
|
defer cancel()
|
|
scheduleResult, err := sched.Algorithm.Schedule(schedulingCycleCtx, state, pod)
|
|
if err != nil {
|
|
sched.recordSchedulingFailure(podInfo.DeepCopy(), err, v1.PodReasonUnschedulable, err.Error())
|
|
// Schedule() may have failed because the pod would not fit on any host, so we try to
|
|
// preempt, with the expectation that the next time the pod is tried for scheduling it
|
|
// will fit due to the preemption. It is also possible that a different pod will schedule
|
|
// into the resources that were preempted, but this is harmless.
|
|
if fitError, ok := err.(*core.FitError); ok {
|
|
if sched.DisablePreemption {
|
|
klog.V(3).Infof("Pod priority feature is not enabled or preemption is disabled by scheduler configuration." +
|
|
" No preemption is performed.")
|
|
} else {
|
|
preemptionStartTime := time.Now()
|
|
sched.preempt(schedulingCycleCtx, state, fwk, pod, fitError)
|
|
metrics.PreemptionAttempts.Inc()
|
|
metrics.SchedulingAlgorithmPreemptionEvaluationDuration.Observe(metrics.SinceInSeconds(preemptionStartTime))
|
|
metrics.DeprecatedSchedulingAlgorithmPreemptionEvaluationDuration.Observe(metrics.SinceInMicroseconds(preemptionStartTime))
|
|
metrics.SchedulingLatency.WithLabelValues(metrics.PreemptionEvaluation).Observe(metrics.SinceInSeconds(preemptionStartTime))
|
|
metrics.DeprecatedSchedulingLatency.WithLabelValues(metrics.PreemptionEvaluation).Observe(metrics.SinceInSeconds(preemptionStartTime))
|
|
}
|
|
// Pod did not fit anywhere, so it is counted as a failure. If preemption
|
|
// succeeds, the pod should get counted as a success the next time we try to
|
|
// schedule it. (hopefully)
|
|
metrics.PodScheduleFailures.Inc()
|
|
} else {
|
|
klog.Errorf("error selecting node for pod: %v", err)
|
|
metrics.PodScheduleErrors.Inc()
|
|
}
|
|
return
|
|
}
|
|
metrics.SchedulingAlgorithmLatency.Observe(metrics.SinceInSeconds(start))
|
|
metrics.DeprecatedSchedulingAlgorithmLatency.Observe(metrics.SinceInMicroseconds(start))
|
|
// Tell the cache to assume that a pod now is running on a given node, even though it hasn't been bound yet.
|
|
// This allows us to keep scheduling without waiting on binding to occur.
|
|
assumedPodInfo := podInfo.DeepCopy()
|
|
assumedPod := assumedPodInfo.Pod
|
|
|
|
// Assume volumes first before assuming the pod.
|
|
//
|
|
// If all volumes are completely bound, then allBound is true and binding will be skipped.
|
|
//
|
|
// Otherwise, binding of volumes is started after the pod is assumed, but before pod binding.
|
|
//
|
|
// This function modifies 'assumedPod' if volume binding is required.
|
|
allBound, err := sched.VolumeBinder.Binder.AssumePodVolumes(assumedPod, scheduleResult.SuggestedHost)
|
|
if err != nil {
|
|
sched.recordSchedulingFailure(assumedPodInfo, err, SchedulerError,
|
|
fmt.Sprintf("AssumePodVolumes failed: %v", err))
|
|
metrics.PodScheduleErrors.Inc()
|
|
return
|
|
}
|
|
|
|
// Run "reserve" plugins.
|
|
if sts := fwk.RunReservePlugins(schedulingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost); !sts.IsSuccess() {
|
|
sched.recordSchedulingFailure(assumedPodInfo, sts.AsError(), SchedulerError, sts.Message())
|
|
metrics.PodScheduleErrors.Inc()
|
|
return
|
|
}
|
|
|
|
// assume modifies `assumedPod` by setting NodeName=scheduleResult.SuggestedHost
|
|
err = sched.assume(assumedPod, scheduleResult.SuggestedHost)
|
|
if err != nil {
|
|
// This is most probably result of a BUG in retrying logic.
|
|
// We report an error here so that pod scheduling can be retried.
|
|
// This relies on the fact that Error will check if the pod has been bound
|
|
// to a node and if so will not add it back to the unscheduled pods queue
|
|
// (otherwise this would cause an infinite loop).
|
|
sched.recordSchedulingFailure(assumedPodInfo, err, SchedulerError, fmt.Sprintf("AssumePod failed: %v", err))
|
|
metrics.PodScheduleErrors.Inc()
|
|
// trigger un-reserve plugins to clean up state associated with the reserved Pod
|
|
fwk.RunUnreservePlugins(schedulingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
return
|
|
}
|
|
// bind the pod to its host asynchronously (we can do this b/c of the assumption step above).
|
|
go func() {
|
|
bindingCycleCtx, cancel := context.WithCancel(ctx)
|
|
defer cancel()
|
|
metrics.SchedulerGoroutines.WithLabelValues("binding").Inc()
|
|
defer metrics.SchedulerGoroutines.WithLabelValues("binding").Dec()
|
|
|
|
// Run "permit" plugins.
|
|
permitStatus := fwk.RunPermitPlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
if !permitStatus.IsSuccess() {
|
|
var reason string
|
|
if permitStatus.IsUnschedulable() {
|
|
metrics.PodScheduleFailures.Inc()
|
|
reason = v1.PodReasonUnschedulable
|
|
} else {
|
|
metrics.PodScheduleErrors.Inc()
|
|
reason = SchedulerError
|
|
}
|
|
if forgetErr := sched.Cache().ForgetPod(assumedPod); forgetErr != nil {
|
|
klog.Errorf("scheduler cache ForgetPod failed: %v", forgetErr)
|
|
}
|
|
// trigger un-reserve plugins to clean up state associated with the reserved Pod
|
|
fwk.RunUnreservePlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
sched.recordSchedulingFailure(assumedPodInfo, permitStatus.AsError(), reason, permitStatus.Message())
|
|
return
|
|
}
|
|
|
|
// Bind volumes first before Pod
|
|
if !allBound {
|
|
err := sched.bindVolumes(assumedPod)
|
|
if err != nil {
|
|
sched.recordSchedulingFailure(assumedPodInfo, err, "VolumeBindingFailed", err.Error())
|
|
metrics.PodScheduleErrors.Inc()
|
|
// trigger un-reserve plugins to clean up state associated with the reserved Pod
|
|
fwk.RunUnreservePlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
return
|
|
}
|
|
}
|
|
|
|
// Run "prebind" plugins.
|
|
preBindStatus := fwk.RunPreBindPlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
if !preBindStatus.IsSuccess() {
|
|
var reason string
|
|
metrics.PodScheduleErrors.Inc()
|
|
reason = SchedulerError
|
|
if forgetErr := sched.Cache().ForgetPod(assumedPod); forgetErr != nil {
|
|
klog.Errorf("scheduler cache ForgetPod failed: %v", forgetErr)
|
|
}
|
|
// trigger un-reserve plugins to clean up state associated with the reserved Pod
|
|
fwk.RunUnreservePlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
sched.recordSchedulingFailure(assumedPodInfo, preBindStatus.AsError(), reason, preBindStatus.Message())
|
|
return
|
|
}
|
|
|
|
err := sched.bind(bindingCycleCtx, assumedPod, scheduleResult.SuggestedHost, state)
|
|
metrics.E2eSchedulingLatency.Observe(metrics.SinceInSeconds(start))
|
|
metrics.DeprecatedE2eSchedulingLatency.Observe(metrics.SinceInMicroseconds(start))
|
|
if err != nil {
|
|
metrics.PodScheduleErrors.Inc()
|
|
// trigger un-reserve plugins to clean up state associated with the reserved Pod
|
|
fwk.RunUnreservePlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
sched.recordSchedulingFailure(assumedPodInfo, err, SchedulerError, fmt.Sprintf("Binding rejected: %v", err))
|
|
} else {
|
|
// Calculating nodeResourceString can be heavy. Avoid it if klog verbosity is below 2.
|
|
if klog.V(2) {
|
|
klog.Infof("pod %v/%v is bound successfully on node %q, %d nodes evaluated, %d nodes were found feasible.", assumedPod.Namespace, assumedPod.Name, scheduleResult.SuggestedHost, scheduleResult.EvaluatedNodes, scheduleResult.FeasibleNodes)
|
|
}
|
|
|
|
metrics.PodScheduleSuccesses.Inc()
|
|
metrics.PodSchedulingAttempts.Observe(float64(podInfo.Attempts))
|
|
metrics.PodSchedulingDuration.Observe(metrics.SinceInSeconds(podInfo.InitialAttemptTimestamp))
|
|
|
|
// Run "postbind" plugins.
|
|
fwk.RunPostBindPlugins(bindingCycleCtx, state, assumedPod, scheduleResult.SuggestedHost)
|
|
}
|
|
}()
|
|
}
|
|
|
|
type podConditionUpdaterImpl struct {
|
|
Client clientset.Interface
|
|
}
|
|
|
|
func (p *podConditionUpdaterImpl) update(pod *v1.Pod, condition *v1.PodCondition) error {
|
|
klog.V(3).Infof("Updating pod condition for %s/%s to (%s==%s, Reason=%s)", pod.Namespace, pod.Name, condition.Type, condition.Status, condition.Reason)
|
|
if podutil.UpdatePodCondition(&pod.Status, condition) {
|
|
_, err := p.Client.CoreV1().Pods(pod.Namespace).UpdateStatus(pod)
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
type podPreemptorImpl struct {
|
|
Client clientset.Interface
|
|
}
|
|
|
|
func (p *podPreemptorImpl) getUpdatedPod(pod *v1.Pod) (*v1.Pod, error) {
|
|
return p.Client.CoreV1().Pods(pod.Namespace).Get(pod.Name, metav1.GetOptions{})
|
|
}
|
|
|
|
func (p *podPreemptorImpl) deletePod(pod *v1.Pod) error {
|
|
return p.Client.CoreV1().Pods(pod.Namespace).Delete(pod.Name, &metav1.DeleteOptions{})
|
|
}
|
|
|
|
func (p *podPreemptorImpl) setNominatedNodeName(pod *v1.Pod, nominatedNodeName string) error {
|
|
podCopy := pod.DeepCopy()
|
|
podCopy.Status.NominatedNodeName = nominatedNodeName
|
|
_, err := p.Client.CoreV1().Pods(pod.Namespace).UpdateStatus(podCopy)
|
|
return err
|
|
}
|
|
|
|
func (p *podPreemptorImpl) removeNominatedNodeName(pod *v1.Pod) error {
|
|
if len(pod.Status.NominatedNodeName) == 0 {
|
|
return nil
|
|
}
|
|
return p.setNominatedNodeName(pod, "")
|
|
}
|
|
|
|
func defaultAlgorithmSourceProviderName() *string {
|
|
provider := schedulerapi.SchedulerDefaultProviderName
|
|
return &provider
|
|
}
|