--- apiVersion: v1 kind: Pod metadata: annotations: kubectl.kubernetes.io/default-container: kube-scheduler kubernetes.io/config.hash: 8dd3d3608fe9c86b0f65904ec2353df4 kubernetes.io/config.mirror: 8dd3d3608fe9c86b0f65904ec2353df4 kubernetes.io/config.seen: "2026-03-20T08:40:20.099279240Z" kubernetes.io/config.source: file target.workload.openshift.io/management: '{"effect": "PreferredDuringScheduling"}' creationTimestamp: "2026-03-20T08:40:34Z" labels: app: openshift-kube-scheduler revision: "3" scheduler: "true" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:kubectl.kubernetes.io/default-container: {} f:kubernetes.io/config.hash: {} f:kubernetes.io/config.mirror: {} f:kubernetes.io/config.seen: {} f:kubernetes.io/config.source: {} f:target.workload.openshift.io/management: {} f:labels: .: {} f:app: {} f:revision: {} f:scheduler: {} f:ownerReferences: .: {} k:{"uid":"7bc50120-3879-4454-8162-ec2f348aada5"}: {} f:spec: f:containers: k:{"name":"kube-scheduler"}: .: {} f:args: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:initialDelaySeconds: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:ports: .: {} k:{"containerPort":10259,"protocol":"TCP"}: .: {} f:containerPort: {} f:hostPort: {} f:protocol: {} f:readinessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:initialDelaySeconds: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/kubernetes/static-pod-certs"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/kubernetes/static-pod-resources"}: .: {} f:mountPath: {} f:name: {} k:{"name":"kube-scheduler-cert-syncer"}: .: {} f:args: {} f:command: {} f:env: .: {} k:{"name":"POD_NAME"}: .: {} f:name: {} f:valueFrom: .: {} f:fieldRef: {} k:{"name":"POD_NAMESPACE"}: .: {} f:name: {} f:valueFrom: .: {} f:fieldRef: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/kubernetes/static-pod-certs"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/kubernetes/static-pod-resources"}: .: {} f:mountPath: {} f:name: {} k:{"name":"kube-scheduler-recovery-controller"}: .: {} f:args: {} f:command: {} f:env: .: {} k:{"name":"POD_NAMESPACE"}: .: {} f:name: {} f:valueFrom: .: {} f:fieldRef: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/kubernetes/static-pod-certs"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/kubernetes/static-pod-resources"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:initContainers: .: {} k:{"name":"wait-for-host-port"}: .: {} f:args: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:nodeName: {} f:priority: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"cert-dir"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"resource-dir"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} manager: kubelet operation: Update time: "2026-03-20T08:40:34Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: .: {} k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodScheduled"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:initContainerStatuses: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"192.168.32.10"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-03-20T08:49:38Z" name: openshift-kube-scheduler-master-0 namespace: openshift-kube-scheduler ownerReferences: - apiVersion: v1 controller: true kind: Node name: master-0 uid: 7bc50120-3879-4454-8162-ec2f348aada5 resourceVersion: "14461" uid: d9385aba-9a1c-4a49-a546-6f093a92f3d1 spec: containers: - args: - --config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml - --cert-dir=/var/run/kubernetes - --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig - --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig - --feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=true,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,StreamingCollectionEncodingToJSON=true,StreamingCollectionEncodingToProtobuf=true,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false - -v=2 - --tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt - --tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key - --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 - --tls-min-version=VersionTLS12 command: - hyperkube - kube-scheduler image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 3 httpGet: path: healthz port: 10259 scheme: HTTPS initialDelaySeconds: 45 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 name: kube-scheduler ports: - containerPort: 10259 hostPort: 10259 protocol: TCP readinessProbe: failureThreshold: 3 httpGet: path: healthz port: 10259 scheme: HTTPS initialDelaySeconds: 45 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 resources: requests: cpu: 15m memory: 50Mi terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/kubernetes/static-pod-resources name: resource-dir - mountPath: /etc/kubernetes/static-pod-certs name: cert-dir - args: - --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig - --namespace=$(POD_NAMESPACE) - --destination-dir=/etc/kubernetes/static-pod-certs command: - cluster-kube-scheduler-operator - cert-syncer env: - name: POD_NAME valueFrom: fieldRef: apiVersion: v1 fieldPath: metadata.name - name: POD_NAMESPACE valueFrom: fieldRef: apiVersion: v1 fieldPath: metadata.namespace image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 imagePullPolicy: IfNotPresent name: kube-scheduler-cert-syncer resources: requests: cpu: 5m memory: 50Mi terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/kubernetes/static-pod-resources name: resource-dir - mountPath: /etc/kubernetes/static-pod-certs name: cert-dir - args: - | timeout 3m /bin/bash -exuo pipefail -c 'while [ -n "$(ss -Htanop \( sport = 11443 \))" ]; do sleep 1; done' exec cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=${POD_NAMESPACE} --listen=0.0.0.0:11443 -v=2 command: - /bin/bash - -euxo - pipefail - -c env: - name: POD_NAMESPACE valueFrom: fieldRef: apiVersion: v1 fieldPath: metadata.namespace image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 imagePullPolicy: IfNotPresent name: kube-scheduler-recovery-controller resources: requests: cpu: 5m memory: 50Mi terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/kubernetes/static-pod-resources name: resource-dir - mountPath: /etc/kubernetes/static-pod-certs name: cert-dir dnsPolicy: ClusterFirst enableServiceLinks: true hostNetwork: true initContainers: - args: - | echo -n "Waiting for port :10259 to be released." while [ -n "$(ss -Htan '( sport = 10259 )')" ]; do echo -n "." sleep 1 done command: - /usr/bin/timeout - "30" - /bin/bash - -c image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 imagePullPolicy: IfNotPresent name: wait-for-host-port resources: requests: cpu: 15m memory: 50Mi terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError nodeName: master-0 preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - hostPath: path: /etc/kubernetes/static-pod-resources/kube-scheduler-pod-3 type: "" name: resource-dir - hostPath: path: /etc/kubernetes/static-pod-resources/kube-scheduler-certs type: "" name: cert-dir status: conditions: - lastProbeTime: null lastTransitionTime: "2026-03-20T08:40:56Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-03-20T08:40:56Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-03-20T08:48:18Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-03-20T08:48:18Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-03-20T08:40:56Z" status: "True" type: PodScheduled containerStatuses: - containerID: cri-o://90e686e94f5392e75756edfe996a2d225a8ede8bd72124fcf1876da1d3244223 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 lastState: terminated: containerID: cri-o://5aef46f74824b7bd8319047c24bacbb9cbf6ac782ef810c2be78f3961d31d75e exitCode: 1 finishedAt: "2026-03-20T08:47:17Z" message: | le_one.go:314] "Successfully bound pod to node" pod="openshift-multus/multus-admission-controller-649577484c-p72cd" node="master-0" evaluatedNodes=1 feasibleNodes=1 I0320 08:44:18.428728 1 schedule_one.go:314] "Successfully bound pod to node" pod="openshift-network-console/networking-console-plugin-7c6b76c555-zbpk4" node="master-0" evaluatedNodes=1 feasibleNodes=1 I0320 08:44:21.688646 1 schedule_one.go:314] "Successfully bound pod to node" pod="openshift-monitoring/thanos-querier-5dc898db58-5m6qg" node="master-0" evaluatedNodes=1 feasibleNodes=1 I0320 08:44:24.420059 1 schedule_one.go:314] "Successfully bound pod to node" pod="openshift-monitoring/metrics-server-646d59cb8b-622bl" node="master-0" evaluatedNodes=1 feasibleNodes=1 I0320 08:44:30.638302 1 schedule_one.go:314] "Successfully bound pod to node" pod="openshift-monitoring/alertmanager-main-0" node="master-0" evaluatedNodes=1 feasibleNodes=1 I0320 08:44:37.197025 1 schedule_one.go:314] "Successfully bound pod to node" pod="openshift-monitoring/prometheus-k8s-0" node="master-0" evaluatedNodes=1 feasibleNodes=1 E0320 08:45:30.553722 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path E0320 08:46:24.054797 1 leaderelection.go:436] error retrieving resource lock openshift-kube-scheduler/kube-scheduler: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-scheduler/leases/kube-scheduler?timeout=53.5s": net/http: request canceled (Client.Timeout exceeded while awaiting headers) I0320 08:46:43.550833 1 leaderelection.go:297] failed to renew lease openshift-kube-scheduler/kube-scheduler: timed out waiting for the condition E0320 08:47:17.553748 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded E0320 08:47:17.553836 1 server.go:309] "Leaderelection lost" reason: Error startedAt: "2026-03-20T08:40:35Z" name: kube-scheduler ready: true restartCount: 1 started: true state: running: startedAt: "2026-03-20T08:47:18Z" - containerID: cri-o://aa54b546db103668e772defaa4b64e4e2c01b2bc8d91706ab1484cd99b14f9d9 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 lastState: {} name: kube-scheduler-cert-syncer ready: true restartCount: 0 started: true state: running: startedAt: "2026-03-20T08:40:36Z" - containerID: cri-o://21af743d9ba623a159aff49736a2e159bcf30ef8644fe5c3ed418758349b2110 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ef199844317b7b012879ed8d29f9b6bc37fad8a6fdb336103cbd5cabc74c4302 lastState: terminated: containerID: cri-o://3c189090fb625d43e4a0aad0248864aa122ec54e7ab96d232c95dbcba79fcc95 exitCode: 0 finishedAt: "2026-03-20T08:47:19Z" reason: Completed startedAt: "2026-03-20T08:40:36Z" name: kube-scheduler-recovery-controller ready: true restartCount: 1 started: true state: running: startedAt: "2026-03-20T08:47:20Z" hostIP: 192.168.32.10 hostIPs: - ip: 192.168.32.10 initContainerStatuses: - containerID: cri-o://2bb061585e269ad50f22944212861cbe8de65df048c827dffb60a910fb8f58b1 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b23c544d3894e5b31f66a18c554f03b0d29f92c2000c46b57b1c96da7ec25db9 lastState: {} name: wait-for-host-port ready: true restartCount: 0 started: false state: terminated: containerID: cri-o://2bb061585e269ad50f22944212861cbe8de65df048c827dffb60a910fb8f58b1 exitCode: 0 finishedAt: "2026-03-20T08:40:34Z" reason: Completed startedAt: "2026-03-20T08:40:34Z" phase: Running podIP: 192.168.32.10 podIPs: - ip: 192.168.32.10 qosClass: Burstable startTime: "2026-03-20T08:40:56Z"