Name:                 control-plane-machine-set-operator-78cbb6b69f-tbh9v
Namespace:            openshift-machine-api
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Service Account:      control-plane-machine-set-operator
Node:                 crc/192.168.126.11
Start Time:           Wed, 25 Feb 2026 08:04:30 +0000
Labels:               k8s-app=control-plane-machine-set-operator
                      pod-template-hash=78cbb6b69f
Annotations:          k8s.ovn.org/pod-networks:
                        {"default":{"ip_addresses":["10.217.0.33/23"],"mac_address":"0a:58:0a:d9:00:21","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                      k8s.v1.cni.cncf.io/network-status:
                        [{
                            "name": "ovn-kubernetes",
                            "interface": "eth0",
                            "ips": [
                                "10.217.0.33"
                            ],
                            "mac": "0a:58:0a:d9:00:21",
                            "default": true,
                            "dns": {}
                        }]
                      openshift.io/required-scc: restricted-v2
                      openshift.io/scc: restricted-v2
                      seccomp.security.alpha.kubernetes.io/pod: runtime/default
Status:               Running
SeccompProfile:       RuntimeDefault
IP:                   10.217.0.33
IPs:
  IP:           10.217.0.33
Controlled By:  ReplicaSet/control-plane-machine-set-operator-78cbb6b69f
Containers:
  control-plane-machine-set-operator:
    Container ID:  cri-o://4adcc69d9d1bdb782414d8f4c0a9f456b2d229e27808712878646fc3096eb6e3
    Image:         quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717
    Image ID:      quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717
    Port:          9443/TCP
    Host Port:     0/TCP
    Command:
      /manager
    Args:
      -v=2
      --leader-elect=true
      --leader-elect-lease-duration=137s
      --leader-elect-renew-deadline=107s
      --leader-elect-retry-period=26s
      --leader-elect-resource-namespace=openshift-machine-api
    State:       Running
      Started:   Fri, 27 Feb 2026 16:35:24 +0000
    Last State:  Terminated
      Reason:    Error
      Message:   d-b4ba-b4fb2159070a" namespace="openshift-machine-api" name="cluster"
I0227 16:26:28.141398       1 controller.go:184] "Finished reconciling control plane machine set" controller="controlplanemachineset" reconcileID="44c7b007-41fd-475d-b4ba-b4fb2159070a" namespace="openshift-machine-api" name="cluster"
E0227 16:33:36.728587       1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path
E0227 16:34:36.734637       1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/control-plane-machine-set-leader: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io control-plane-machine-set-leader)
I0227 16:34:49.724912       1 leaderelection.go:297] failed to renew lease openshift-machine-api/control-plane-machine-set-leader: timed out waiting for the condition
E0227 16:35:23.741432       1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded
I0227 16:35:23.743253       1 internal.go:538] "Stopping and waiting for non leader election runnables"
I0227 16:35:23.743554       1 internal.go:542] "Stopping and waiting for leader election runnables"
I0227 16:35:23.743624       1 internal.go:550] "Stopping and waiting for caches"
I0227 16:35:23.743444       1 recorder.go:104] "control-plane-machine-set-operator-78cbb6b69f-tbh9v_66b14e49-2772-4c07-af99-dabac31519f6 stopped leading" logger="events" type="Normal" object={"kind":"Lease","namespace":"openshift-machine-api","name":"control-plane-machine-set-leader","uid":"9cdaee0e-f943-4120-a7ce-8e87ad485f3d","apiVersion":"coordination.k8s.io/v1","resourceVersion":"35198"} reason="LeaderElection"
I0227 16:35:23.743710       1 internal.go:554] "Stopping and waiting for webhooks"
E0227 16:35:23.743670       1 main.go:233] "problem running manager" err="leader election lost" logger="setup"

      Exit Code:    1
      Started:      Fri, 27 Feb 2026 16:25:50 +0000
      Finished:     Fri, 27 Feb 2026 16:35:23 +0000
    Ready:          True
    Restart Count:  2
    Requests:
      cpu:     10m
      memory:  50Mi
    Environment:
      RELEASE_VERSION:      4.18.1
      COMPONENT_NAMESPACE:  openshift-machine-api (v1:metadata.namespace)
    Mounts:
      /tmp/k8s-webhook-server/serving-certs from control-plane-machine-set-operator-tls (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-66ftl (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  control-plane-machine-set-operator-tls:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  control-plane-machine-set-operator-tls
    Optional:    false
  kube-api-access-66ftl:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              node-role.kubernetes.io/master=
Tolerations:                 node-role.kubernetes.io/master:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason                  Age   From               Message
  ----     ------                  ----  ----               -------
  Warning  FailedScheduling        2d8h  default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/unreachable: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
  Normal   Scheduled               2d8h  default-scheduler  Successfully assigned openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-tbh9v to crc
  Normal   AddedInterface          2d8h  multus             Add eth0 [10.217.0.33/23] from ovn-kubernetes
  Normal   Pulled                  2d8h  kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717" already present on machine
  Normal   Created                 2d8h  kubelet            Created container control-plane-machine-set-operator
  Normal   Started                 2d8h  kubelet            Started container control-plane-machine-set-operator
  Warning  FailedMount             22m   kubelet            MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" : failed to sync secret cache: timed out waiting for the condition
  Warning  FailedCreatePodSandBox  21m   kubelet            Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_control-plane-machine-set-operator-78cbb6b69f-tbh9v_openshift-machine-api_6ddf71c1-114a-4d9f-b40c-deb00b9738a2_0(467e9155ee9a14a5761c6f5656bd5d50dabb49500cde467233d0335064697907): error adding pod openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-tbh9v to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): timed out waiting for the condition
  Normal   AddedInterface          20m   multus             Add eth0 [10.217.0.33/23] from ovn-kubernetes
  Normal   Pulled                  20m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717" already present on machine
  Normal   Created                 20m   kubelet            Created container control-plane-machine-set-operator
  Normal   Started                 20m   kubelet            Started container control-plane-machine-set-operator
  Normal   Pulled                  11m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717" already present on machine
  Normal   Created                 11m   kubelet            Created container control-plane-machine-set-operator
  Normal   Started                 11m   kubelet            Started container control-plane-machine-set-operator
