Name:                 machine-api-operator-5694c8668f-2dgn2
Namespace:            openshift-machine-api
Priority:             2000001000
Priority Class Name:  system-node-critical
Service Account:      machine-api-operator
Node:                 crc/192.168.126.11
Start Time:           Wed, 25 Feb 2026 08:04:29 +0000
Labels:               k8s-app=machine-api-operator
                      pod-template-hash=5694c8668f
Annotations:          capability.openshift.io/name: MachineAPI
                      k8s.ovn.org/pod-networks:
                        {"default":{"ip_addresses":["10.217.0.19/23"],"mac_address":"0a:58:0a:d9:00:13","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                      k8s.v1.cni.cncf.io/network-status:
                        [{
                            "name": "ovn-kubernetes",
                            "interface": "eth0",
                            "ips": [
                                "10.217.0.19"
                            ],
                            "mac": "0a:58:0a:d9:00:13",
                            "default": true,
                            "dns": {}
                        }]
                      kubectl.kubernetes.io/default-container: machine-api-operator
                      openshift.io/required-scc: restricted-v2
                      openshift.io/scc: restricted-v2
                      seccomp.security.alpha.kubernetes.io/pod: runtime/default
Status:               Running
SeccompProfile:       RuntimeDefault
IP:                   10.217.0.19
IPs:
  IP:           10.217.0.19
Controlled By:  ReplicaSet/machine-api-operator-5694c8668f
Containers:
  kube-rbac-proxy:
    Container ID:  cri-o://8bbe56c1261ef10ef9171f618e2311891b7695582146ebdb29934749b03febba
    Image:         quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09
    Image ID:      cc56d852e3f7cb630b7d2731dbf5b253fbcf674d3734431686e66aa2bb74810b
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://localhost:8080/
      --tls-cert-file=/etc/tls/private/tls.crt
      --tls-private-key-file=/etc/tls/private/tls.key
      --config-file=/etc/kube-rbac-proxy/config-file.yaml
      --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305
      --logtostderr=true
      --v=3
    State:          Running
      Started:      Thu, 26 Feb 2026 08:47:15 +0000
    Ready:          True
    Restart Count:  1
    Requests:
      cpu:        10m
      memory:     20Mi
    Environment:  <none>
    Mounts:
      /etc/kube-rbac-proxy from config (rw)
      /etc/tls/private from machine-api-operator-tls (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-lbfhq (ro)
  machine-api-operator:
    Container ID:  cri-o://78b42d91b4329743a0971fca83e94d2f2c656a3fdee912e78eba0a295d8a5fb5
    Image:         quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93
    Image ID:      quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93
    Port:          <none>
    Host Port:     <none>
    Command:
      /machine-api-operator
    Args:
      start
      --images-json=/etc/machine-api-operator-config/images/images.json
      --alsologtostderr
      --v=3
    State:       Running
      Started:   Thu, 26 Feb 2026 08:59:11 +0000
    Last State:  Terminated
      Reason:    Error
      Message:   9] Syncing status: re-syncing
I0226 08:47:45.388011       1 sync.go:77] Provider is NoOp, skipping synchronisation
I0226 08:47:45.405729       1 status.go:99] Syncing status: available
I0226 08:47:45.594902       1 status.go:69] Syncing status: re-syncing
I0226 08:47:45.610620       1 sync.go:77] Provider is NoOp, skipping synchronisation
I0226 08:47:45.813343       1 request.go:632] Waited for 202.009424ms due to client-side throttling, not priority and fairness, request: GET:https://10.217.4.1:443/apis/config.openshift.io/v1/clusteroperators/machine-api
I0226 08:47:45.855332       1 status.go:99] Syncing status: available
I0226 08:47:46.000055       1 request.go:632] Waited for 143.590583ms due to client-side throttling, not priority and fairness, request: PUT:https://10.217.4.1:443/apis/config.openshift.io/v1/clusteroperators/machine-api/status
E0226 08:55:16.022295       1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path
E0226 08:56:16.026563       1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-api-operator)
E0226 08:57:50.035823       1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path
E0226 08:58:42.017781       1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: Get "https://10.217.4.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-machine-api/leases/machine-api-operator": context deadline exceeded
I0226 08:58:42.018446       1 leaderelection.go:297] failed to renew lease openshift-machine-api/machine-api-operator: timed out waiting for the condition
F0226 08:59:10.632104       1 start.go:104] Leader election lost

      Exit Code:    255
      Started:      Thu, 26 Feb 2026 08:47:17 +0000
      Finished:     Thu, 26 Feb 2026 08:59:10 +0000
    Ready:          True
    Restart Count:  2
    Requests:
      cpu:     10m
      memory:  50Mi
    Environment:
      RELEASE_VERSION:      4.18.1
      COMPONENT_NAMESPACE:  openshift-machine-api (v1:metadata.namespace)
      POD_NAME:             machine-api-operator-5694c8668f-2dgn2 (v1:metadata.name)
      METRICS_PORT:         8080
    Mounts:
      /etc/machine-api-operator-config/images from images (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-lbfhq (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  config:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      kube-rbac-proxy
    Optional:  false
  images:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      machine-api-operator-images
    Optional:  false
  machine-api-operator-tls:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  machine-api-operator-tls
    Optional:    false
  kube-api-access-lbfhq:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              node-role.kubernetes.io/master=
Tolerations:                 node-role.kubernetes.io/master:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events:
  Type     Reason            Age   From               Message
  ----     ------            ----  ----               -------
  Warning  FailedScheduling  25h   default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/unreachable: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
  Normal   Scheduled         25h   default-scheduler  Successfully assigned openshift-machine-api/machine-api-operator-5694c8668f-2dgn2 to crc
  Normal   AddedInterface    25h   multus             Add eth0 [10.217.0.19/23] from ovn-kubernetes
  Normal   Pulled            25h   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09" already present on machine
  Normal   Created           25h   kubelet            Created container kube-rbac-proxy
  Normal   Started           25h   kubelet            Started container kube-rbac-proxy
  Normal   Pulled            25h   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           25h   kubelet            Created container machine-api-operator
  Normal   Started           25h   kubelet            Started container machine-api-operator
  Normal   AddedInterface    18m   multus             Add eth0 [10.217.0.19/23] from ovn-kubernetes
  Normal   Pulled            18m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09" already present on machine
  Normal   Created           18m   kubelet            Created container kube-rbac-proxy
  Normal   Started           18m   kubelet            Started container kube-rbac-proxy
  Normal   Pulled            18m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           18m   kubelet            Created container machine-api-operator
  Normal   Started           18m   kubelet            Started container machine-api-operator
  Warning  FailedMount       12m   kubelet            MountVolume.SetUp failed for volume "images" : failed to sync configmap cache: timed out waiting for the condition
  Warning  FailedMount       12m   kubelet            MountVolume.SetUp failed for volume "machine-api-operator-tls" : failed to sync secret cache: timed out waiting for the condition
  Warning  FailedMount       12m   kubelet            MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition
  Normal   Pulled            7m2s  kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           7m2s  kubelet            Created container machine-api-operator
  Normal   Started           7m2s  kubelet            Started container machine-api-operator
