Name:                 machine-api-operator-5694c8668f-chf57
Namespace:            openshift-machine-api
Priority:             2000001000
Priority Class Name:  system-node-critical
Service Account:      machine-api-operator
Node:                 crc/192.168.126.11
Start Time:           Mon, 24 Nov 2025 13:21:13 +0000
Labels:               k8s-app=machine-api-operator
                      pod-template-hash=5694c8668f
Annotations:          capability.openshift.io/name: MachineAPI
                      k8s.ovn.org/pod-networks:
                        {"default":{"ip_addresses":["10.217.0.12/23"],"mac_address":"0a:58:0a:d9:00:0c","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                      k8s.v1.cni.cncf.io/network-status:
                        [{
                            "name": "ovn-kubernetes",
                            "interface": "eth0",
                            "ips": [
                                "10.217.0.12"
                            ],
                            "mac": "0a:58:0a:d9:00:0c",
                            "default": true,
                            "dns": {}
                        }]
                      kubectl.kubernetes.io/default-container: machine-api-operator
                      openshift.io/required-scc: restricted-v2
                      openshift.io/scc: restricted-v2
                      seccomp.security.alpha.kubernetes.io/pod: runtime/default
Status:               Running
SeccompProfile:       RuntimeDefault
IP:                   10.217.0.12
IPs:
  IP:           10.217.0.12
Controlled By:  ReplicaSet/machine-api-operator-5694c8668f
Containers:
  kube-rbac-proxy:
    Container ID:  cri-o://5e128cf1aca2ce50fc1ee1aa2345c0352f4d4cb175cc9adffe4b343264b3a747
    Image:         quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09
    Image ID:      cc56d852e3f7cb630b7d2731dbf5b253fbcf674d3734431686e66aa2bb74810b
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://localhost:8080/
      --tls-cert-file=/etc/tls/private/tls.crt
      --tls-private-key-file=/etc/tls/private/tls.key
      --config-file=/etc/kube-rbac-proxy/config-file.yaml
      --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305
      --logtostderr=true
      --v=3
    State:          Running
      Started:      Wed, 03 Dec 2025 12:44:40 +0000
    Ready:          True
    Restart Count:  1
    Requests:
      cpu:        10m
      memory:     20Mi
    Environment:  <none>
    Mounts:
      /etc/kube-rbac-proxy from config (rw)
      /etc/tls/private from machine-api-operator-tls (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-r8x9q (ro)
  machine-api-operator:
    Container ID:  cri-o://9e43b7aba69cd264b157590f8a398585ef0ebe6599b32eab2bdb27db24c0158b
    Image:         quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93
    Image ID:      quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93
    Port:          <none>
    Host Port:     <none>
    Command:
      /machine-api-operator
    Args:
      start
      --images-json=/etc/machine-api-operator-config/images/images.json
      --alsologtostderr
      --v=3
    State:       Running
      Started:   Wed, 03 Dec 2025 13:01:10 +0000
    Last State:  Terminated
      Reason:    Error
      Message:   d watching *v1beta1.Machine from github.com/openshift/client-go/machine/informers/externalversions/factory.go:125
I1203 12:50:39.365359       1 reflector.go:368] Caches populated for *v1beta1.Machine from github.com/openshift/client-go/machine/informers/externalversions/factory.go:125
I1203 12:50:57.241817       1 reflector.go:341] Listing and watching *v1.ClusterOperator from github.com/openshift/client-go/config/informers/externalversions/factory.go:125
I1203 12:50:57.258792       1 reflector.go:368] Caches populated for *v1.ClusterOperator from github.com/openshift/client-go/config/informers/externalversions/factory.go:125
I1203 12:50:57.277149       1 status.go:69] Syncing status: re-syncing
I1203 12:50:57.286558       1 sync.go:77] Provider is NoOp, skipping synchronisation
I1203 12:50:57.290404       1 status.go:99] Syncing status: available
E1203 12:57:39.516069       1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path
E1203 12:58:39.520516       1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-api-operator)
E1203 13:00:13.530072       1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path
E1203 13:01:05.507478       1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: Get "https://10.217.4.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-machine-api/leases/machine-api-operator": context deadline exceeded
I1203 13:01:05.508245       1 leaderelection.go:297] failed to renew lease openshift-machine-api/machine-api-operator: timed out waiting for the condition
F1203 13:01:09.173156       1 start.go:104] Leader election lost

      Exit Code:    255
      Started:      Wed, 03 Dec 2025 12:44:43 +0000
      Finished:     Wed, 03 Dec 2025 13:01:09 +0000
    Ready:          True
    Restart Count:  2
    Requests:
      cpu:     10m
      memory:  50Mi
    Environment:
      RELEASE_VERSION:      4.18.1
      COMPONENT_NAMESPACE:  openshift-machine-api (v1:metadata.namespace)
      POD_NAME:             machine-api-operator-5694c8668f-chf57 (v1:metadata.name)
      METRICS_PORT:         8080
    Mounts:
      /etc/machine-api-operator-config/images from images (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-r8x9q (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  config:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      kube-rbac-proxy
    Optional:  false
  images:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      machine-api-operator-images
    Optional:  false
  machine-api-operator-tls:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  machine-api-operator-tls
    Optional:    false
  kube-api-access-r8x9q:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              node-role.kubernetes.io/master=
Tolerations:                 node-role.kubernetes.io/master:NoSchedule op=Exists
                             node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events:
  Type     Reason            Age   From               Message
  ----     ------            ----  ----               -------
  Warning  FailedScheduling  9d    default-scheduler  0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/unreachable: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.
  Normal   Scheduled         9d    default-scheduler  Successfully assigned openshift-machine-api/machine-api-operator-5694c8668f-chf57 to crc
  Normal   AddedInterface    9d    multus             Add eth0 [10.217.0.12/23] from ovn-kubernetes
  Normal   Pulled            9d    kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09" already present on machine
  Normal   Created           9d    kubelet            Created container kube-rbac-proxy
  Normal   Started           9d    kubelet            Started container kube-rbac-proxy
  Normal   Pulled            9d    kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           9d    kubelet            Created container machine-api-operator
  Normal   Started           9d    kubelet            Started container machine-api-operator
  Normal   AddedInterface    98m   multus             Add eth0 [10.217.0.12/23] from ovn-kubernetes
  Normal   Pulled            98m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09" already present on machine
  Normal   Created           98m   kubelet            Created container kube-rbac-proxy
  Normal   Started           98m   kubelet            Started container kube-rbac-proxy
  Normal   Pulled            98m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           98m   kubelet            Created container machine-api-operator
  Normal   Started           98m   kubelet            Started container machine-api-operator
  Warning  FailedMount       88m   kubelet            MountVolume.SetUp failed for volume "machine-api-operator-tls" : failed to sync secret cache: timed out waiting for the condition
  Warning  FailedMount       88m   kubelet            MountVolume.SetUp failed for volume "images" : failed to sync configmap cache: timed out waiting for the condition
  Warning  FailedMount       88m   kubelet            MountVolume.SetUp failed for volume "config" : failed to sync configmap cache: timed out waiting for the condition
  Normal   Pulled            81m   kubelet            Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93" already present on machine
  Normal   Created           81m   kubelet            Created container machine-api-operator
  Normal   Started           81m   kubelet            Started container machine-api-operator
