Name:             watcher-operator-controller-manager-5c9559cd7b-g4s97
Namespace:        openstack-operators
Priority:         0
Service Account:  watcher-operator-controller-manager
Node:             crc/192.168.126.11
Start Time:       Wed, 28 Jan 2026 16:58:11 +0000
Labels:           control-plane=controller-manager
                  openstack.org/operator-name=watcher
                  pod-template-hash=5c9559cd7b
Annotations:      k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.87/23"],"mac_address":"0a:58:0a:d9:00:57","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.87"
                        ],
                        "mac": "0a:58:0a:d9:00:57",
                        "default": true,
                        "dns": {}
                    }]
                  kubectl.kubernetes.io/default-container: manager
                  openshift.io/scc: anyuid
Status:           Running
IP:               10.217.0.87
IPs:
  IP:           10.217.0.87
Controlled By:  ReplicaSet/watcher-operator-controller-manager-5c9559cd7b
Containers:
  manager:
    Container ID:  cri-o://d960199df588a1f89119002cfca778fa0ae27b706f35545c69c89c50d80b2a91
    Image:         quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64
    Image ID:      quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64
    Port:          <none>
    Host Port:     <none>
    Command:
      /manager
    Args:
      --health-probe-bind-address=:8081
      --metrics-bind-address=127.0.0.1:8080
      --leader-elect
    State:          Running
      Started:      Wed, 28 Jan 2026 17:15:58 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Wed, 28 Jan 2026 17:07:35 +0000
      Finished:     Wed, 28 Jan 2026 17:15:49 +0000
    Ready:          True
    Restart Count:  2
    Limits:
      cpu:     500m
      memory:  512Mi
    Requests:
      cpu:      10m
      memory:   256Mi
    Liveness:   http-get http://:8081/healthz delay=15s timeout=1s period=20s #success=1 #failure=3
    Readiness:  http-get http://:8081/readyz delay=5s timeout=1s period=10s #success=1 #failure=3
    Environment:
      LEASE_DURATION:   30
      RENEW_DEADLINE:   20
      RETRY_PERIOD:     5
      ENABLE_WEBHOOKS:  false
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-m9gkf (ro)
  kube-rbac-proxy:
    Container ID:  cri-o://6f418b8c1b07875f8bbc985832889bf6acf754e976c773e0a67fcda0c862f7c9
    Image:         quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a
    Image ID:      quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://127.0.0.1:8080/
      --logtostderr=true
      --v=0
    State:          Running
      Started:      Wed, 28 Jan 2026 16:59:37 +0000
    Ready:          True
    Restart Count:  0
    Limits:
      cpu:     500m
      memory:  128Mi
    Requests:
      cpu:        5m
      memory:     64Mi
    Environment:  <none>
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-m9gkf (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  kube-api-access-m9gkf:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events:
  Type     Reason                           Age                From               Message
  ----     ------                           ----               ----               -------
  Normal   Scheduled                        36m                default-scheduler  Successfully assigned openstack-operators/watcher-operator-controller-manager-5c9559cd7b-g4s97 to crc
  Normal   AddedInterface                   36m                multus             Add eth0 [10.217.0.87/23] from ovn-kubernetes
  Warning  Failed                           35m                kubelet            Failed to pull image "quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64": rpc error: code = Canceled desc = copying config: context canceled
  Warning  Failed                           35m                kubelet            Error: ErrImagePull
  Normal   Pulled                           35m                kubelet            Container image "quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a" already present on machine
  Normal   Pulling                          35m (x2 over 36m)  kubelet            Pulling image "quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64"
  Normal   Created                          35m                kubelet            Created container kube-rbac-proxy
  Normal   Started                          35m                kubelet            Started container kube-rbac-proxy
  Normal   Pulled                           34m                kubelet            Successfully pulled image "quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64" in 9.036s (9.036s including waiting). Image size: 178199440 bytes.
  Normal   Created                          27m (x2 over 34m)  kubelet            Created container manager
  Normal   Started                          27m (x2 over 34m)  kubelet            Started container manager
  Warning  Unhealthy                        18m (x3 over 27m)  kubelet            Readiness probe failed: Get "http://10.217.0.87:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy                        18m (x3 over 27m)  kubelet            Liveness probe failed: Get "http://10.217.0.87:8081/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy                        18m (x2 over 27m)  kubelet            Liveness probe failed: Get "http://10.217.0.87:8081/healthz": dial tcp 10.217.0.87:8081: connect: connection refused
  Warning  FailedToRetrieveImagePullSecret  18m                kubelet            Unable to retrieve some image pull secrets (watcher-operator-controller-manager-dockercfg-rmw2s); attempting to pull the image may not succeed.
  Normal   Killing                          18m                kubelet            Container manager failed liveness probe, will be restarted
  Normal   Pulled                           18m (x2 over 27m)  kubelet            Container image "quay.io/openstack-k8s-operators/watcher-operator@sha256:b2aa2de20b023b2e0a5a130e7f5e0adaa628e639ce65985e3c4bc80f7ee7cd64" already present on machine
  Warning  Unhealthy                        18m (x4 over 27m)  kubelet            Readiness probe failed: Get "http://10.217.0.87:8081/readyz": dial tcp 10.217.0.87:8081: connect: connection refused
