Name:             barbican-operator-controller-manager-7b64f4fb85-x22x8
Namespace:        openstack-operators
Priority:         0
Service Account:  barbican-operator-controller-manager
Node:             crc/192.168.126.11
Start Time:       Tue, 25 Nov 2025 20:32:39 +0000
Labels:           app.kubernetes.io/name=barbican-operator
                  control-plane=controller-manager
                  openstack.org/operator-name=barbican
                  pod-template-hash=7b64f4fb85
Annotations:      k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.71/23"],"mac_address":"0a:58:0a:d9:00:47","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.71"
                        ],
                        "mac": "0a:58:0a:d9:00:47",
                        "default": true,
                        "dns": {}
                    }]
                  kubectl.kubernetes.io/default-container: manager
                  openshift.io/scc: anyuid
Status:           Running
IP:               10.217.0.71
IPs:
  IP:           10.217.0.71
Controlled By:  ReplicaSet/barbican-operator-controller-manager-7b64f4fb85
Containers:
  manager:
    Container ID:  cri-o://42c9601b82036d3c7f53e120918358ad210e832ddc06eb6afe9ba0ff6cf347b7
    Image:         quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c
    Image ID:      quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c
    Port:          <none>
    Host Port:     <none>
    Command:
      /manager
    Args:
      --leader-elect
      --health-probe-bind-address=:8081
      --metrics-bind-address=127.0.0.1:8080
    State:          Running
      Started:      Tue, 25 Nov 2025 21:02:16 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 25 Nov 2025 21:01:40 +0000
      Finished:     Tue, 25 Nov 2025 21:01:41 +0000
    Ready:          True
    Restart Count:  3
    Limits:
      cpu:     500m
      memory:  512Mi
    Requests:
      cpu:      10m
      memory:   256Mi
    Liveness:   http-get http://:8081/healthz delay=15s timeout=1s period=20s #success=1 #failure=3
    Readiness:  http-get http://:8081/readyz delay=5s timeout=1s period=10s #success=1 #failure=3
    Environment:
      LEASE_DURATION:   30
      RENEW_DEADLINE:   20
      RETRY_PERIOD:     5
      ENABLE_WEBHOOKS:  false
      METRICS_CERTS:    false
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gsxjq (ro)
  kube-rbac-proxy:
    Container ID:  cri-o://68e6f84df5e6cad09a9a0d9aed07036ff81fc1d2110b8e89103e8eb5db630fea
    Image:         quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0
    Image ID:      quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://127.0.0.1:8080/
      --logtostderr=true
      --v=0
    State:          Running
      Started:      Tue, 25 Nov 2025 20:35:06 +0000
    Ready:          True
    Restart Count:  0
    Limits:
      cpu:     500m
      memory:  128Mi
    Requests:
      cpu:        5m
      memory:     64Mi
    Environment:  <none>
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-gsxjq (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  kube-api-access-gsxjq:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events:
  Type     Reason          Age                From               Message
  ----     ------          ----               ----               -------
  Normal   Scheduled       56m                default-scheduler  Successfully assigned openstack-operators/barbican-operator-controller-manager-7b64f4fb85-x22x8 to crc
  Normal   AddedInterface  56m                multus             Add eth0 [10.217.0.71/23] from ovn-kubernetes
  Warning  Failed          55m                kubelet            Failed to pull image "quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c": rpc error: code = Canceled desc = copying config: context canceled
  Warning  Failed          55m                kubelet            Error: ErrImagePull
  Normal   Pulling         55m                kubelet            Pulling image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0"
  Warning  Failed          54m                kubelet            Failed to pull image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0": rpc error: code = Canceled desc = copying layer: context canceled
  Warning  Failed          54m                kubelet            Error: ErrImagePull
  Normal   Pulling         54m (x2 over 56m)  kubelet            Pulling image "quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c"
  Normal   Pulled          54m                kubelet            Successfully pulled image "quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c" in 7.411s (7.411s including waiting). Image size: 190758874 bytes.
  Normal   Pulled          54m                kubelet            Container image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" already present on machine
  Normal   Created         54m                kubelet            Created container kube-rbac-proxy
  Normal   Started         54m                kubelet            Started container kube-rbac-proxy
  Warning  Unhealthy       54m (x4 over 54m)  kubelet            Readiness probe failed: Get "http://10.217.0.71:8081/readyz": dial tcp 10.217.0.71:8081: connect: connection refused
  Warning  Unhealthy       54m                kubelet            Readiness probe failed: Get "http://10.217.0.71:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Normal   Created         27m (x3 over 54m)  kubelet            Created container manager
  Normal   Started         27m (x3 over 54m)  kubelet            Started container manager
  Normal   Pulled          27m (x2 over 27m)  kubelet            Container image "quay.io/openstack-k8s-operators/barbican-operator@sha256:3dbf9fd9dce75f1fb250ee4c4097ad77d2f34110b61d85e37abd9c472e022e6c" already present on machine
  Warning  BackOff         27m (x6 over 27m)  kubelet            Back-off restarting failed container manager in pod barbican-operator-controller-manager-7b64f4fb85-x22x8_openstack-operators(6a6e9dee-261f-4a07-b71c-c3eebc403e16)
