Name:             octavia-operator-controller-manager-6f8c446f86-nfcgp
Namespace:        openstack-operators
Priority:         0
Service Account:  octavia-operator-controller-manager
Node:             crc/192.168.126.11
Start Time:       Sat, 29 Nov 2025 05:58:30 +0000
Labels:           app.kubernetes.io/name=octavia-operator
                  control-plane=controller-manager
                  openstack.org/operator-name=octavia
                  pod-template-hash=6f8c446f86
Annotations:      k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.80/23"],"mac_address":"0a:58:0a:d9:00:50","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.80"
                        ],
                        "mac": "0a:58:0a:d9:00:50",
                        "default": true,
                        "dns": {}
                    }]
                  kubectl.kubernetes.io/default-container: manager
                  openshift.io/scc: anyuid
Status:           Running
IP:               10.217.0.80
IPs:
  IP:           10.217.0.80
Controlled By:  ReplicaSet/octavia-operator-controller-manager-6f8c446f86
Containers:
  manager:
    Container ID:  cri-o://18c7fe408fd62fe417a2e0379780a8906798355c0d03cfdbd1e2b6b4242d01dd
    Image:         quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168
    Image ID:      quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168
    Port:          <none>
    Host Port:     <none>
    Command:
      /manager
    Args:
      --leader-elect
      --health-probe-bind-address=:8081
      --metrics-bind-address=127.0.0.1:8080
    State:          Running
      Started:      Sat, 29 Nov 2025 06:53:34 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Sat, 29 Nov 2025 05:59:03 +0000
      Finished:     Sat, 29 Nov 2025 06:53:19 +0000
    Ready:          True
    Restart Count:  1
    Limits:
      cpu:     500m
      memory:  512Mi
    Requests:
      cpu:      10m
      memory:   256Mi
    Liveness:   http-get http://:8081/healthz delay=15s timeout=1s period=20s #success=1 #failure=3
    Readiness:  http-get http://:8081/readyz delay=5s timeout=1s period=10s #success=1 #failure=3
    Environment:
      LEASE_DURATION:   30
      RENEW_DEADLINE:   20
      RETRY_PERIOD:     5
      ENABLE_WEBHOOKS:  false
      METRICS_CERTS:    false
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-fxzrn (ro)
  kube-rbac-proxy:
    Container ID:  cri-o://cb979db8289219d37c94ece396edc453add2a9a6468b148efda8f0195a5eabb9
    Image:         quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0
    Image ID:      quay.io/openstack-k8s-operators/kube-rbac-proxy@sha256:d28df2924a366ed857d6c2c14baac9741238032d41f3d02c12cd757189b68b8a
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://127.0.0.1:8080/
      --logtostderr=true
      --v=0
    State:          Running
      Started:      Sat, 29 Nov 2025 05:59:07 +0000
    Ready:          True
    Restart Count:  0
    Limits:
      cpu:     500m
      memory:  128Mi
    Requests:
      cpu:        5m
      memory:     64Mi
    Environment:  <none>
    Mounts:
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-fxzrn (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  kube-api-access-fxzrn:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   Burstable
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/memory-pressure:NoSchedule op=Exists
                             node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events:
  Type     Reason          Age                    From               Message
  ----     ------          ----                   ----               -------
  Normal   Scheduled       62m                    default-scheduler  Successfully assigned openstack-operators/octavia-operator-controller-manager-6f8c446f86-nfcgp to crc
  Normal   AddedInterface  62m                    multus             Add eth0 [10.217.0.80/23] from ovn-kubernetes
  Warning  Failed          62m                    kubelet            Failed to pull image "quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168": pull QPS exceeded
  Warning  Failed          62m                    kubelet            Error: ErrImagePull
  Warning  Failed          62m                    kubelet            Failed to pull image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0": pull QPS exceeded
  Warning  Failed          62m                    kubelet            Error: ErrImagePull
  Normal   BackOff         62m (x2 over 62m)      kubelet            Back-off pulling image "quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168"
  Warning  Failed          62m (x2 over 62m)      kubelet            Error: ImagePullBackOff
  Normal   BackOff         62m (x2 over 62m)      kubelet            Back-off pulling image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0"
  Warning  Failed          62m (x2 over 62m)      kubelet            Error: ImagePullBackOff
  Normal   Pulling         62m (x2 over 62m)      kubelet            Pulling image "quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168"
  Normal   Pulled          62m                    kubelet            Successfully pulled image "quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" in 11.892s (11.892s including waiting)
  Normal   Pulling         62m (x2 over 62m)      kubelet            Pulling image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0"
  Normal   Pulled          62m                    kubelet            Successfully pulled image "quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" in 2.481s (2.481s including waiting)
  Normal   Created         62m                    kubelet            Created container kube-rbac-proxy
  Normal   Started         62m                    kubelet            Started container kube-rbac-proxy
  Normal   Pulled          7m53s                  kubelet            Container image "quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" already present on machine
  Normal   Created         7m40s (x2 over 62m)    kubelet            Created container manager
  Normal   Started         7m38s (x2 over 62m)    kubelet            Started container manager
  Warning  Unhealthy       7m33s (x2 over 7m33s)  kubelet            Readiness probe failed: Get "http://10.217.0.80:8081/readyz": dial tcp 10.217.0.80:8081: connect: connection refused
