Name:             ceilometer-0
Namespace:        openstack
Priority:         0
Service Account:  telemetry-ceilometer
Node:             crc/192.168.126.11
Start Time:       Fri, 05 Dec 2025 12:48:57 +0000
Labels:           apps.kubernetes.io/pod-index=0
                  controller-revision-hash=ceilometer-689bc57d47
                  owner=ceilometer
                  service=ceilometer
                  statefulset.kubernetes.io/pod-name=ceilometer-0
Annotations:      k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.224/23"],"mac_address":"0a:58:0a:d9:00:e0","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0....
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.224"
                        ],
                        "mac": "0a:58:0a:d9:00:e0",
                        "default": true,
                        "dns": {}
                    }]
                  k8s.v1.cni.cncf.io/networks: []
                  openshift.io/scc: anyuid
Status:           Running
IP:               10.217.0.224
IPs:
  IP:           10.217.0.224
Controlled By:  StatefulSet/ceilometer
Containers:
  ceilometer-central-agent:
    Container ID:  cri-o://d717437910c93d9d2568879808ee962446e10526f12d5b81d76de7e0416ed7b8
    Image:         38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest
    Image ID:      38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-central@sha256:5c782cbcdff83002eba7fad900784d0ca3732d18ae2a36d60e8d6b174d8124bf
    Port:          <none>
    Host Port:     <none>
    Command:
      /bin/bash
    Args:
      -c
      /usr/local/bin/kolla_start
    State:          Running
      Started:      Fri, 05 Dec 2025 12:48:58 +0000
    Ready:          True
    Restart Count:  0
    Liveness:       exec [/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py] delay=300s timeout=5s period=5s #success=1 #failure=3
    Environment:
      CONFIG_HASH:            n5fh648h55ch584h68hcfh656h678h5fbh595h5dchf7h85hb8h7h5c9h9h58bh55fh67dh577h96h5cdhf9h9bh69hc4hc5h566h7hf4h677q
      KOLLA_CONFIG_STRATEGY:  COPY_ALWAYS
    Mounts:
      /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem from combined-ca-bundle (ro,path="tls-ca-bundle.pem")
      /var/lib/kolla/config_files/config.json from config-data (ro,path="ceilometer-central-config.json")
      /var/lib/openstack/bin from scripts (ro)
      /var/lib/openstack/config from config-data (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-l6g85 (ro)
  ceilometer-notification-agent:
    Container ID:  cri-o://777a08bfcc533e258bab34217b1328f2659ee0eb77a0b3a59e372c6991c15894
    Image:         38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-notification:watcher_latest
    Image ID:      38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-notification@sha256:06ebaec7f7d962a3e00b188cae1cc8af44cec120a288f8bf0e7f6da123144fb4
    Port:          <none>
    Host Port:     <none>
    Command:
      /bin/bash
    Args:
      -c
      /usr/local/bin/kolla_start
    State:          Running
      Started:      Fri, 05 Dec 2025 12:48:58 +0000
    Ready:          True
    Restart Count:  0
    Liveness:       exec [/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py] delay=300s timeout=5s period=30s #success=1 #failure=3
    Environment:
      CONFIG_HASH:            n5fh648h55ch584h68hcfh656h678h5fbh595h5dchf7h85hb8h7h5c9h9h58bh55fh67dh577h96h5cdhf9h9bh69hc4hc5h566h7hf4h677q
      KOLLA_CONFIG_STRATEGY:  COPY_ALWAYS
    Mounts:
      /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem from combined-ca-bundle (ro,path="tls-ca-bundle.pem")
      /var/lib/kolla/config_files/config.json from config-data (ro,path="ceilometer-notification-config.json")
      /var/lib/openstack/bin from scripts (ro)
      /var/lib/openstack/config from config-data (ro)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-l6g85 (ro)
  sg-core:
    Container ID:   cri-o://5f92b16ea9a3a69cb30b790f4508dcf35c3a047c7335752685dd8f57f7fa743c
    Image:          quay.io/openstack-k8s-operators/sg-core:latest
    Image ID:       quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1
    Port:           <none>
    Host Port:      <none>
    State:          Running
      Started:      Fri, 05 Dec 2025 12:48:59 +0000
    Ready:          True
    Restart Count:  0
    Environment:    <none>
    Mounts:
      /etc/sg-core.conf.yaml from sg-core-conf-yaml (rw,path="sg-core.conf.yaml")
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-l6g85 (ro)
  proxy-httpd:
    Container ID:  cri-o://cfe12d12c664071d0a9e1772ffaca324796ef5dce707e17cfd8daa242183f309
    Image:         registry.redhat.io/ubi9/httpd-24:latest
    Image ID:      registry.redhat.io/ubi9/httpd-24@sha256:379ba2fc60106c2026bafa8e7d0ad585d87e6e7fe3f0a6a3d714f744e450bdf2
    Port:          3000/TCP
    Host Port:     0/TCP
    Command:
      /usr/sbin/httpd
    Args:
      -DFOREGROUND
    State:          Running
      Started:      Fri, 05 Dec 2025 12:49:01 +0000
    Ready:          True
    Restart Count:  0
    Liveness:       http-get https://:3000/ delay=300s timeout=30s period=30s #success=1 #failure=3
    Readiness:      http-get https://:3000/ delay=10s timeout=30s period=30s #success=1 #failure=3
    Environment:    <none>
    Mounts:
      /etc/httpd/conf.d/ssl.conf from config-data (ro,path="ssl.conf")
      /etc/httpd/conf/httpd.conf from config-data (ro,path="httpd.conf")
      /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem from combined-ca-bundle (ro,path="tls-ca-bundle.pem")
      /etc/pki/tls/certs/tls.crt from ceilometer-tls-certs (ro,path="tls.crt")
      /etc/pki/tls/private/tls.key from ceilometer-tls-certs (ro,path="tls.key")
      /run/httpd from run-httpd (rw)
      /var/log/httpd from log-httpd (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-l6g85 (ro)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  scripts:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  ceilometer-scripts
    Optional:    false
  config-data:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  ceilometer-config-data
    Optional:    false
  sg-core-conf-yaml:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  ceilometer-config-data
    Optional:    false
  run-httpd:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:     
    SizeLimit:  <unset>
  log-httpd:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:     
    SizeLimit:  <unset>
  ceilometer-tls-certs:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  cert-ceilometer-internal-svc
    Optional:    false
  combined-ca-bundle:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  combined-ca-bundle
    Optional:    false
  kube-api-access-l6g85:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   BestEffort
Node-Selectors:              <none>
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type    Reason          Age   From               Message
  ----    ------          ----  ----               -------
  Normal  Scheduled       83m   default-scheduler  Successfully assigned openstack/ceilometer-0 to crc
  Normal  AddedInterface  83m   multus             Add eth0 [10.217.0.224/23] from ovn-kubernetes
  Normal  Pulling         83m   kubelet            Pulling image "38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest"
  Normal  Pulled          83m   kubelet            Successfully pulled image "38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-central:watcher_latest" in 52ms (52ms including waiting). Image size: 601984160 bytes.
  Normal  Created         83m   kubelet            Created container ceilometer-central-agent
  Normal  Started         83m   kubelet            Started container ceilometer-central-agent
  Normal  Pulling         83m   kubelet            Pulling image "38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-notification:watcher_latest"
  Normal  Pulled          83m   kubelet            Successfully pulled image "38.102.83.151:5001/podified-master-centos10/openstack-ceilometer-notification:watcher_latest" in 47ms (47ms including waiting). Image size: 591520932 bytes.
  Normal  Created         83m   kubelet            Created container ceilometer-notification-agent
  Normal  Started         83m   kubelet            Started container ceilometer-notification-agent
  Normal  Pulling         83m   kubelet            Pulling image "quay.io/openstack-k8s-operators/sg-core:latest"
  Normal  Pulled          83m   kubelet            Successfully pulled image "quay.io/openstack-k8s-operators/sg-core:latest" in 475ms (475ms including waiting). Image size: 299354633 bytes.
  Normal  Created         83m   kubelet            Created container sg-core
  Normal  Started         83m   kubelet            Started container sg-core
  Normal  Pulling         83m   kubelet            Pulling image "registry.redhat.io/ubi9/httpd-24:latest"
  Normal  Pulled          83m   kubelet            Successfully pulled image "registry.redhat.io/ubi9/httpd-24:latest" in 1.294s (1.294s including waiting). Image size: 313797465 bytes.
  Normal  Created         83m   kubelet            Created container proxy-httpd
  Normal  Started         83m   kubelet            Started container proxy-httpd
